dataproc

package
v0.0.0-...-c684e27 Latest Latest
Warning

This package is not in the latest version of its module.

Go to latest
Published: Sep 20, 2024 License: BSD-3-Clause Imports: 4 Imported by: 0

Documentation

Index

Constants

View Source
const (
	CLUSTER_SKIA  = "cluster-173d"
	JOB_ID_PREFIX = "  jobId: "
)

Variables

This section is empty.

Functions

This section is empty.

Types

type PySparkJob

type PySparkJob struct {
	// The main Python file to run.
	PyFile string
	// Any additional arguments to pass onto the job command line.
	Args []string
	// Which cluster should run the job.
	Cluster string
	// Any files to provide to the job.
	Files []string
	// contains filtered or unexported fields
}

PySparkJob describes a PySpark job which runs on a GCE cluster.

func (*PySparkJob) Command

func (j *PySparkJob) Command() *exec.Command

Return the command used to trigger this job.

func (*PySparkJob) Run

func (j *PySparkJob) Run(ctx context.Context) (string, error)

Run the job and return its output.

func (*PySparkJob) Submit

func (j *PySparkJob) Submit(ctx context.Context) (string, error)

Trigger the job and return its ID.

func (*PySparkJob) Wait

func (j *PySparkJob) Wait(ctx context.Context) (string, error)

Wait for the job to complete and return its output.

Jump to

Keyboard shortcuts

? : This menu
/ : Search site
f or F : Jump to
y or Y : Canonical URL