Documentation ¶
Index ¶
- func DataflowJob_IsConstruct(x interface{}) *bool
- func DataflowJob_IsTerraformElement(x interface{}) *bool
- func DataflowJob_IsTerraformResource(x interface{}) *bool
- func DataflowJob_TfResourceType() *string
- func NewDataflowJobTimeoutsOutputReference_Override(d DataflowJobTimeoutsOutputReference, ...)
- func NewDataflowJob_Override(d DataflowJob, scope constructs.Construct, id *string, ...)
- type DataflowJob
- type DataflowJobConfig
- type DataflowJobTimeouts
- type DataflowJobTimeoutsOutputReference
Constants ¶
This section is empty.
Variables ¶
This section is empty.
Functions ¶
func DataflowJob_IsConstruct ¶
func DataflowJob_IsConstruct(x interface{}) *bool
Checks if `x` is a construct.
Use this method instead of `instanceof` to properly detect `Construct` instances, even when the construct library is symlinked.
Explanation: in JavaScript, multiple copies of the `constructs` library on disk are seen as independent, completely different libraries. As a consequence, the class `Construct` in each copy of the `constructs` library is seen as a different class, and an instance of one class will not test as `instanceof` the other class. `npm install` will not create installations like this, but users may manually symlink construct libraries together or use a monorepo tool: in those cases, multiple copies of the `constructs` library can be accidentally installed, and `instanceof` will behave unpredictably. It is safest to avoid using `instanceof`, and using this type-testing method instead.
Returns: true if `x` is an object created from a class which extends `Construct`.
func DataflowJob_IsTerraformElement ¶
func DataflowJob_IsTerraformElement(x interface{}) *bool
Experimental.
func DataflowJob_IsTerraformResource ¶
func DataflowJob_IsTerraformResource(x interface{}) *bool
Experimental.
func DataflowJob_TfResourceType ¶
func DataflowJob_TfResourceType() *string
func NewDataflowJobTimeoutsOutputReference_Override ¶
func NewDataflowJobTimeoutsOutputReference_Override(d DataflowJobTimeoutsOutputReference, terraformResource cdktf.IInterpolatingParent, terraformAttribute *string)
func NewDataflowJob_Override ¶
func NewDataflowJob_Override(d DataflowJob, scope constructs.Construct, id *string, config *DataflowJobConfig)
Create a new {@link https://registry.terraform.io/providers/hashicorp/google/4.79.0/docs/resources/dataflow_job google_dataflow_job} Resource.
Types ¶
type DataflowJob ¶
type DataflowJob interface { cdktf.TerraformResource AdditionalExperiments() *[]*string SetAdditionalExperiments(val *[]*string) AdditionalExperimentsInput() *[]*string // Experimental. CdktfStack() cdktf.TerraformStack // Experimental. Connection() interface{} // Experimental. SetConnection(val interface{}) // Experimental. ConstructNodeMetadata() *map[string]interface{} // Experimental. Count() interface{} // Experimental. SetCount(val interface{}) // Experimental. DependsOn() *[]*string // Experimental. SetDependsOn(val *[]*string) EnableStreamingEngine() interface{} SetEnableStreamingEngine(val interface{}) EnableStreamingEngineInput() interface{} // Experimental. ForEach() cdktf.ITerraformIterator // Experimental. SetForEach(val cdktf.ITerraformIterator) // Experimental. Fqn() *string // Experimental. FriendlyUniqueId() *string Id() *string SetId(val *string) IdInput() *string IpConfiguration() *string SetIpConfiguration(val *string) IpConfigurationInput() *string JobId() *string KmsKeyName() *string SetKmsKeyName(val *string) KmsKeyNameInput() *string Labels() *map[string]*string SetLabels(val *map[string]*string) LabelsInput() *map[string]*string // Experimental. Lifecycle() *cdktf.TerraformResourceLifecycle // Experimental. SetLifecycle(val *cdktf.TerraformResourceLifecycle) MachineType() *string SetMachineType(val *string) MachineTypeInput() *string MaxWorkers() *float64 SetMaxWorkers(val *float64) MaxWorkersInput() *float64 Name() *string SetName(val *string) NameInput() *string Network() *string SetNetwork(val *string) NetworkInput() *string // The tree node. Node() constructs.Node OnDelete() *string SetOnDelete(val *string) OnDeleteInput() *string Parameters() *map[string]*string SetParameters(val *map[string]*string) ParametersInput() *map[string]*string Project() *string SetProject(val *string) ProjectInput() *string // Experimental. Provider() cdktf.TerraformProvider // Experimental. SetProvider(val cdktf.TerraformProvider) // Experimental. Provisioners() *[]interface{} // Experimental. SetProvisioners(val *[]interface{}) // Experimental. RawOverrides() interface{} Region() *string SetRegion(val *string) RegionInput() *string ServiceAccountEmail() *string SetServiceAccountEmail(val *string) ServiceAccountEmailInput() *string SkipWaitOnJobTermination() interface{} SetSkipWaitOnJobTermination(val interface{}) SkipWaitOnJobTerminationInput() interface{} State() *string Subnetwork() *string SetSubnetwork(val *string) SubnetworkInput() *string TempGcsLocation() *string SetTempGcsLocation(val *string) TempGcsLocationInput() *string TemplateGcsPath() *string SetTemplateGcsPath(val *string) TemplateGcsPathInput() *string // Experimental. TerraformGeneratorMetadata() *cdktf.TerraformProviderGeneratorMetadata // Experimental. TerraformMetaArguments() *map[string]interface{} // Experimental. TerraformResourceType() *string Timeouts() DataflowJobTimeoutsOutputReference TimeoutsInput() interface{} TransformNameMapping() *map[string]*string SetTransformNameMapping(val *map[string]*string) TransformNameMappingInput() *map[string]*string Type() *string Zone() *string SetZone(val *string) ZoneInput() *string // Experimental. AddOverride(path *string, value interface{}) // Experimental. GetAnyMapAttribute(terraformAttribute *string) *map[string]interface{} // Experimental. GetBooleanAttribute(terraformAttribute *string) cdktf.IResolvable // Experimental. GetBooleanMapAttribute(terraformAttribute *string) *map[string]*bool // Experimental. GetListAttribute(terraformAttribute *string) *[]*string // Experimental. GetNumberAttribute(terraformAttribute *string) *float64 // Experimental. GetNumberListAttribute(terraformAttribute *string) *[]*float64 // Experimental. GetNumberMapAttribute(terraformAttribute *string) *map[string]*float64 // Experimental. GetStringAttribute(terraformAttribute *string) *string // Experimental. GetStringMapAttribute(terraformAttribute *string) *map[string]*string // Experimental. InterpolationForAttribute(terraformAttribute *string) cdktf.IResolvable // Overrides the auto-generated logical ID with a specific ID. // Experimental. OverrideLogicalId(newLogicalId *string) PutTimeouts(value *DataflowJobTimeouts) ResetAdditionalExperiments() ResetEnableStreamingEngine() ResetId() ResetIpConfiguration() ResetKmsKeyName() ResetLabels() ResetMachineType() ResetMaxWorkers() ResetNetwork() ResetOnDelete() // Resets a previously passed logical Id to use the auto-generated logical id again. // Experimental. ResetOverrideLogicalId() ResetParameters() ResetProject() ResetRegion() ResetServiceAccountEmail() ResetSkipWaitOnJobTermination() ResetSubnetwork() ResetTimeouts() ResetTransformNameMapping() ResetZone() SynthesizeAttributes() *map[string]interface{} // Experimental. ToMetadata() interface{} // Returns a string representation of this construct. ToString() *string // Adds this resource to the terraform JSON output. // Experimental. ToTerraform() interface{} }
Represents a {@link https://registry.terraform.io/providers/hashicorp/google/4.79.0/docs/resources/dataflow_job google_dataflow_job}.
func NewDataflowJob ¶
func NewDataflowJob(scope constructs.Construct, id *string, config *DataflowJobConfig) DataflowJob
Create a new {@link https://registry.terraform.io/providers/hashicorp/google/4.79.0/docs/resources/dataflow_job google_dataflow_job} Resource.
type DataflowJobConfig ¶
type DataflowJobConfig struct { // Experimental. Connection interface{} `field:"optional" json:"connection" yaml:"connection"` // Experimental. Count interface{} `field:"optional" json:"count" yaml:"count"` // Experimental. DependsOn *[]cdktf.ITerraformDependable `field:"optional" json:"dependsOn" yaml:"dependsOn"` // Experimental. ForEach cdktf.ITerraformIterator `field:"optional" json:"forEach" yaml:"forEach"` // Experimental. Lifecycle *cdktf.TerraformResourceLifecycle `field:"optional" json:"lifecycle" yaml:"lifecycle"` // Experimental. Provider cdktf.TerraformProvider `field:"optional" json:"provider" yaml:"provider"` // Experimental. Provisioners *[]interface{} `field:"optional" json:"provisioners" yaml:"provisioners"` // A unique name for the resource, required by Dataflow. // // Docs at Terraform Registry: {@link https://registry.terraform.io/providers/hashicorp/google/4.79.0/docs/resources/dataflow_job#name DataflowJob#name} Name *string `field:"required" json:"name" yaml:"name"` // A writeable location on Google Cloud Storage for the Dataflow job to dump its temporary data. // // Docs at Terraform Registry: {@link https://registry.terraform.io/providers/hashicorp/google/4.79.0/docs/resources/dataflow_job#temp_gcs_location DataflowJob#temp_gcs_location} TempGcsLocation *string `field:"required" json:"tempGcsLocation" yaml:"tempGcsLocation"` // The Google Cloud Storage path to the Dataflow job template. // // Docs at Terraform Registry: {@link https://registry.terraform.io/providers/hashicorp/google/4.79.0/docs/resources/dataflow_job#template_gcs_path DataflowJob#template_gcs_path} TemplateGcsPath *string `field:"required" json:"templateGcsPath" yaml:"templateGcsPath"` // List of experiments that should be used by the job. An example value is ["enable_stackdriver_agent_metrics"]. // // Docs at Terraform Registry: {@link https://registry.terraform.io/providers/hashicorp/google/4.79.0/docs/resources/dataflow_job#additional_experiments DataflowJob#additional_experiments} AdditionalExperiments *[]*string `field:"optional" json:"additionalExperiments" yaml:"additionalExperiments"` // Indicates if the job should use the streaming engine feature. // // Docs at Terraform Registry: {@link https://registry.terraform.io/providers/hashicorp/google/4.79.0/docs/resources/dataflow_job#enable_streaming_engine DataflowJob#enable_streaming_engine} EnableStreamingEngine interface{} `field:"optional" json:"enableStreamingEngine" yaml:"enableStreamingEngine"` // Docs at Terraform Registry: {@link https://registry.terraform.io/providers/hashicorp/google/4.79.0/docs/resources/dataflow_job#id DataflowJob#id}. // // Please be aware that the id field is automatically added to all resources in Terraform providers using a Terraform provider SDK version below 2. // If you experience problems setting this value it might not be settable. Please take a look at the provider documentation to ensure it should be settable. Id *string `field:"optional" json:"id" yaml:"id"` // The configuration for VM IPs. Options are "WORKER_IP_PUBLIC" or "WORKER_IP_PRIVATE". // // Docs at Terraform Registry: {@link https://registry.terraform.io/providers/hashicorp/google/4.79.0/docs/resources/dataflow_job#ip_configuration DataflowJob#ip_configuration} IpConfiguration *string `field:"optional" json:"ipConfiguration" yaml:"ipConfiguration"` // The name for the Cloud KMS key for the job. Key format is: projects/PROJECT_ID/locations/LOCATION/keyRings/KEY_RING/cryptoKeys/KEY. // // Docs at Terraform Registry: {@link https://registry.terraform.io/providers/hashicorp/google/4.79.0/docs/resources/dataflow_job#kms_key_name DataflowJob#kms_key_name} KmsKeyName *string `field:"optional" json:"kmsKeyName" yaml:"kmsKeyName"` // User labels to be specified for the job. // // Keys and values should follow the restrictions specified in the labeling restrictions page. NOTE: Google-provided Dataflow templates often provide default labels that begin with goog-dataflow-provided. Unless explicitly set in config, these labels will be ignored to prevent diffs on re-apply. // // Docs at Terraform Registry: {@link https://registry.terraform.io/providers/hashicorp/google/4.79.0/docs/resources/dataflow_job#labels DataflowJob#labels} Labels *map[string]*string `field:"optional" json:"labels" yaml:"labels"` // The machine type to use for the job. // // Docs at Terraform Registry: {@link https://registry.terraform.io/providers/hashicorp/google/4.79.0/docs/resources/dataflow_job#machine_type DataflowJob#machine_type} MachineType *string `field:"optional" json:"machineType" yaml:"machineType"` // The number of workers permitted to work on the job. More workers may improve processing speed at additional cost. // // Docs at Terraform Registry: {@link https://registry.terraform.io/providers/hashicorp/google/4.79.0/docs/resources/dataflow_job#max_workers DataflowJob#max_workers} MaxWorkers *float64 `field:"optional" json:"maxWorkers" yaml:"maxWorkers"` // The network to which VMs will be assigned. If it is not provided, "default" will be used. // // Docs at Terraform Registry: {@link https://registry.terraform.io/providers/hashicorp/google/4.79.0/docs/resources/dataflow_job#network DataflowJob#network} Network *string `field:"optional" json:"network" yaml:"network"` // One of "drain" or "cancel". Specifies behavior of deletion during terraform destroy. // // Docs at Terraform Registry: {@link https://registry.terraform.io/providers/hashicorp/google/4.79.0/docs/resources/dataflow_job#on_delete DataflowJob#on_delete} OnDelete *string `field:"optional" json:"onDelete" yaml:"onDelete"` // Key/Value pairs to be passed to the Dataflow job (as used in the template). // // Docs at Terraform Registry: {@link https://registry.terraform.io/providers/hashicorp/google/4.79.0/docs/resources/dataflow_job#parameters DataflowJob#parameters} Parameters *map[string]*string `field:"optional" json:"parameters" yaml:"parameters"` // The project in which the resource belongs. // // Docs at Terraform Registry: {@link https://registry.terraform.io/providers/hashicorp/google/4.79.0/docs/resources/dataflow_job#project DataflowJob#project} Project *string `field:"optional" json:"project" yaml:"project"` // The region in which the created job should run. // // Docs at Terraform Registry: {@link https://registry.terraform.io/providers/hashicorp/google/4.79.0/docs/resources/dataflow_job#region DataflowJob#region} Region *string `field:"optional" json:"region" yaml:"region"` // The Service Account email used to create the job. // // Docs at Terraform Registry: {@link https://registry.terraform.io/providers/hashicorp/google/4.79.0/docs/resources/dataflow_job#service_account_email DataflowJob#service_account_email} ServiceAccountEmail *string `field:"optional" json:"serviceAccountEmail" yaml:"serviceAccountEmail"` // If true, treat DRAINING and CANCELLING as terminal job states and do not wait for further changes before removing from terraform state and moving on. // // WARNING: this will lead to job name conflicts if you do not ensure that the job names are different, e.g. by embedding a release ID or by using a random_id. // // Docs at Terraform Registry: {@link https://registry.terraform.io/providers/hashicorp/google/4.79.0/docs/resources/dataflow_job#skip_wait_on_job_termination DataflowJob#skip_wait_on_job_termination} SkipWaitOnJobTermination interface{} `field:"optional" json:"skipWaitOnJobTermination" yaml:"skipWaitOnJobTermination"` // The subnetwork to which VMs will be assigned. Should be of the form "regions/REGION/subnetworks/SUBNETWORK". // // Docs at Terraform Registry: {@link https://registry.terraform.io/providers/hashicorp/google/4.79.0/docs/resources/dataflow_job#subnetwork DataflowJob#subnetwork} Subnetwork *string `field:"optional" json:"subnetwork" yaml:"subnetwork"` // timeouts block. // // Docs at Terraform Registry: {@link https://registry.terraform.io/providers/hashicorp/google/4.79.0/docs/resources/dataflow_job#timeouts DataflowJob#timeouts} Timeouts *DataflowJobTimeouts `field:"optional" json:"timeouts" yaml:"timeouts"` // Only applicable when updating a pipeline. // // Map of transform name prefixes of the job to be replaced with the corresponding name prefixes of the new job. // // Docs at Terraform Registry: {@link https://registry.terraform.io/providers/hashicorp/google/4.79.0/docs/resources/dataflow_job#transform_name_mapping DataflowJob#transform_name_mapping} TransformNameMapping *map[string]*string `field:"optional" json:"transformNameMapping" yaml:"transformNameMapping"` // The zone in which the created job should run. If it is not provided, the provider zone is used. // // Docs at Terraform Registry: {@link https://registry.terraform.io/providers/hashicorp/google/4.79.0/docs/resources/dataflow_job#zone DataflowJob#zone} Zone *string `field:"optional" json:"zone" yaml:"zone"` }
type DataflowJobTimeouts ¶
type DataflowJobTimeouts struct { // Docs at Terraform Registry: {@link https://registry.terraform.io/providers/hashicorp/google/4.79.0/docs/resources/dataflow_job#update DataflowJob#update}. Update *string `field:"optional" json:"update" yaml:"update"` }
type DataflowJobTimeoutsOutputReference ¶
type DataflowJobTimeoutsOutputReference interface { cdktf.ComplexObject // the index of the complex object in a list. // Experimental. ComplexObjectIndex() interface{} // Experimental. SetComplexObjectIndex(val interface{}) // set to true if this item is from inside a set and needs tolist() for accessing it set to "0" for single list items. // Experimental. ComplexObjectIsFromSet() *bool // Experimental. SetComplexObjectIsFromSet(val *bool) // The creation stack of this resolvable which will be appended to errors thrown during resolution. // // If this returns an empty array the stack will not be attached. // Experimental. CreationStack() *[]*string // Experimental. Fqn() *string InternalValue() interface{} SetInternalValue(val interface{}) // Experimental. TerraformAttribute() *string // Experimental. SetTerraformAttribute(val *string) // Experimental. TerraformResource() cdktf.IInterpolatingParent // Experimental. SetTerraformResource(val cdktf.IInterpolatingParent) Update() *string SetUpdate(val *string) UpdateInput() *string // Experimental. ComputeFqn() *string // Experimental. GetAnyMapAttribute(terraformAttribute *string) *map[string]interface{} // Experimental. GetBooleanAttribute(terraformAttribute *string) cdktf.IResolvable // Experimental. GetBooleanMapAttribute(terraformAttribute *string) *map[string]*bool // Experimental. GetListAttribute(terraformAttribute *string) *[]*string // Experimental. GetNumberAttribute(terraformAttribute *string) *float64 // Experimental. GetNumberListAttribute(terraformAttribute *string) *[]*float64 // Experimental. GetNumberMapAttribute(terraformAttribute *string) *map[string]*float64 // Experimental. GetStringAttribute(terraformAttribute *string) *string // Experimental. GetStringMapAttribute(terraformAttribute *string) *map[string]*string // Experimental. InterpolationAsList() cdktf.IResolvable // Experimental. InterpolationForAttribute(property *string) cdktf.IResolvable ResetUpdate() // Produce the Token's value at resolution time. // Experimental. Resolve(_context cdktf.IResolveContext) interface{} // Return a string representation of this resolvable object. // // Returns a reversible string representation. // Experimental. ToString() *string }
func NewDataflowJobTimeoutsOutputReference ¶
func NewDataflowJobTimeoutsOutputReference(terraformResource cdktf.IInterpolatingParent, terraformAttribute *string) DataflowJobTimeoutsOutputReference