Documentation ¶
Index ¶
- func DataDatabricksSparkVersion_IsConstruct(x interface{}) *bool
- func DataDatabricksSparkVersion_TfResourceType() *string
- func NewDataDatabricksSparkVersion_Override(d DataDatabricksSparkVersion, scope constructs.Construct, id *string, ...)
- type DataDatabricksSparkVersion
- type DataDatabricksSparkVersionConfig
Constants ¶
This section is empty.
Variables ¶
This section is empty.
Functions ¶
func DataDatabricksSparkVersion_IsConstruct ¶
func DataDatabricksSparkVersion_IsConstruct(x interface{}) *bool
Checks if `x` is a construct.
Use this method instead of `instanceof` to properly detect `Construct` instances, even when the construct library is symlinked.
Explanation: in JavaScript, multiple copies of the `constructs` library on disk are seen as independent, completely different libraries. As a consequence, the class `Construct` in each copy of the `constructs` library is seen as a different class, and an instance of one class will not test as `instanceof` the other class. `npm install` will not create installations like this, but users may manually symlink construct libraries together or use a monorepo tool: in those cases, multiple copies of the `constructs` library can be accidentally installed, and `instanceof` will behave unpredictably. It is safest to avoid using `instanceof`, and using this type-testing method instead.
Returns: true if `x` is an object created from a class which extends `Construct`.
func DataDatabricksSparkVersion_TfResourceType ¶
func DataDatabricksSparkVersion_TfResourceType() *string
func NewDataDatabricksSparkVersion_Override ¶
func NewDataDatabricksSparkVersion_Override(d DataDatabricksSparkVersion, scope constructs.Construct, id *string, config *DataDatabricksSparkVersionConfig)
Create a new {@link https://www.terraform.io/docs/providers/databricks/d/spark_version databricks_spark_version} Data Source.
Types ¶
type DataDatabricksSparkVersion ¶
type DataDatabricksSparkVersion interface { cdktf.TerraformDataSource Beta() interface{} SetBeta(val interface{}) BetaInput() interface{} // Experimental. CdktfStack() cdktf.TerraformStack // Experimental. ConstructNodeMetadata() *map[string]interface{} // Experimental. Count() *float64 // Experimental. SetCount(val *float64) // Experimental. DependsOn() *[]*string // Experimental. SetDependsOn(val *[]*string) // Experimental. ForEach() cdktf.ITerraformIterator // Experimental. SetForEach(val cdktf.ITerraformIterator) // Experimental. Fqn() *string // Experimental. FriendlyUniqueId() *string Genomics() interface{} SetGenomics(val interface{}) GenomicsInput() interface{} Gpu() interface{} SetGpu(val interface{}) GpuInput() interface{} Graviton() interface{} SetGraviton(val interface{}) GravitonInput() interface{} Id() *string SetId(val *string) IdInput() *string Latest() interface{} SetLatest(val interface{}) LatestInput() interface{} // Experimental. Lifecycle() *cdktf.TerraformResourceLifecycle // Experimental. SetLifecycle(val *cdktf.TerraformResourceLifecycle) LongTermSupport() interface{} SetLongTermSupport(val interface{}) LongTermSupportInput() interface{} Ml() interface{} SetMl(val interface{}) MlInput() interface{} // The tree node. Node() constructs.Node Photon() interface{} SetPhoton(val interface{}) PhotonInput() interface{} // Experimental. Provider() cdktf.TerraformProvider // Experimental. SetProvider(val cdktf.TerraformProvider) // Experimental. RawOverrides() interface{} Scala() *string SetScala(val *string) ScalaInput() *string SparkVersion() *string SetSparkVersion(val *string) SparkVersionInput() *string // Experimental. TerraformGeneratorMetadata() *cdktf.TerraformProviderGeneratorMetadata // Experimental. TerraformMetaArguments() *map[string]interface{} // Experimental. TerraformResourceType() *string // Experimental. AddOverride(path *string, value interface{}) // Experimental. GetAnyMapAttribute(terraformAttribute *string) *map[string]interface{} // Experimental. GetBooleanAttribute(terraformAttribute *string) cdktf.IResolvable // Experimental. GetBooleanMapAttribute(terraformAttribute *string) *map[string]*bool // Experimental. GetListAttribute(terraformAttribute *string) *[]*string // Experimental. GetNumberAttribute(terraformAttribute *string) *float64 // Experimental. GetNumberListAttribute(terraformAttribute *string) *[]*float64 // Experimental. GetNumberMapAttribute(terraformAttribute *string) *map[string]*float64 // Experimental. GetStringAttribute(terraformAttribute *string) *string // Experimental. GetStringMapAttribute(terraformAttribute *string) *map[string]*string // Experimental. InterpolationForAttribute(terraformAttribute *string) cdktf.IResolvable // Overrides the auto-generated logical ID with a specific ID. // Experimental. OverrideLogicalId(newLogicalId *string) ResetBeta() ResetGenomics() ResetGpu() ResetGraviton() ResetId() ResetLatest() ResetLongTermSupport() ResetMl() // Resets a previously passed logical Id to use the auto-generated logical id again. // Experimental. ResetOverrideLogicalId() ResetPhoton() ResetScala() ResetSparkVersion() SynthesizeAttributes() *map[string]interface{} // Experimental. ToMetadata() interface{} // Returns a string representation of this construct. ToString() *string // Adds this resource to the terraform JSON output. // Experimental. ToTerraform() interface{} }
Represents a {@link https://www.terraform.io/docs/providers/databricks/d/spark_version databricks_spark_version}.
func NewDataDatabricksSparkVersion ¶
func NewDataDatabricksSparkVersion(scope constructs.Construct, id *string, config *DataDatabricksSparkVersionConfig) DataDatabricksSparkVersion
Create a new {@link https://www.terraform.io/docs/providers/databricks/d/spark_version databricks_spark_version} Data Source.
type DataDatabricksSparkVersionConfig ¶
type DataDatabricksSparkVersionConfig struct { // Experimental. Connection interface{} `field:"optional" json:"connection" yaml:"connection"` // Experimental. Count *float64 `field:"optional" json:"count" yaml:"count"` // Experimental. DependsOn *[]cdktf.ITerraformDependable `field:"optional" json:"dependsOn" yaml:"dependsOn"` // Experimental. ForEach cdktf.ITerraformIterator `field:"optional" json:"forEach" yaml:"forEach"` // Experimental. Lifecycle *cdktf.TerraformResourceLifecycle `field:"optional" json:"lifecycle" yaml:"lifecycle"` // Experimental. Provider cdktf.TerraformProvider `field:"optional" json:"provider" yaml:"provider"` // Experimental. Provisioners *[]interface{} `field:"optional" json:"provisioners" yaml:"provisioners"` // Docs at Terraform Registry: {@link https://www.terraform.io/docs/providers/databricks/d/spark_version#beta DataDatabricksSparkVersion#beta}. Beta interface{} `field:"optional" json:"beta" yaml:"beta"` // Docs at Terraform Registry: {@link https://www.terraform.io/docs/providers/databricks/d/spark_version#genomics DataDatabricksSparkVersion#genomics}. Genomics interface{} `field:"optional" json:"genomics" yaml:"genomics"` // Docs at Terraform Registry: {@link https://www.terraform.io/docs/providers/databricks/d/spark_version#gpu DataDatabricksSparkVersion#gpu}. Gpu interface{} `field:"optional" json:"gpu" yaml:"gpu"` // Docs at Terraform Registry: {@link https://www.terraform.io/docs/providers/databricks/d/spark_version#graviton DataDatabricksSparkVersion#graviton}. Graviton interface{} `field:"optional" json:"graviton" yaml:"graviton"` // Docs at Terraform Registry: {@link https://www.terraform.io/docs/providers/databricks/d/spark_version#id DataDatabricksSparkVersion#id}. // // Please be aware that the id field is automatically added to all resources in Terraform providers using a Terraform provider SDK version below 2. // If you experience problems setting this value it might not be settable. Please take a look at the provider documentation to ensure it should be settable. Id *string `field:"optional" json:"id" yaml:"id"` // Docs at Terraform Registry: {@link https://www.terraform.io/docs/providers/databricks/d/spark_version#latest DataDatabricksSparkVersion#latest}. Latest interface{} `field:"optional" json:"latest" yaml:"latest"` // Docs at Terraform Registry: {@link https://www.terraform.io/docs/providers/databricks/d/spark_version#long_term_support DataDatabricksSparkVersion#long_term_support}. LongTermSupport interface{} `field:"optional" json:"longTermSupport" yaml:"longTermSupport"` // Docs at Terraform Registry: {@link https://www.terraform.io/docs/providers/databricks/d/spark_version#ml DataDatabricksSparkVersion#ml}. Ml interface{} `field:"optional" json:"ml" yaml:"ml"` // Docs at Terraform Registry: {@link https://www.terraform.io/docs/providers/databricks/d/spark_version#photon DataDatabricksSparkVersion#photon}. Photon interface{} `field:"optional" json:"photon" yaml:"photon"` // Docs at Terraform Registry: {@link https://www.terraform.io/docs/providers/databricks/d/spark_version#scala DataDatabricksSparkVersion#scala}. Scala *string `field:"optional" json:"scala" yaml:"scala"` // Docs at Terraform Registry: {@link https://www.terraform.io/docs/providers/databricks/d/spark_version#spark_version DataDatabricksSparkVersion#spark_version}. SparkVersion *string `field:"optional" json:"sparkVersion" yaml:"sparkVersion"` }