bigquery

package
v3.5.0 Latest Latest
Warning

This package is not in the latest version of its module.

Go to latest
Published: May 14, 2020 License: Apache-2.0 Imports: 4 Imported by: 0

Documentation

Index

Constants

This section is empty.

Variables

This section is empty.

Functions

This section is empty.

Types

type AppProfile

type AppProfile struct {
	pulumi.CustomResourceState

	// The unique name of the app profile in the form `[_a-zA-Z0-9][-_.a-zA-Z0-9]*`.
	AppProfileId pulumi.StringOutput `pulumi:"appProfileId"`
	// Long form description of the use case for this app profile.
	Description pulumi.StringPtrOutput `pulumi:"description"`
	// If true, ignore safety checks when deleting/updating the app profile.
	IgnoreWarnings pulumi.BoolPtrOutput `pulumi:"ignoreWarnings"`
	// The name of the instance to create the app profile within.
	Instance pulumi.StringPtrOutput `pulumi:"instance"`
	// If true, read/write requests are routed to the nearest cluster in the instance, and will fail over to the nearest cluster that is available
	// in the event of transient errors or delays. Clusters in a region are considered equidistant. Choosing this option sacrifices read-your-writes
	// consistency to improve availability.
	MultiClusterRoutingUseAny pulumi.BoolPtrOutput `pulumi:"multiClusterRoutingUseAny"`
	// The unique name of the requested app profile. Values are of the form
	// 'projects/<project>/instances/<instance>/appProfiles/<appProfileId>'.
	Name pulumi.StringOutput `pulumi:"name"`
	// The ID of the project in which the resource belongs.
	// If it is not provided, the provider project is used.
	Project pulumi.StringOutput `pulumi:"project"`
	// Use a single-cluster routing policy.  Structure is documented below.
	SingleClusterRouting AppProfileSingleClusterRoutingPtrOutput `pulumi:"singleClusterRouting"`
}

App profile is a configuration object describing how Cloud Bigtable should treat traffic from a particular end user application.

func GetAppProfile

func GetAppProfile(ctx *pulumi.Context,
	name string, id pulumi.IDInput, state *AppProfileState, opts ...pulumi.ResourceOption) (*AppProfile, error)

GetAppProfile gets an existing AppProfile resource's state with the given name, ID, and optional state properties that are used to uniquely qualify the lookup (nil if not required).

func NewAppProfile

func NewAppProfile(ctx *pulumi.Context,
	name string, args *AppProfileArgs, opts ...pulumi.ResourceOption) (*AppProfile, error)

NewAppProfile registers a new resource with the given unique name, arguments, and options.

type AppProfileArgs

type AppProfileArgs struct {
	// The unique name of the app profile in the form `[_a-zA-Z0-9][-_.a-zA-Z0-9]*`.
	AppProfileId pulumi.StringInput
	// Long form description of the use case for this app profile.
	Description pulumi.StringPtrInput
	// If true, ignore safety checks when deleting/updating the app profile.
	IgnoreWarnings pulumi.BoolPtrInput
	// The name of the instance to create the app profile within.
	Instance pulumi.StringPtrInput
	// If true, read/write requests are routed to the nearest cluster in the instance, and will fail over to the nearest cluster that is available
	// in the event of transient errors or delays. Clusters in a region are considered equidistant. Choosing this option sacrifices read-your-writes
	// consistency to improve availability.
	MultiClusterRoutingUseAny pulumi.BoolPtrInput
	// The ID of the project in which the resource belongs.
	// If it is not provided, the provider project is used.
	Project pulumi.StringPtrInput
	// Use a single-cluster routing policy.  Structure is documented below.
	SingleClusterRouting AppProfileSingleClusterRoutingPtrInput
}

The set of arguments for constructing a AppProfile resource.

func (AppProfileArgs) ElementType

func (AppProfileArgs) ElementType() reflect.Type

type AppProfileSingleClusterRouting

type AppProfileSingleClusterRouting struct {
	// If true, CheckAndMutateRow and ReadModifyWriteRow requests are allowed by this app profile.
	// It is unsafe to send these requests to the same table/row/column in multiple clusters.
	AllowTransactionalWrites *bool `pulumi:"allowTransactionalWrites"`
	// The cluster to which read/write requests should be routed.
	ClusterId string `pulumi:"clusterId"`
}

type AppProfileSingleClusterRoutingArgs

type AppProfileSingleClusterRoutingArgs struct {
	// If true, CheckAndMutateRow and ReadModifyWriteRow requests are allowed by this app profile.
	// It is unsafe to send these requests to the same table/row/column in multiple clusters.
	AllowTransactionalWrites pulumi.BoolPtrInput `pulumi:"allowTransactionalWrites"`
	// The cluster to which read/write requests should be routed.
	ClusterId pulumi.StringInput `pulumi:"clusterId"`
}

func (AppProfileSingleClusterRoutingArgs) ElementType

func (AppProfileSingleClusterRoutingArgs) ToAppProfileSingleClusterRoutingOutput

func (i AppProfileSingleClusterRoutingArgs) ToAppProfileSingleClusterRoutingOutput() AppProfileSingleClusterRoutingOutput

func (AppProfileSingleClusterRoutingArgs) ToAppProfileSingleClusterRoutingOutputWithContext

func (i AppProfileSingleClusterRoutingArgs) ToAppProfileSingleClusterRoutingOutputWithContext(ctx context.Context) AppProfileSingleClusterRoutingOutput

func (AppProfileSingleClusterRoutingArgs) ToAppProfileSingleClusterRoutingPtrOutput

func (i AppProfileSingleClusterRoutingArgs) ToAppProfileSingleClusterRoutingPtrOutput() AppProfileSingleClusterRoutingPtrOutput

func (AppProfileSingleClusterRoutingArgs) ToAppProfileSingleClusterRoutingPtrOutputWithContext

func (i AppProfileSingleClusterRoutingArgs) ToAppProfileSingleClusterRoutingPtrOutputWithContext(ctx context.Context) AppProfileSingleClusterRoutingPtrOutput

type AppProfileSingleClusterRoutingInput

type AppProfileSingleClusterRoutingInput interface {
	pulumi.Input

	ToAppProfileSingleClusterRoutingOutput() AppProfileSingleClusterRoutingOutput
	ToAppProfileSingleClusterRoutingOutputWithContext(context.Context) AppProfileSingleClusterRoutingOutput
}

AppProfileSingleClusterRoutingInput is an input type that accepts AppProfileSingleClusterRoutingArgs and AppProfileSingleClusterRoutingOutput values. You can construct a concrete instance of `AppProfileSingleClusterRoutingInput` via:

AppProfileSingleClusterRoutingArgs{...}

type AppProfileSingleClusterRoutingOutput

type AppProfileSingleClusterRoutingOutput struct{ *pulumi.OutputState }

func (AppProfileSingleClusterRoutingOutput) AllowTransactionalWrites

func (o AppProfileSingleClusterRoutingOutput) AllowTransactionalWrites() pulumi.BoolPtrOutput

If true, CheckAndMutateRow and ReadModifyWriteRow requests are allowed by this app profile. It is unsafe to send these requests to the same table/row/column in multiple clusters.

func (AppProfileSingleClusterRoutingOutput) ClusterId

The cluster to which read/write requests should be routed.

func (AppProfileSingleClusterRoutingOutput) ElementType

func (AppProfileSingleClusterRoutingOutput) ToAppProfileSingleClusterRoutingOutput

func (o AppProfileSingleClusterRoutingOutput) ToAppProfileSingleClusterRoutingOutput() AppProfileSingleClusterRoutingOutput

func (AppProfileSingleClusterRoutingOutput) ToAppProfileSingleClusterRoutingOutputWithContext

func (o AppProfileSingleClusterRoutingOutput) ToAppProfileSingleClusterRoutingOutputWithContext(ctx context.Context) AppProfileSingleClusterRoutingOutput

func (AppProfileSingleClusterRoutingOutput) ToAppProfileSingleClusterRoutingPtrOutput

func (o AppProfileSingleClusterRoutingOutput) ToAppProfileSingleClusterRoutingPtrOutput() AppProfileSingleClusterRoutingPtrOutput

func (AppProfileSingleClusterRoutingOutput) ToAppProfileSingleClusterRoutingPtrOutputWithContext

func (o AppProfileSingleClusterRoutingOutput) ToAppProfileSingleClusterRoutingPtrOutputWithContext(ctx context.Context) AppProfileSingleClusterRoutingPtrOutput

type AppProfileSingleClusterRoutingPtrInput

type AppProfileSingleClusterRoutingPtrInput interface {
	pulumi.Input

	ToAppProfileSingleClusterRoutingPtrOutput() AppProfileSingleClusterRoutingPtrOutput
	ToAppProfileSingleClusterRoutingPtrOutputWithContext(context.Context) AppProfileSingleClusterRoutingPtrOutput
}

AppProfileSingleClusterRoutingPtrInput is an input type that accepts AppProfileSingleClusterRoutingArgs, AppProfileSingleClusterRoutingPtr and AppProfileSingleClusterRoutingPtrOutput values. You can construct a concrete instance of `AppProfileSingleClusterRoutingPtrInput` via:

		 AppProfileSingleClusterRoutingArgs{...}

 or:

		 nil

type AppProfileSingleClusterRoutingPtrOutput

type AppProfileSingleClusterRoutingPtrOutput struct{ *pulumi.OutputState }

func (AppProfileSingleClusterRoutingPtrOutput) AllowTransactionalWrites

func (o AppProfileSingleClusterRoutingPtrOutput) AllowTransactionalWrites() pulumi.BoolPtrOutput

If true, CheckAndMutateRow and ReadModifyWriteRow requests are allowed by this app profile. It is unsafe to send these requests to the same table/row/column in multiple clusters.

func (AppProfileSingleClusterRoutingPtrOutput) ClusterId

The cluster to which read/write requests should be routed.

func (AppProfileSingleClusterRoutingPtrOutput) Elem

func (AppProfileSingleClusterRoutingPtrOutput) ElementType

func (AppProfileSingleClusterRoutingPtrOutput) ToAppProfileSingleClusterRoutingPtrOutput

func (o AppProfileSingleClusterRoutingPtrOutput) ToAppProfileSingleClusterRoutingPtrOutput() AppProfileSingleClusterRoutingPtrOutput

func (AppProfileSingleClusterRoutingPtrOutput) ToAppProfileSingleClusterRoutingPtrOutputWithContext

func (o AppProfileSingleClusterRoutingPtrOutput) ToAppProfileSingleClusterRoutingPtrOutputWithContext(ctx context.Context) AppProfileSingleClusterRoutingPtrOutput

type AppProfileState

type AppProfileState struct {
	// The unique name of the app profile in the form `[_a-zA-Z0-9][-_.a-zA-Z0-9]*`.
	AppProfileId pulumi.StringPtrInput
	// Long form description of the use case for this app profile.
	Description pulumi.StringPtrInput
	// If true, ignore safety checks when deleting/updating the app profile.
	IgnoreWarnings pulumi.BoolPtrInput
	// The name of the instance to create the app profile within.
	Instance pulumi.StringPtrInput
	// If true, read/write requests are routed to the nearest cluster in the instance, and will fail over to the nearest cluster that is available
	// in the event of transient errors or delays. Clusters in a region are considered equidistant. Choosing this option sacrifices read-your-writes
	// consistency to improve availability.
	MultiClusterRoutingUseAny pulumi.BoolPtrInput
	// The unique name of the requested app profile. Values are of the form
	// 'projects/<project>/instances/<instance>/appProfiles/<appProfileId>'.
	Name pulumi.StringPtrInput
	// The ID of the project in which the resource belongs.
	// If it is not provided, the provider project is used.
	Project pulumi.StringPtrInput
	// Use a single-cluster routing policy.  Structure is documented below.
	SingleClusterRouting AppProfileSingleClusterRoutingPtrInput
}

func (AppProfileState) ElementType

func (AppProfileState) ElementType() reflect.Type

type Connection added in v3.4.0

type Connection struct {
	pulumi.CustomResourceState

	// Cloud SQL properties.  Structure is documented below.
	CloudSql ConnectionCloudSqlOutput `pulumi:"cloudSql"`
	// Optional connection id that should be assigned to the created connection.
	ConnectionId pulumi.StringPtrOutput `pulumi:"connectionId"`
	// A descriptive description for the connection
	Description pulumi.StringPtrOutput `pulumi:"description"`
	// A descriptive name for the connection
	FriendlyName pulumi.StringPtrOutput `pulumi:"friendlyName"`
	// True if the connection has credential assigned.
	HasCredential pulumi.BoolOutput `pulumi:"hasCredential"`
	// The geographic location where the connection should reside.
	// Cloud SQL instance must be in the same location as the connection
	// with following exceptions: Cloud SQL us-central1 maps to BigQuery US, Cloud SQL europe-west1 maps to BigQuery EU.
	// Examples: US, EU, asia-northeast1, us-central1, europe-west1. The default value is US.
	Location pulumi.StringPtrOutput `pulumi:"location"`
	// The resource name of the connection in the form of:
	// "projects/{project_id}/locations/{location_id}/connections/{connectionId}"
	Name pulumi.StringOutput `pulumi:"name"`
	// The ID of the project in which the resource belongs.
	// If it is not provided, the provider project is used.
	Project pulumi.StringOutput `pulumi:"project"`
}

A connection allows BigQuery connections to external data sources..

To get more information about Connection, see:

* [API documentation](https://cloud.google.com/bigquery/docs/reference/bigqueryconnection/rest/v1beta1/projects.locations.connections/create) * How-to Guides

func GetConnection added in v3.4.0

func GetConnection(ctx *pulumi.Context,
	name string, id pulumi.IDInput, state *ConnectionState, opts ...pulumi.ResourceOption) (*Connection, error)

GetConnection gets an existing Connection resource's state with the given name, ID, and optional state properties that are used to uniquely qualify the lookup (nil if not required).

func NewConnection added in v3.4.0

func NewConnection(ctx *pulumi.Context,
	name string, args *ConnectionArgs, opts ...pulumi.ResourceOption) (*Connection, error)

NewConnection registers a new resource with the given unique name, arguments, and options.

type ConnectionArgs added in v3.4.0

type ConnectionArgs struct {
	// Cloud SQL properties.  Structure is documented below.
	CloudSql ConnectionCloudSqlInput
	// Optional connection id that should be assigned to the created connection.
	ConnectionId pulumi.StringPtrInput
	// A descriptive description for the connection
	Description pulumi.StringPtrInput
	// A descriptive name for the connection
	FriendlyName pulumi.StringPtrInput
	// The geographic location where the connection should reside.
	// Cloud SQL instance must be in the same location as the connection
	// with following exceptions: Cloud SQL us-central1 maps to BigQuery US, Cloud SQL europe-west1 maps to BigQuery EU.
	// Examples: US, EU, asia-northeast1, us-central1, europe-west1. The default value is US.
	Location pulumi.StringPtrInput
	// The ID of the project in which the resource belongs.
	// If it is not provided, the provider project is used.
	Project pulumi.StringPtrInput
}

The set of arguments for constructing a Connection resource.

func (ConnectionArgs) ElementType added in v3.4.0

func (ConnectionArgs) ElementType() reflect.Type

type ConnectionCloudSql added in v3.4.0

type ConnectionCloudSql struct {
	// Database name.
	Database string `pulumi:"database"`
	// Cloud SQL instance ID in the form project:location:instance.
	InstanceId string `pulumi:"instanceId"`
	// Type of the Cloud SQL database.
	Type string `pulumi:"type"`
}

type ConnectionCloudSqlArgs added in v3.4.0

type ConnectionCloudSqlArgs struct {
	// Database name.
	Database pulumi.StringInput `pulumi:"database"`
	// Cloud SQL instance ID in the form project:location:instance.
	InstanceId pulumi.StringInput `pulumi:"instanceId"`
	// Type of the Cloud SQL database.
	Type pulumi.StringInput `pulumi:"type"`
}

func (ConnectionCloudSqlArgs) ElementType added in v3.4.0

func (ConnectionCloudSqlArgs) ElementType() reflect.Type

func (ConnectionCloudSqlArgs) ToConnectionCloudSqlOutput added in v3.4.0

func (i ConnectionCloudSqlArgs) ToConnectionCloudSqlOutput() ConnectionCloudSqlOutput

func (ConnectionCloudSqlArgs) ToConnectionCloudSqlOutputWithContext added in v3.4.0

func (i ConnectionCloudSqlArgs) ToConnectionCloudSqlOutputWithContext(ctx context.Context) ConnectionCloudSqlOutput

func (ConnectionCloudSqlArgs) ToConnectionCloudSqlPtrOutput added in v3.4.0

func (i ConnectionCloudSqlArgs) ToConnectionCloudSqlPtrOutput() ConnectionCloudSqlPtrOutput

func (ConnectionCloudSqlArgs) ToConnectionCloudSqlPtrOutputWithContext added in v3.4.0

func (i ConnectionCloudSqlArgs) ToConnectionCloudSqlPtrOutputWithContext(ctx context.Context) ConnectionCloudSqlPtrOutput

type ConnectionCloudSqlInput added in v3.4.0

type ConnectionCloudSqlInput interface {
	pulumi.Input

	ToConnectionCloudSqlOutput() ConnectionCloudSqlOutput
	ToConnectionCloudSqlOutputWithContext(context.Context) ConnectionCloudSqlOutput
}

ConnectionCloudSqlInput is an input type that accepts ConnectionCloudSqlArgs and ConnectionCloudSqlOutput values. You can construct a concrete instance of `ConnectionCloudSqlInput` via:

ConnectionCloudSqlArgs{...}

type ConnectionCloudSqlOutput added in v3.4.0

type ConnectionCloudSqlOutput struct{ *pulumi.OutputState }

func (ConnectionCloudSqlOutput) Database added in v3.4.0

Database name.

func (ConnectionCloudSqlOutput) ElementType added in v3.4.0

func (ConnectionCloudSqlOutput) ElementType() reflect.Type

func (ConnectionCloudSqlOutput) InstanceId added in v3.4.0

Cloud SQL instance ID in the form project:location:instance.

func (ConnectionCloudSqlOutput) ToConnectionCloudSqlOutput added in v3.4.0

func (o ConnectionCloudSqlOutput) ToConnectionCloudSqlOutput() ConnectionCloudSqlOutput

func (ConnectionCloudSqlOutput) ToConnectionCloudSqlOutputWithContext added in v3.4.0

func (o ConnectionCloudSqlOutput) ToConnectionCloudSqlOutputWithContext(ctx context.Context) ConnectionCloudSqlOutput

func (ConnectionCloudSqlOutput) ToConnectionCloudSqlPtrOutput added in v3.4.0

func (o ConnectionCloudSqlOutput) ToConnectionCloudSqlPtrOutput() ConnectionCloudSqlPtrOutput

func (ConnectionCloudSqlOutput) ToConnectionCloudSqlPtrOutputWithContext added in v3.4.0

func (o ConnectionCloudSqlOutput) ToConnectionCloudSqlPtrOutputWithContext(ctx context.Context) ConnectionCloudSqlPtrOutput

func (ConnectionCloudSqlOutput) Type added in v3.4.0

Type of the Cloud SQL database.

type ConnectionCloudSqlPtrInput added in v3.4.0

type ConnectionCloudSqlPtrInput interface {
	pulumi.Input

	ToConnectionCloudSqlPtrOutput() ConnectionCloudSqlPtrOutput
	ToConnectionCloudSqlPtrOutputWithContext(context.Context) ConnectionCloudSqlPtrOutput
}

ConnectionCloudSqlPtrInput is an input type that accepts ConnectionCloudSqlArgs, ConnectionCloudSqlPtr and ConnectionCloudSqlPtrOutput values. You can construct a concrete instance of `ConnectionCloudSqlPtrInput` via:

		 ConnectionCloudSqlArgs{...}

 or:

		 nil

func ConnectionCloudSqlPtr added in v3.4.0

func ConnectionCloudSqlPtr(v *ConnectionCloudSqlArgs) ConnectionCloudSqlPtrInput

type ConnectionCloudSqlPtrOutput added in v3.4.0

type ConnectionCloudSqlPtrOutput struct{ *pulumi.OutputState }

func (ConnectionCloudSqlPtrOutput) Database added in v3.4.0

Database name.

func (ConnectionCloudSqlPtrOutput) Elem added in v3.4.0

func (ConnectionCloudSqlPtrOutput) ElementType added in v3.4.0

func (ConnectionCloudSqlPtrOutput) InstanceId added in v3.4.0

Cloud SQL instance ID in the form project:location:instance.

func (ConnectionCloudSqlPtrOutput) ToConnectionCloudSqlPtrOutput added in v3.4.0

func (o ConnectionCloudSqlPtrOutput) ToConnectionCloudSqlPtrOutput() ConnectionCloudSqlPtrOutput

func (ConnectionCloudSqlPtrOutput) ToConnectionCloudSqlPtrOutputWithContext added in v3.4.0

func (o ConnectionCloudSqlPtrOutput) ToConnectionCloudSqlPtrOutputWithContext(ctx context.Context) ConnectionCloudSqlPtrOutput

func (ConnectionCloudSqlPtrOutput) Type added in v3.4.0

Type of the Cloud SQL database.

type ConnectionState added in v3.4.0

type ConnectionState struct {
	// Cloud SQL properties.  Structure is documented below.
	CloudSql ConnectionCloudSqlPtrInput
	// Optional connection id that should be assigned to the created connection.
	ConnectionId pulumi.StringPtrInput
	// A descriptive description for the connection
	Description pulumi.StringPtrInput
	// A descriptive name for the connection
	FriendlyName pulumi.StringPtrInput
	// True if the connection has credential assigned.
	HasCredential pulumi.BoolPtrInput
	// The geographic location where the connection should reside.
	// Cloud SQL instance must be in the same location as the connection
	// with following exceptions: Cloud SQL us-central1 maps to BigQuery US, Cloud SQL europe-west1 maps to BigQuery EU.
	// Examples: US, EU, asia-northeast1, us-central1, europe-west1. The default value is US.
	Location pulumi.StringPtrInput
	// The resource name of the connection in the form of:
	// "projects/{project_id}/locations/{location_id}/connections/{connectionId}"
	Name pulumi.StringPtrInput
	// The ID of the project in which the resource belongs.
	// If it is not provided, the provider project is used.
	Project pulumi.StringPtrInput
}

func (ConnectionState) ElementType added in v3.4.0

func (ConnectionState) ElementType() reflect.Type

type DataTransferConfig

type DataTransferConfig struct {
	pulumi.CustomResourceState

	// The number of days to look back to automatically refresh the data.
	// For example, if dataRefreshWindowDays = 10, then every day BigQuery
	// reingests data for [today-10, today-1], rather than ingesting data for
	// just [today-1]. Only valid if the data source supports the feature.
	// Set the value to 0 to use the default value.
	DataRefreshWindowDays pulumi.IntPtrOutput `pulumi:"dataRefreshWindowDays"`
	// The data source id. Cannot be changed once the transfer config is created.
	DataSourceId pulumi.StringOutput `pulumi:"dataSourceId"`
	// The BigQuery target dataset id.
	DestinationDatasetId pulumi.StringOutput `pulumi:"destinationDatasetId"`
	// When set to true, no runs are scheduled for a given transfer.
	Disabled pulumi.BoolPtrOutput `pulumi:"disabled"`
	// The user specified display name for the transfer config.
	DisplayName pulumi.StringOutput `pulumi:"displayName"`
	// The geographic location where the transfer config should reside.
	// Examples: US, EU, asia-northeast1. The default value is US.
	Location pulumi.StringPtrOutput `pulumi:"location"`
	// The resource name of the transfer config. Transfer config names have the form
	// projects/{projectId}/locations/{location}/transferConfigs/{configId}. Where configId is usually a uuid, but this is not
	// required. The name is ignored when creating a transfer config.
	Name pulumi.StringOutput `pulumi:"name"`
	// These parameters are specific to each data source.
	Params pulumi.StringMapOutput `pulumi:"params"`
	// The ID of the project in which the resource belongs.
	// If it is not provided, the provider project is used.
	Project pulumi.StringOutput `pulumi:"project"`
	// Data transfer schedule. If the data source does not support a custom
	// schedule, this should be empty. If it is empty, the default value for
	// the data source will be used. The specified times are in UTC. Examples
	// of valid format: 1st,3rd monday of month 15:30, every wed,fri of jan,
	// jun 13:15, and first sunday of quarter 00:00. See more explanation
	// about the format here:
	// https://cloud.google.com/appengine/docs/flexible/python/scheduling-jobs-with-cron-yaml#the_schedule_format
	// NOTE: the granularity should be at least 8 hours, or less frequent.
	Schedule pulumi.StringPtrOutput `pulumi:"schedule"`
	// Optional service account name. If this field is set, transfer config will
	// be created with this service account credentials. It requires that
	// requesting user calling this API has permissions to act as this service account.
	ServiceAccountName pulumi.StringPtrOutput `pulumi:"serviceAccountName"`
}

Represents a data transfer configuration. A transfer configuration contains all metadata needed to perform a data transfer.

To get more information about Config, see:

* [API documentation](https://cloud.google.com/bigquery/docs/reference/datatransfer/rest/v1/projects.locations.transferConfigs/create) * How-to Guides

func GetDataTransferConfig

func GetDataTransferConfig(ctx *pulumi.Context,
	name string, id pulumi.IDInput, state *DataTransferConfigState, opts ...pulumi.ResourceOption) (*DataTransferConfig, error)

GetDataTransferConfig gets an existing DataTransferConfig resource's state with the given name, ID, and optional state properties that are used to uniquely qualify the lookup (nil if not required).

func NewDataTransferConfig

func NewDataTransferConfig(ctx *pulumi.Context,
	name string, args *DataTransferConfigArgs, opts ...pulumi.ResourceOption) (*DataTransferConfig, error)

NewDataTransferConfig registers a new resource with the given unique name, arguments, and options.

type DataTransferConfigArgs

type DataTransferConfigArgs struct {
	// The number of days to look back to automatically refresh the data.
	// For example, if dataRefreshWindowDays = 10, then every day BigQuery
	// reingests data for [today-10, today-1], rather than ingesting data for
	// just [today-1]. Only valid if the data source supports the feature.
	// Set the value to 0 to use the default value.
	DataRefreshWindowDays pulumi.IntPtrInput
	// The data source id. Cannot be changed once the transfer config is created.
	DataSourceId pulumi.StringInput
	// The BigQuery target dataset id.
	DestinationDatasetId pulumi.StringInput
	// When set to true, no runs are scheduled for a given transfer.
	Disabled pulumi.BoolPtrInput
	// The user specified display name for the transfer config.
	DisplayName pulumi.StringInput
	// The geographic location where the transfer config should reside.
	// Examples: US, EU, asia-northeast1. The default value is US.
	Location pulumi.StringPtrInput
	// These parameters are specific to each data source.
	Params pulumi.StringMapInput
	// The ID of the project in which the resource belongs.
	// If it is not provided, the provider project is used.
	Project pulumi.StringPtrInput
	// Data transfer schedule. If the data source does not support a custom
	// schedule, this should be empty. If it is empty, the default value for
	// the data source will be used. The specified times are in UTC. Examples
	// of valid format: 1st,3rd monday of month 15:30, every wed,fri of jan,
	// jun 13:15, and first sunday of quarter 00:00. See more explanation
	// about the format here:
	// https://cloud.google.com/appengine/docs/flexible/python/scheduling-jobs-with-cron-yaml#the_schedule_format
	// NOTE: the granularity should be at least 8 hours, or less frequent.
	Schedule pulumi.StringPtrInput
	// Optional service account name. If this field is set, transfer config will
	// be created with this service account credentials. It requires that
	// requesting user calling this API has permissions to act as this service account.
	ServiceAccountName pulumi.StringPtrInput
}

The set of arguments for constructing a DataTransferConfig resource.

func (DataTransferConfigArgs) ElementType

func (DataTransferConfigArgs) ElementType() reflect.Type

type DataTransferConfigState

type DataTransferConfigState struct {
	// The number of days to look back to automatically refresh the data.
	// For example, if dataRefreshWindowDays = 10, then every day BigQuery
	// reingests data for [today-10, today-1], rather than ingesting data for
	// just [today-1]. Only valid if the data source supports the feature.
	// Set the value to 0 to use the default value.
	DataRefreshWindowDays pulumi.IntPtrInput
	// The data source id. Cannot be changed once the transfer config is created.
	DataSourceId pulumi.StringPtrInput
	// The BigQuery target dataset id.
	DestinationDatasetId pulumi.StringPtrInput
	// When set to true, no runs are scheduled for a given transfer.
	Disabled pulumi.BoolPtrInput
	// The user specified display name for the transfer config.
	DisplayName pulumi.StringPtrInput
	// The geographic location where the transfer config should reside.
	// Examples: US, EU, asia-northeast1. The default value is US.
	Location pulumi.StringPtrInput
	// The resource name of the transfer config. Transfer config names have the form
	// projects/{projectId}/locations/{location}/transferConfigs/{configId}. Where configId is usually a uuid, but this is not
	// required. The name is ignored when creating a transfer config.
	Name pulumi.StringPtrInput
	// These parameters are specific to each data source.
	Params pulumi.StringMapInput
	// The ID of the project in which the resource belongs.
	// If it is not provided, the provider project is used.
	Project pulumi.StringPtrInput
	// Data transfer schedule. If the data source does not support a custom
	// schedule, this should be empty. If it is empty, the default value for
	// the data source will be used. The specified times are in UTC. Examples
	// of valid format: 1st,3rd monday of month 15:30, every wed,fri of jan,
	// jun 13:15, and first sunday of quarter 00:00. See more explanation
	// about the format here:
	// https://cloud.google.com/appengine/docs/flexible/python/scheduling-jobs-with-cron-yaml#the_schedule_format
	// NOTE: the granularity should be at least 8 hours, or less frequent.
	Schedule pulumi.StringPtrInput
	// Optional service account name. If this field is set, transfer config will
	// be created with this service account credentials. It requires that
	// requesting user calling this API has permissions to act as this service account.
	ServiceAccountName pulumi.StringPtrInput
}

func (DataTransferConfigState) ElementType

func (DataTransferConfigState) ElementType() reflect.Type

type Dataset

type Dataset struct {
	pulumi.CustomResourceState

	// An array of objects that define dataset access for one or more entities.  Structure is documented below.
	Accesses DatasetAccessTypeArrayOutput `pulumi:"accesses"`
	// The time when this dataset was created, in milliseconds since the epoch.
	CreationTime pulumi.IntOutput `pulumi:"creationTime"`
	// The ID of the dataset containing this table.
	DatasetId pulumi.StringOutput `pulumi:"datasetId"`
	// The default encryption key for all tables in the dataset. Once this property is set,
	// all newly-created partitioned tables in the dataset will have encryption key set to
	// this value, unless table creation request (or query) overrides the key.  Structure is documented below.
	DefaultEncryptionConfiguration DatasetDefaultEncryptionConfigurationPtrOutput `pulumi:"defaultEncryptionConfiguration"`
	// The default partition expiration for all partitioned tables in
	// the dataset, in milliseconds.
	DefaultPartitionExpirationMs pulumi.IntPtrOutput `pulumi:"defaultPartitionExpirationMs"`
	// The default lifetime of all tables in the dataset, in milliseconds.
	// The minimum value is 3600000 milliseconds (one hour).
	DefaultTableExpirationMs pulumi.IntPtrOutput `pulumi:"defaultTableExpirationMs"`
	// If set to `true`, delete all the tables in the
	// dataset when destroying the resource; otherwise,
	// destroying the resource will fail if tables are present.
	DeleteContentsOnDestroy pulumi.BoolPtrOutput `pulumi:"deleteContentsOnDestroy"`
	// A user-friendly description of the dataset
	Description pulumi.StringPtrOutput `pulumi:"description"`
	// A hash of the resource.
	Etag pulumi.StringOutput `pulumi:"etag"`
	// A descriptive name for the dataset
	FriendlyName pulumi.StringPtrOutput `pulumi:"friendlyName"`
	// The labels associated with this dataset. You can use these to
	// organize and group your datasets
	Labels pulumi.StringMapOutput `pulumi:"labels"`
	// The date when this dataset or any of its tables was last modified, in milliseconds since the epoch.
	LastModifiedTime pulumi.IntOutput `pulumi:"lastModifiedTime"`
	// The geographic location where the dataset should reside.
	// See [official docs](https://cloud.google.com/bigquery/docs/dataset-locations).
	Location pulumi.StringPtrOutput `pulumi:"location"`
	// The ID of the project in which the resource belongs.
	// If it is not provided, the provider project is used.
	Project pulumi.StringOutput `pulumi:"project"`
	// The URI of the created resource.
	SelfLink pulumi.StringOutput `pulumi:"selfLink"`
}

Datasets allow you to organize and control access to your tables.

To get more information about Dataset, see:

* [API documentation](https://cloud.google.com/bigquery/docs/reference/rest/v2/datasets) * How-to Guides

func GetDataset

func GetDataset(ctx *pulumi.Context,
	name string, id pulumi.IDInput, state *DatasetState, opts ...pulumi.ResourceOption) (*Dataset, error)

GetDataset gets an existing Dataset resource's state with the given name, ID, and optional state properties that are used to uniquely qualify the lookup (nil if not required).

func NewDataset

func NewDataset(ctx *pulumi.Context,
	name string, args *DatasetArgs, opts ...pulumi.ResourceOption) (*Dataset, error)

NewDataset registers a new resource with the given unique name, arguments, and options.

type DatasetAccess

type DatasetAccess struct {
	pulumi.CustomResourceState

	// The ID of the dataset containing this table.
	DatasetId pulumi.StringOutput `pulumi:"datasetId"`
	// A domain to grant access to. Any users signed in with the
	// domain specified will be granted the specified access
	Domain pulumi.StringPtrOutput `pulumi:"domain"`
	// An email address of a Google Group to grant access to.
	GroupByEmail pulumi.StringPtrOutput `pulumi:"groupByEmail"`
	// Some other type of member that appears in the IAM Policy but isn't a user,
	// group, domain, or special group. For example: `allUsers`
	IamMember pulumi.StringPtrOutput `pulumi:"iamMember"`
	// The ID of the project in which the resource belongs.
	// If it is not provided, the provider project is used.
	Project pulumi.StringOutput `pulumi:"project"`
	// Describes the rights granted to the user specified by the other
	// member of the access object. Primitive, Predefined and custom
	// roles are supported. Predefined roles that have equivalent
	// primitive roles are swapped by the API to their Primitive
	// counterparts, and will show a diff post-create. See
	// [official docs](https://cloud.google.com/bigquery/docs/access-control).
	Role pulumi.StringPtrOutput `pulumi:"role"`
	// A special group to grant access to. Possible values include:
	SpecialGroup pulumi.StringPtrOutput `pulumi:"specialGroup"`
	// An email address of a user to grant access to. For example:
	// fred@example.com
	UserByEmail pulumi.StringPtrOutput `pulumi:"userByEmail"`
	// A view from a different dataset to grant access to. Queries
	// executed against that view will have read access to tables in
	// this dataset. The role field is not required when this field is
	// set. If that view is updated by any user, access to the view
	// needs to be granted again via an update operation.  Structure is documented below.
	View DatasetAccessViewPtrOutput `pulumi:"view"`
}

Gives dataset access for a single entity. This resource is intended to be used in cases where it is not possible to compile a full list of access blocks to include in a `bigquery.Dataset` resource, to enable them to be added separately.

> **Note:** If this resource is used alongside a `bigquery.Dataset` resource, the dataset resource must either have no defined `access` blocks or a `lifecycle` block with `ignoreChanges = [access]` so they don't fight over which accesses should be on the dataset.

To get more information about DatasetAccess, see:

* [API documentation](https://cloud.google.com/bigquery/docs/reference/rest/v2/datasets) * How-to Guides

func GetDatasetAccess added in v3.1.0

func GetDatasetAccess(ctx *pulumi.Context,
	name string, id pulumi.IDInput, state *DatasetAccessState, opts ...pulumi.ResourceOption) (*DatasetAccess, error)

GetDatasetAccess gets an existing DatasetAccess resource's state with the given name, ID, and optional state properties that are used to uniquely qualify the lookup (nil if not required).

func NewDatasetAccess added in v3.1.0

func NewDatasetAccess(ctx *pulumi.Context,
	name string, args *DatasetAccessArgs, opts ...pulumi.ResourceOption) (*DatasetAccess, error)

NewDatasetAccess registers a new resource with the given unique name, arguments, and options.

type DatasetAccessArgs

type DatasetAccessArgs struct {
	// The ID of the dataset containing this table.
	DatasetId pulumi.StringInput
	// A domain to grant access to. Any users signed in with the
	// domain specified will be granted the specified access
	Domain pulumi.StringPtrInput
	// An email address of a Google Group to grant access to.
	GroupByEmail pulumi.StringPtrInput
	// Some other type of member that appears in the IAM Policy but isn't a user,
	// group, domain, or special group. For example: `allUsers`
	IamMember pulumi.StringPtrInput
	// The ID of the project in which the resource belongs.
	// If it is not provided, the provider project is used.
	Project pulumi.StringPtrInput
	// Describes the rights granted to the user specified by the other
	// member of the access object. Primitive, Predefined and custom
	// roles are supported. Predefined roles that have equivalent
	// primitive roles are swapped by the API to their Primitive
	// counterparts, and will show a diff post-create. See
	// [official docs](https://cloud.google.com/bigquery/docs/access-control).
	Role pulumi.StringPtrInput
	// A special group to grant access to. Possible values include:
	SpecialGroup pulumi.StringPtrInput
	// An email address of a user to grant access to. For example:
	// fred@example.com
	UserByEmail pulumi.StringPtrInput
	// A view from a different dataset to grant access to. Queries
	// executed against that view will have read access to tables in
	// this dataset. The role field is not required when this field is
	// set. If that view is updated by any user, access to the view
	// needs to be granted again via an update operation.  Structure is documented below.
	View DatasetAccessViewPtrInput
}

The set of arguments for constructing a DatasetAccess resource.

func (DatasetAccessArgs) ElementType

func (DatasetAccessArgs) ElementType() reflect.Type

type DatasetAccessState added in v3.1.0

type DatasetAccessState struct {
	// The ID of the dataset containing this table.
	DatasetId pulumi.StringPtrInput
	// A domain to grant access to. Any users signed in with the
	// domain specified will be granted the specified access
	Domain pulumi.StringPtrInput
	// An email address of a Google Group to grant access to.
	GroupByEmail pulumi.StringPtrInput
	// Some other type of member that appears in the IAM Policy but isn't a user,
	// group, domain, or special group. For example: `allUsers`
	IamMember pulumi.StringPtrInput
	// The ID of the project in which the resource belongs.
	// If it is not provided, the provider project is used.
	Project pulumi.StringPtrInput
	// Describes the rights granted to the user specified by the other
	// member of the access object. Primitive, Predefined and custom
	// roles are supported. Predefined roles that have equivalent
	// primitive roles are swapped by the API to their Primitive
	// counterparts, and will show a diff post-create. See
	// [official docs](https://cloud.google.com/bigquery/docs/access-control).
	Role pulumi.StringPtrInput
	// A special group to grant access to. Possible values include:
	SpecialGroup pulumi.StringPtrInput
	// An email address of a user to grant access to. For example:
	// fred@example.com
	UserByEmail pulumi.StringPtrInput
	// A view from a different dataset to grant access to. Queries
	// executed against that view will have read access to tables in
	// this dataset. The role field is not required when this field is
	// set. If that view is updated by any user, access to the view
	// needs to be granted again via an update operation.  Structure is documented below.
	View DatasetAccessViewPtrInput
}

func (DatasetAccessState) ElementType added in v3.1.0

func (DatasetAccessState) ElementType() reflect.Type

type DatasetAccessType added in v3.1.0

type DatasetAccessType struct {
	// A domain to grant access to. Any users signed in with the
	// domain specified will be granted the specified access
	Domain *string `pulumi:"domain"`
	// An email address of a Google Group to grant access to.
	GroupByEmail *string `pulumi:"groupByEmail"`
	// Describes the rights granted to the user specified by the other
	// member of the access object. Primitive, Predefined and custom
	// roles are supported. Predefined roles that have equivalent
	// primitive roles are swapped by the API to their Primitive
	// counterparts. See
	// [official docs](https://cloud.google.com/bigquery/docs/access-control).
	Role *string `pulumi:"role"`
	// A special group to grant access to. Possible values include:
	SpecialGroup *string `pulumi:"specialGroup"`
	// An email address of a user to grant access to. For example:
	// fred@example.com
	UserByEmail *string `pulumi:"userByEmail"`
	// A view from a different dataset to grant access to. Queries
	// executed against that view will have read access to tables in
	// this dataset. The role field is not required when this field is
	// set. If that view is updated by any user, access to the view
	// needs to be granted again via an update operation.  Structure is documented below.
	View *DatasetAccessView `pulumi:"view"`
}

type DatasetAccessTypeArgs added in v3.1.0

type DatasetAccessTypeArgs struct {
	// A domain to grant access to. Any users signed in with the
	// domain specified will be granted the specified access
	Domain pulumi.StringPtrInput `pulumi:"domain"`
	// An email address of a Google Group to grant access to.
	GroupByEmail pulumi.StringPtrInput `pulumi:"groupByEmail"`
	// Describes the rights granted to the user specified by the other
	// member of the access object. Primitive, Predefined and custom
	// roles are supported. Predefined roles that have equivalent
	// primitive roles are swapped by the API to their Primitive
	// counterparts. See
	// [official docs](https://cloud.google.com/bigquery/docs/access-control).
	Role pulumi.StringPtrInput `pulumi:"role"`
	// A special group to grant access to. Possible values include:
	SpecialGroup pulumi.StringPtrInput `pulumi:"specialGroup"`
	// An email address of a user to grant access to. For example:
	// fred@example.com
	UserByEmail pulumi.StringPtrInput `pulumi:"userByEmail"`
	// A view from a different dataset to grant access to. Queries
	// executed against that view will have read access to tables in
	// this dataset. The role field is not required when this field is
	// set. If that view is updated by any user, access to the view
	// needs to be granted again via an update operation.  Structure is documented below.
	View DatasetAccessViewPtrInput `pulumi:"view"`
}

func (DatasetAccessTypeArgs) ElementType added in v3.1.0

func (DatasetAccessTypeArgs) ElementType() reflect.Type

func (DatasetAccessTypeArgs) ToDatasetAccessTypeOutput added in v3.1.0

func (i DatasetAccessTypeArgs) ToDatasetAccessTypeOutput() DatasetAccessTypeOutput

func (DatasetAccessTypeArgs) ToDatasetAccessTypeOutputWithContext added in v3.1.0

func (i DatasetAccessTypeArgs) ToDatasetAccessTypeOutputWithContext(ctx context.Context) DatasetAccessTypeOutput

type DatasetAccessTypeArray added in v3.1.0

type DatasetAccessTypeArray []DatasetAccessTypeInput

func (DatasetAccessTypeArray) ElementType added in v3.1.0

func (DatasetAccessTypeArray) ElementType() reflect.Type

func (DatasetAccessTypeArray) ToDatasetAccessTypeArrayOutput added in v3.1.0

func (i DatasetAccessTypeArray) ToDatasetAccessTypeArrayOutput() DatasetAccessTypeArrayOutput

func (DatasetAccessTypeArray) ToDatasetAccessTypeArrayOutputWithContext added in v3.1.0

func (i DatasetAccessTypeArray) ToDatasetAccessTypeArrayOutputWithContext(ctx context.Context) DatasetAccessTypeArrayOutput

type DatasetAccessTypeArrayInput added in v3.1.0

type DatasetAccessTypeArrayInput interface {
	pulumi.Input

	ToDatasetAccessTypeArrayOutput() DatasetAccessTypeArrayOutput
	ToDatasetAccessTypeArrayOutputWithContext(context.Context) DatasetAccessTypeArrayOutput
}

DatasetAccessTypeArrayInput is an input type that accepts DatasetAccessTypeArray and DatasetAccessTypeArrayOutput values. You can construct a concrete instance of `DatasetAccessTypeArrayInput` via:

DatasetAccessTypeArray{ DatasetAccessTypeArgs{...} }

type DatasetAccessTypeArrayOutput added in v3.1.0

type DatasetAccessTypeArrayOutput struct{ *pulumi.OutputState }

func (DatasetAccessTypeArrayOutput) ElementType added in v3.1.0

func (DatasetAccessTypeArrayOutput) Index added in v3.1.0

func (DatasetAccessTypeArrayOutput) ToDatasetAccessTypeArrayOutput added in v3.1.0

func (o DatasetAccessTypeArrayOutput) ToDatasetAccessTypeArrayOutput() DatasetAccessTypeArrayOutput

func (DatasetAccessTypeArrayOutput) ToDatasetAccessTypeArrayOutputWithContext added in v3.1.0

func (o DatasetAccessTypeArrayOutput) ToDatasetAccessTypeArrayOutputWithContext(ctx context.Context) DatasetAccessTypeArrayOutput

type DatasetAccessTypeInput added in v3.1.0

type DatasetAccessTypeInput interface {
	pulumi.Input

	ToDatasetAccessTypeOutput() DatasetAccessTypeOutput
	ToDatasetAccessTypeOutputWithContext(context.Context) DatasetAccessTypeOutput
}

DatasetAccessTypeInput is an input type that accepts DatasetAccessTypeArgs and DatasetAccessTypeOutput values. You can construct a concrete instance of `DatasetAccessTypeInput` via:

DatasetAccessTypeArgs{...}

type DatasetAccessTypeOutput added in v3.1.0

type DatasetAccessTypeOutput struct{ *pulumi.OutputState }

func (DatasetAccessTypeOutput) Domain added in v3.1.0

A domain to grant access to. Any users signed in with the domain specified will be granted the specified access

func (DatasetAccessTypeOutput) ElementType added in v3.1.0

func (DatasetAccessTypeOutput) ElementType() reflect.Type

func (DatasetAccessTypeOutput) GroupByEmail added in v3.1.0

An email address of a Google Group to grant access to.

func (DatasetAccessTypeOutput) Role added in v3.1.0

Describes the rights granted to the user specified by the other member of the access object. Primitive, Predefined and custom roles are supported. Predefined roles that have equivalent primitive roles are swapped by the API to their Primitive counterparts. See [official docs](https://cloud.google.com/bigquery/docs/access-control).

func (DatasetAccessTypeOutput) SpecialGroup added in v3.1.0

A special group to grant access to. Possible values include:

func (DatasetAccessTypeOutput) ToDatasetAccessTypeOutput added in v3.1.0

func (o DatasetAccessTypeOutput) ToDatasetAccessTypeOutput() DatasetAccessTypeOutput

func (DatasetAccessTypeOutput) ToDatasetAccessTypeOutputWithContext added in v3.1.0

func (o DatasetAccessTypeOutput) ToDatasetAccessTypeOutputWithContext(ctx context.Context) DatasetAccessTypeOutput

func (DatasetAccessTypeOutput) UserByEmail added in v3.1.0

An email address of a user to grant access to. For example: fred@example.com

func (DatasetAccessTypeOutput) View added in v3.1.0

A view from a different dataset to grant access to. Queries executed against that view will have read access to tables in this dataset. The role field is not required when this field is set. If that view is updated by any user, access to the view needs to be granted again via an update operation. Structure is documented below.

type DatasetAccessView

type DatasetAccessView struct {
	// The ID of the dataset containing this table.
	DatasetId string `pulumi:"datasetId"`
	// The ID of the project containing this table.
	ProjectId string `pulumi:"projectId"`
	// The ID of the table. The ID must contain only letters (a-z,
	// A-Z), numbers (0-9), or underscores (_). The maximum length
	// is 1,024 characters.
	TableId string `pulumi:"tableId"`
}

type DatasetAccessViewArgs

type DatasetAccessViewArgs struct {
	// The ID of the dataset containing this table.
	DatasetId pulumi.StringInput `pulumi:"datasetId"`
	// The ID of the project containing this table.
	ProjectId pulumi.StringInput `pulumi:"projectId"`
	// The ID of the table. The ID must contain only letters (a-z,
	// A-Z), numbers (0-9), or underscores (_). The maximum length
	// is 1,024 characters.
	TableId pulumi.StringInput `pulumi:"tableId"`
}

func (DatasetAccessViewArgs) ElementType

func (DatasetAccessViewArgs) ElementType() reflect.Type

func (DatasetAccessViewArgs) ToDatasetAccessViewOutput

func (i DatasetAccessViewArgs) ToDatasetAccessViewOutput() DatasetAccessViewOutput

func (DatasetAccessViewArgs) ToDatasetAccessViewOutputWithContext

func (i DatasetAccessViewArgs) ToDatasetAccessViewOutputWithContext(ctx context.Context) DatasetAccessViewOutput

func (DatasetAccessViewArgs) ToDatasetAccessViewPtrOutput

func (i DatasetAccessViewArgs) ToDatasetAccessViewPtrOutput() DatasetAccessViewPtrOutput

func (DatasetAccessViewArgs) ToDatasetAccessViewPtrOutputWithContext

func (i DatasetAccessViewArgs) ToDatasetAccessViewPtrOutputWithContext(ctx context.Context) DatasetAccessViewPtrOutput

type DatasetAccessViewInput

type DatasetAccessViewInput interface {
	pulumi.Input

	ToDatasetAccessViewOutput() DatasetAccessViewOutput
	ToDatasetAccessViewOutputWithContext(context.Context) DatasetAccessViewOutput
}

DatasetAccessViewInput is an input type that accepts DatasetAccessViewArgs and DatasetAccessViewOutput values. You can construct a concrete instance of `DatasetAccessViewInput` via:

DatasetAccessViewArgs{...}

type DatasetAccessViewOutput

type DatasetAccessViewOutput struct{ *pulumi.OutputState }

func (DatasetAccessViewOutput) DatasetId

The ID of the dataset containing this table.

func (DatasetAccessViewOutput) ElementType

func (DatasetAccessViewOutput) ElementType() reflect.Type

func (DatasetAccessViewOutput) ProjectId

The ID of the project containing this table.

func (DatasetAccessViewOutput) TableId

The ID of the table. The ID must contain only letters (a-z, A-Z), numbers (0-9), or underscores (_). The maximum length is 1,024 characters.

func (DatasetAccessViewOutput) ToDatasetAccessViewOutput

func (o DatasetAccessViewOutput) ToDatasetAccessViewOutput() DatasetAccessViewOutput

func (DatasetAccessViewOutput) ToDatasetAccessViewOutputWithContext

func (o DatasetAccessViewOutput) ToDatasetAccessViewOutputWithContext(ctx context.Context) DatasetAccessViewOutput

func (DatasetAccessViewOutput) ToDatasetAccessViewPtrOutput

func (o DatasetAccessViewOutput) ToDatasetAccessViewPtrOutput() DatasetAccessViewPtrOutput

func (DatasetAccessViewOutput) ToDatasetAccessViewPtrOutputWithContext

func (o DatasetAccessViewOutput) ToDatasetAccessViewPtrOutputWithContext(ctx context.Context) DatasetAccessViewPtrOutput

type DatasetAccessViewPtrInput

type DatasetAccessViewPtrInput interface {
	pulumi.Input

	ToDatasetAccessViewPtrOutput() DatasetAccessViewPtrOutput
	ToDatasetAccessViewPtrOutputWithContext(context.Context) DatasetAccessViewPtrOutput
}

DatasetAccessViewPtrInput is an input type that accepts DatasetAccessViewArgs, DatasetAccessViewPtr and DatasetAccessViewPtrOutput values. You can construct a concrete instance of `DatasetAccessViewPtrInput` via:

		 DatasetAccessViewArgs{...}

 or:

		 nil

type DatasetAccessViewPtrOutput

type DatasetAccessViewPtrOutput struct{ *pulumi.OutputState }

func (DatasetAccessViewPtrOutput) DatasetId

The ID of the dataset containing this table.

func (DatasetAccessViewPtrOutput) Elem

func (DatasetAccessViewPtrOutput) ElementType

func (DatasetAccessViewPtrOutput) ElementType() reflect.Type

func (DatasetAccessViewPtrOutput) ProjectId

The ID of the project containing this table.

func (DatasetAccessViewPtrOutput) TableId

The ID of the table. The ID must contain only letters (a-z, A-Z), numbers (0-9), or underscores (_). The maximum length is 1,024 characters.

func (DatasetAccessViewPtrOutput) ToDatasetAccessViewPtrOutput

func (o DatasetAccessViewPtrOutput) ToDatasetAccessViewPtrOutput() DatasetAccessViewPtrOutput

func (DatasetAccessViewPtrOutput) ToDatasetAccessViewPtrOutputWithContext

func (o DatasetAccessViewPtrOutput) ToDatasetAccessViewPtrOutputWithContext(ctx context.Context) DatasetAccessViewPtrOutput

type DatasetArgs

type DatasetArgs struct {
	// An array of objects that define dataset access for one or more entities.  Structure is documented below.
	Accesses DatasetAccessTypeArrayInput
	// The ID of the dataset containing this table.
	DatasetId pulumi.StringInput
	// The default encryption key for all tables in the dataset. Once this property is set,
	// all newly-created partitioned tables in the dataset will have encryption key set to
	// this value, unless table creation request (or query) overrides the key.  Structure is documented below.
	DefaultEncryptionConfiguration DatasetDefaultEncryptionConfigurationPtrInput
	// The default partition expiration for all partitioned tables in
	// the dataset, in milliseconds.
	DefaultPartitionExpirationMs pulumi.IntPtrInput
	// The default lifetime of all tables in the dataset, in milliseconds.
	// The minimum value is 3600000 milliseconds (one hour).
	DefaultTableExpirationMs pulumi.IntPtrInput
	// If set to `true`, delete all the tables in the
	// dataset when destroying the resource; otherwise,
	// destroying the resource will fail if tables are present.
	DeleteContentsOnDestroy pulumi.BoolPtrInput
	// A user-friendly description of the dataset
	Description pulumi.StringPtrInput
	// A descriptive name for the dataset
	FriendlyName pulumi.StringPtrInput
	// The labels associated with this dataset. You can use these to
	// organize and group your datasets
	Labels pulumi.StringMapInput
	// The geographic location where the dataset should reside.
	// See [official docs](https://cloud.google.com/bigquery/docs/dataset-locations).
	Location pulumi.StringPtrInput
	// The ID of the project in which the resource belongs.
	// If it is not provided, the provider project is used.
	Project pulumi.StringPtrInput
}

The set of arguments for constructing a Dataset resource.

func (DatasetArgs) ElementType

func (DatasetArgs) ElementType() reflect.Type

type DatasetDefaultEncryptionConfiguration

type DatasetDefaultEncryptionConfiguration struct {
	// Describes the Cloud KMS encryption key that will be used to protect destination
	// BigQuery table. The BigQuery Service Account associated with your project requires
	// access to this encryption key.
	KmsKeyName string `pulumi:"kmsKeyName"`
}

type DatasetDefaultEncryptionConfigurationArgs

type DatasetDefaultEncryptionConfigurationArgs struct {
	// Describes the Cloud KMS encryption key that will be used to protect destination
	// BigQuery table. The BigQuery Service Account associated with your project requires
	// access to this encryption key.
	KmsKeyName pulumi.StringInput `pulumi:"kmsKeyName"`
}

func (DatasetDefaultEncryptionConfigurationArgs) ElementType

func (DatasetDefaultEncryptionConfigurationArgs) ToDatasetDefaultEncryptionConfigurationOutput

func (i DatasetDefaultEncryptionConfigurationArgs) ToDatasetDefaultEncryptionConfigurationOutput() DatasetDefaultEncryptionConfigurationOutput

func (DatasetDefaultEncryptionConfigurationArgs) ToDatasetDefaultEncryptionConfigurationOutputWithContext

func (i DatasetDefaultEncryptionConfigurationArgs) ToDatasetDefaultEncryptionConfigurationOutputWithContext(ctx context.Context) DatasetDefaultEncryptionConfigurationOutput

func (DatasetDefaultEncryptionConfigurationArgs) ToDatasetDefaultEncryptionConfigurationPtrOutput

func (i DatasetDefaultEncryptionConfigurationArgs) ToDatasetDefaultEncryptionConfigurationPtrOutput() DatasetDefaultEncryptionConfigurationPtrOutput

func (DatasetDefaultEncryptionConfigurationArgs) ToDatasetDefaultEncryptionConfigurationPtrOutputWithContext

func (i DatasetDefaultEncryptionConfigurationArgs) ToDatasetDefaultEncryptionConfigurationPtrOutputWithContext(ctx context.Context) DatasetDefaultEncryptionConfigurationPtrOutput

type DatasetDefaultEncryptionConfigurationInput

type DatasetDefaultEncryptionConfigurationInput interface {
	pulumi.Input

	ToDatasetDefaultEncryptionConfigurationOutput() DatasetDefaultEncryptionConfigurationOutput
	ToDatasetDefaultEncryptionConfigurationOutputWithContext(context.Context) DatasetDefaultEncryptionConfigurationOutput
}

DatasetDefaultEncryptionConfigurationInput is an input type that accepts DatasetDefaultEncryptionConfigurationArgs and DatasetDefaultEncryptionConfigurationOutput values. You can construct a concrete instance of `DatasetDefaultEncryptionConfigurationInput` via:

DatasetDefaultEncryptionConfigurationArgs{...}

type DatasetDefaultEncryptionConfigurationOutput

type DatasetDefaultEncryptionConfigurationOutput struct{ *pulumi.OutputState }

func (DatasetDefaultEncryptionConfigurationOutput) ElementType

func (DatasetDefaultEncryptionConfigurationOutput) KmsKeyName

Describes the Cloud KMS encryption key that will be used to protect destination BigQuery table. The BigQuery Service Account associated with your project requires access to this encryption key.

func (DatasetDefaultEncryptionConfigurationOutput) ToDatasetDefaultEncryptionConfigurationOutput

func (o DatasetDefaultEncryptionConfigurationOutput) ToDatasetDefaultEncryptionConfigurationOutput() DatasetDefaultEncryptionConfigurationOutput

func (DatasetDefaultEncryptionConfigurationOutput) ToDatasetDefaultEncryptionConfigurationOutputWithContext

func (o DatasetDefaultEncryptionConfigurationOutput) ToDatasetDefaultEncryptionConfigurationOutputWithContext(ctx context.Context) DatasetDefaultEncryptionConfigurationOutput

func (DatasetDefaultEncryptionConfigurationOutput) ToDatasetDefaultEncryptionConfigurationPtrOutput

func (o DatasetDefaultEncryptionConfigurationOutput) ToDatasetDefaultEncryptionConfigurationPtrOutput() DatasetDefaultEncryptionConfigurationPtrOutput

func (DatasetDefaultEncryptionConfigurationOutput) ToDatasetDefaultEncryptionConfigurationPtrOutputWithContext

func (o DatasetDefaultEncryptionConfigurationOutput) ToDatasetDefaultEncryptionConfigurationPtrOutputWithContext(ctx context.Context) DatasetDefaultEncryptionConfigurationPtrOutput

type DatasetDefaultEncryptionConfigurationPtrInput

type DatasetDefaultEncryptionConfigurationPtrInput interface {
	pulumi.Input

	ToDatasetDefaultEncryptionConfigurationPtrOutput() DatasetDefaultEncryptionConfigurationPtrOutput
	ToDatasetDefaultEncryptionConfigurationPtrOutputWithContext(context.Context) DatasetDefaultEncryptionConfigurationPtrOutput
}

DatasetDefaultEncryptionConfigurationPtrInput is an input type that accepts DatasetDefaultEncryptionConfigurationArgs, DatasetDefaultEncryptionConfigurationPtr and DatasetDefaultEncryptionConfigurationPtrOutput values. You can construct a concrete instance of `DatasetDefaultEncryptionConfigurationPtrInput` via:

		 DatasetDefaultEncryptionConfigurationArgs{...}

 or:

		 nil

type DatasetDefaultEncryptionConfigurationPtrOutput

type DatasetDefaultEncryptionConfigurationPtrOutput struct{ *pulumi.OutputState }

func (DatasetDefaultEncryptionConfigurationPtrOutput) Elem

func (DatasetDefaultEncryptionConfigurationPtrOutput) ElementType

func (DatasetDefaultEncryptionConfigurationPtrOutput) KmsKeyName

Describes the Cloud KMS encryption key that will be used to protect destination BigQuery table. The BigQuery Service Account associated with your project requires access to this encryption key.

func (DatasetDefaultEncryptionConfigurationPtrOutput) ToDatasetDefaultEncryptionConfigurationPtrOutput

func (o DatasetDefaultEncryptionConfigurationPtrOutput) ToDatasetDefaultEncryptionConfigurationPtrOutput() DatasetDefaultEncryptionConfigurationPtrOutput

func (DatasetDefaultEncryptionConfigurationPtrOutput) ToDatasetDefaultEncryptionConfigurationPtrOutputWithContext

func (o DatasetDefaultEncryptionConfigurationPtrOutput) ToDatasetDefaultEncryptionConfigurationPtrOutputWithContext(ctx context.Context) DatasetDefaultEncryptionConfigurationPtrOutput

type DatasetState

type DatasetState struct {
	// An array of objects that define dataset access for one or more entities.  Structure is documented below.
	Accesses DatasetAccessTypeArrayInput
	// The time when this dataset was created, in milliseconds since the epoch.
	CreationTime pulumi.IntPtrInput
	// The ID of the dataset containing this table.
	DatasetId pulumi.StringPtrInput
	// The default encryption key for all tables in the dataset. Once this property is set,
	// all newly-created partitioned tables in the dataset will have encryption key set to
	// this value, unless table creation request (or query) overrides the key.  Structure is documented below.
	DefaultEncryptionConfiguration DatasetDefaultEncryptionConfigurationPtrInput
	// The default partition expiration for all partitioned tables in
	// the dataset, in milliseconds.
	DefaultPartitionExpirationMs pulumi.IntPtrInput
	// The default lifetime of all tables in the dataset, in milliseconds.
	// The minimum value is 3600000 milliseconds (one hour).
	DefaultTableExpirationMs pulumi.IntPtrInput
	// If set to `true`, delete all the tables in the
	// dataset when destroying the resource; otherwise,
	// destroying the resource will fail if tables are present.
	DeleteContentsOnDestroy pulumi.BoolPtrInput
	// A user-friendly description of the dataset
	Description pulumi.StringPtrInput
	// A hash of the resource.
	Etag pulumi.StringPtrInput
	// A descriptive name for the dataset
	FriendlyName pulumi.StringPtrInput
	// The labels associated with this dataset. You can use these to
	// organize and group your datasets
	Labels pulumi.StringMapInput
	// The date when this dataset or any of its tables was last modified, in milliseconds since the epoch.
	LastModifiedTime pulumi.IntPtrInput
	// The geographic location where the dataset should reside.
	// See [official docs](https://cloud.google.com/bigquery/docs/dataset-locations).
	Location pulumi.StringPtrInput
	// The ID of the project in which the resource belongs.
	// If it is not provided, the provider project is used.
	Project pulumi.StringPtrInput
	// The URI of the created resource.
	SelfLink pulumi.StringPtrInput
}

func (DatasetState) ElementType

func (DatasetState) ElementType() reflect.Type

type GetDefaultServiceAccountArgs

type GetDefaultServiceAccountArgs struct {
	// The project the unique service account was created for. If it is not provided, the provider project is used.
	Project *string `pulumi:"project"`
}

A collection of arguments for invoking getDefaultServiceAccount.

type GetDefaultServiceAccountResult

type GetDefaultServiceAccountResult struct {
	// The email address of the service account. This value is often used to refer to the service account
	// in order to grant IAM permissions.
	Email string `pulumi:"email"`
	// The provider-assigned unique ID for this managed resource.
	Id      string `pulumi:"id"`
	Project string `pulumi:"project"`
}

A collection of values returned by getDefaultServiceAccount.

func GetDefaultServiceAccount

func GetDefaultServiceAccount(ctx *pulumi.Context, args *GetDefaultServiceAccountArgs, opts ...pulumi.InvokeOption) (*GetDefaultServiceAccountResult, error)

Get the email address of a project's unique BigQuery service account.

Each Google Cloud project has a unique service account used by BigQuery. When using BigQuery with [customer-managed encryption keys](https://cloud.google.com/bigquery/docs/customer-managed-encryption), this account needs to be granted the `cloudkms.cryptoKeyEncrypterDecrypter` IAM role on the customer-managed Cloud KMS key used to protect the data.

For more information see [the API reference](https://cloud.google.com/bigquery/docs/reference/rest/v2/projects/getServiceAccount).

type Job added in v3.3.0

type Job struct {
	pulumi.CustomResourceState

	// Copies a table.  Structure is documented below.
	Copy JobCopyPtrOutput `pulumi:"copy"`
	// Configures an extract job.  Structure is documented below.
	Extract JobExtractPtrOutput `pulumi:"extract"`
	// The ID of the job. The ID must contain only letters (a-z, A-Z), numbers (0-9), underscores (_), or dashes (-). The maximum length is 1,024 characters.
	JobId pulumi.StringOutput `pulumi:"jobId"`
	// Job timeout in milliseconds. If this time limit is exceeded, BigQuery may attempt to terminate the job.
	JobTimeoutMs pulumi.StringPtrOutput `pulumi:"jobTimeoutMs"`
	// The type of the job.
	JobType pulumi.StringOutput `pulumi:"jobType"`
	// The labels associated with this job. You can use these to organize and group your jobs.
	Labels pulumi.StringMapOutput `pulumi:"labels"`
	// Configures a load job.  Structure is documented below.
	Load JobLoadPtrOutput `pulumi:"load"`
	// The geographic location of the job. The default value is US.
	Location pulumi.StringPtrOutput `pulumi:"location"`
	// The ID of the project in which the resource belongs.
	// If it is not provided, the provider project is used.
	Project pulumi.StringOutput `pulumi:"project"`
	// Configures a query job.  Structure is documented below.
	Query JobQueryPtrOutput `pulumi:"query"`
	// Email address of the user who ran the job.
	UserEmail pulumi.StringOutput `pulumi:"userEmail"`
}

Jobs are actions that BigQuery runs on your behalf to load data, export data, query data, or copy data. Once a BigQuery job is created, it cannot be changed or deleted.

func GetJob added in v3.3.0

func GetJob(ctx *pulumi.Context,
	name string, id pulumi.IDInput, state *JobState, opts ...pulumi.ResourceOption) (*Job, error)

GetJob gets an existing Job resource's state with the given name, ID, and optional state properties that are used to uniquely qualify the lookup (nil if not required).

func NewJob added in v3.3.0

func NewJob(ctx *pulumi.Context,
	name string, args *JobArgs, opts ...pulumi.ResourceOption) (*Job, error)

NewJob registers a new resource with the given unique name, arguments, and options.

type JobArgs added in v3.3.0

type JobArgs struct {
	// Copies a table.  Structure is documented below.
	Copy JobCopyPtrInput
	// Configures an extract job.  Structure is documented below.
	Extract JobExtractPtrInput
	// The ID of the job. The ID must contain only letters (a-z, A-Z), numbers (0-9), underscores (_), or dashes (-). The maximum length is 1,024 characters.
	JobId pulumi.StringInput
	// Job timeout in milliseconds. If this time limit is exceeded, BigQuery may attempt to terminate the job.
	JobTimeoutMs pulumi.StringPtrInput
	// The labels associated with this job. You can use these to organize and group your jobs.
	Labels pulumi.StringMapInput
	// Configures a load job.  Structure is documented below.
	Load JobLoadPtrInput
	// The geographic location of the job. The default value is US.
	Location pulumi.StringPtrInput
	// The ID of the project in which the resource belongs.
	// If it is not provided, the provider project is used.
	Project pulumi.StringPtrInput
	// Configures a query job.  Structure is documented below.
	Query JobQueryPtrInput
}

The set of arguments for constructing a Job resource.

func (JobArgs) ElementType added in v3.3.0

func (JobArgs) ElementType() reflect.Type

type JobCopy added in v3.3.0

type JobCopy struct {
	// Specifies whether the job is allowed to create new tables. The following values are supported:
	// CREATE_IF_NEEDED: If the table does not exist, BigQuery creates the table.
	// CREATE_NEVER: The table must already exist. If it does not, a 'notFound' error is returned in the job result.
	// Creation, truncation and append actions occur as one atomic update upon job completion
	CreateDisposition *string `pulumi:"createDisposition"`
	// Custom encryption configuration (e.g., Cloud KMS keys)  Structure is documented below.
	DestinationEncryptionConfiguration *JobCopyDestinationEncryptionConfiguration `pulumi:"destinationEncryptionConfiguration"`
	// The destination table.  Structure is documented below.
	DestinationTable *JobCopyDestinationTable `pulumi:"destinationTable"`
	// Source tables to copy.  Structure is documented below.
	SourceTables []JobCopySourceTable `pulumi:"sourceTables"`
	// Specifies the action that occurs if the destination table already exists. The following values are supported:
	// WRITE_TRUNCATE: If the table already exists, BigQuery overwrites the table data and uses the schema from the query result.
	// WRITE_APPEND: If the table already exists, BigQuery appends the data to the table.
	// WRITE_EMPTY: If the table already exists and contains data, a 'duplicate' error is returned in the job result.
	// Each action is atomic and only occurs if BigQuery is able to complete the job successfully.
	// Creation, truncation and append actions occur as one atomic update upon job completion.
	WriteDisposition *string `pulumi:"writeDisposition"`
}

type JobCopyArgs added in v3.3.0

type JobCopyArgs struct {
	// Specifies whether the job is allowed to create new tables. The following values are supported:
	// CREATE_IF_NEEDED: If the table does not exist, BigQuery creates the table.
	// CREATE_NEVER: The table must already exist. If it does not, a 'notFound' error is returned in the job result.
	// Creation, truncation and append actions occur as one atomic update upon job completion
	CreateDisposition pulumi.StringPtrInput `pulumi:"createDisposition"`
	// Custom encryption configuration (e.g., Cloud KMS keys)  Structure is documented below.
	DestinationEncryptionConfiguration JobCopyDestinationEncryptionConfigurationPtrInput `pulumi:"destinationEncryptionConfiguration"`
	// The destination table.  Structure is documented below.
	DestinationTable JobCopyDestinationTablePtrInput `pulumi:"destinationTable"`
	// Source tables to copy.  Structure is documented below.
	SourceTables JobCopySourceTableArrayInput `pulumi:"sourceTables"`
	// Specifies the action that occurs if the destination table already exists. The following values are supported:
	// WRITE_TRUNCATE: If the table already exists, BigQuery overwrites the table data and uses the schema from the query result.
	// WRITE_APPEND: If the table already exists, BigQuery appends the data to the table.
	// WRITE_EMPTY: If the table already exists and contains data, a 'duplicate' error is returned in the job result.
	// Each action is atomic and only occurs if BigQuery is able to complete the job successfully.
	// Creation, truncation and append actions occur as one atomic update upon job completion.
	WriteDisposition pulumi.StringPtrInput `pulumi:"writeDisposition"`
}

func (JobCopyArgs) ElementType added in v3.3.0

func (JobCopyArgs) ElementType() reflect.Type

func (JobCopyArgs) ToJobCopyOutput added in v3.3.0

func (i JobCopyArgs) ToJobCopyOutput() JobCopyOutput

func (JobCopyArgs) ToJobCopyOutputWithContext added in v3.3.0

func (i JobCopyArgs) ToJobCopyOutputWithContext(ctx context.Context) JobCopyOutput

func (JobCopyArgs) ToJobCopyPtrOutput added in v3.3.0

func (i JobCopyArgs) ToJobCopyPtrOutput() JobCopyPtrOutput

func (JobCopyArgs) ToJobCopyPtrOutputWithContext added in v3.3.0

func (i JobCopyArgs) ToJobCopyPtrOutputWithContext(ctx context.Context) JobCopyPtrOutput

type JobCopyDestinationEncryptionConfiguration added in v3.3.0

type JobCopyDestinationEncryptionConfiguration struct {
	// Describes the Cloud KMS encryption key that will be used to protect destination BigQuery table.
	// The BigQuery Service Account associated with your project requires access to this encryption key.
	KmsKeyName string `pulumi:"kmsKeyName"`
}

type JobCopyDestinationEncryptionConfigurationArgs added in v3.3.0

type JobCopyDestinationEncryptionConfigurationArgs struct {
	// Describes the Cloud KMS encryption key that will be used to protect destination BigQuery table.
	// The BigQuery Service Account associated with your project requires access to this encryption key.
	KmsKeyName pulumi.StringInput `pulumi:"kmsKeyName"`
}

func (JobCopyDestinationEncryptionConfigurationArgs) ElementType added in v3.3.0

func (JobCopyDestinationEncryptionConfigurationArgs) ToJobCopyDestinationEncryptionConfigurationOutput added in v3.3.0

func (i JobCopyDestinationEncryptionConfigurationArgs) ToJobCopyDestinationEncryptionConfigurationOutput() JobCopyDestinationEncryptionConfigurationOutput

func (JobCopyDestinationEncryptionConfigurationArgs) ToJobCopyDestinationEncryptionConfigurationOutputWithContext added in v3.3.0

func (i JobCopyDestinationEncryptionConfigurationArgs) ToJobCopyDestinationEncryptionConfigurationOutputWithContext(ctx context.Context) JobCopyDestinationEncryptionConfigurationOutput

func (JobCopyDestinationEncryptionConfigurationArgs) ToJobCopyDestinationEncryptionConfigurationPtrOutput added in v3.3.0

func (i JobCopyDestinationEncryptionConfigurationArgs) ToJobCopyDestinationEncryptionConfigurationPtrOutput() JobCopyDestinationEncryptionConfigurationPtrOutput

func (JobCopyDestinationEncryptionConfigurationArgs) ToJobCopyDestinationEncryptionConfigurationPtrOutputWithContext added in v3.3.0

func (i JobCopyDestinationEncryptionConfigurationArgs) ToJobCopyDestinationEncryptionConfigurationPtrOutputWithContext(ctx context.Context) JobCopyDestinationEncryptionConfigurationPtrOutput

type JobCopyDestinationEncryptionConfigurationInput added in v3.3.0

type JobCopyDestinationEncryptionConfigurationInput interface {
	pulumi.Input

	ToJobCopyDestinationEncryptionConfigurationOutput() JobCopyDestinationEncryptionConfigurationOutput
	ToJobCopyDestinationEncryptionConfigurationOutputWithContext(context.Context) JobCopyDestinationEncryptionConfigurationOutput
}

JobCopyDestinationEncryptionConfigurationInput is an input type that accepts JobCopyDestinationEncryptionConfigurationArgs and JobCopyDestinationEncryptionConfigurationOutput values. You can construct a concrete instance of `JobCopyDestinationEncryptionConfigurationInput` via:

JobCopyDestinationEncryptionConfigurationArgs{...}

type JobCopyDestinationEncryptionConfigurationOutput added in v3.3.0

type JobCopyDestinationEncryptionConfigurationOutput struct{ *pulumi.OutputState }

func (JobCopyDestinationEncryptionConfigurationOutput) ElementType added in v3.3.0

func (JobCopyDestinationEncryptionConfigurationOutput) KmsKeyName added in v3.3.0

Describes the Cloud KMS encryption key that will be used to protect destination BigQuery table. The BigQuery Service Account associated with your project requires access to this encryption key.

func (JobCopyDestinationEncryptionConfigurationOutput) ToJobCopyDestinationEncryptionConfigurationOutput added in v3.3.0

func (o JobCopyDestinationEncryptionConfigurationOutput) ToJobCopyDestinationEncryptionConfigurationOutput() JobCopyDestinationEncryptionConfigurationOutput

func (JobCopyDestinationEncryptionConfigurationOutput) ToJobCopyDestinationEncryptionConfigurationOutputWithContext added in v3.3.0

func (o JobCopyDestinationEncryptionConfigurationOutput) ToJobCopyDestinationEncryptionConfigurationOutputWithContext(ctx context.Context) JobCopyDestinationEncryptionConfigurationOutput

func (JobCopyDestinationEncryptionConfigurationOutput) ToJobCopyDestinationEncryptionConfigurationPtrOutput added in v3.3.0

func (o JobCopyDestinationEncryptionConfigurationOutput) ToJobCopyDestinationEncryptionConfigurationPtrOutput() JobCopyDestinationEncryptionConfigurationPtrOutput

func (JobCopyDestinationEncryptionConfigurationOutput) ToJobCopyDestinationEncryptionConfigurationPtrOutputWithContext added in v3.3.0

func (o JobCopyDestinationEncryptionConfigurationOutput) ToJobCopyDestinationEncryptionConfigurationPtrOutputWithContext(ctx context.Context) JobCopyDestinationEncryptionConfigurationPtrOutput

type JobCopyDestinationEncryptionConfigurationPtrInput added in v3.3.0

type JobCopyDestinationEncryptionConfigurationPtrInput interface {
	pulumi.Input

	ToJobCopyDestinationEncryptionConfigurationPtrOutput() JobCopyDestinationEncryptionConfigurationPtrOutput
	ToJobCopyDestinationEncryptionConfigurationPtrOutputWithContext(context.Context) JobCopyDestinationEncryptionConfigurationPtrOutput
}

JobCopyDestinationEncryptionConfigurationPtrInput is an input type that accepts JobCopyDestinationEncryptionConfigurationArgs, JobCopyDestinationEncryptionConfigurationPtr and JobCopyDestinationEncryptionConfigurationPtrOutput values. You can construct a concrete instance of `JobCopyDestinationEncryptionConfigurationPtrInput` via:

		 JobCopyDestinationEncryptionConfigurationArgs{...}

 or:

		 nil

type JobCopyDestinationEncryptionConfigurationPtrOutput added in v3.3.0

type JobCopyDestinationEncryptionConfigurationPtrOutput struct{ *pulumi.OutputState }

func (JobCopyDestinationEncryptionConfigurationPtrOutput) Elem added in v3.3.0

func (JobCopyDestinationEncryptionConfigurationPtrOutput) ElementType added in v3.3.0

func (JobCopyDestinationEncryptionConfigurationPtrOutput) KmsKeyName added in v3.3.0

Describes the Cloud KMS encryption key that will be used to protect destination BigQuery table. The BigQuery Service Account associated with your project requires access to this encryption key.

func (JobCopyDestinationEncryptionConfigurationPtrOutput) ToJobCopyDestinationEncryptionConfigurationPtrOutput added in v3.3.0

func (o JobCopyDestinationEncryptionConfigurationPtrOutput) ToJobCopyDestinationEncryptionConfigurationPtrOutput() JobCopyDestinationEncryptionConfigurationPtrOutput

func (JobCopyDestinationEncryptionConfigurationPtrOutput) ToJobCopyDestinationEncryptionConfigurationPtrOutputWithContext added in v3.3.0

func (o JobCopyDestinationEncryptionConfigurationPtrOutput) ToJobCopyDestinationEncryptionConfigurationPtrOutputWithContext(ctx context.Context) JobCopyDestinationEncryptionConfigurationPtrOutput

type JobCopyDestinationTable added in v3.3.0

type JobCopyDestinationTable struct {
	// The ID of the dataset containing this model.
	DatasetId string `pulumi:"datasetId"`
	// The ID of the project containing this model.
	ProjectId string `pulumi:"projectId"`
	// The ID of the table.
	TableId string `pulumi:"tableId"`
}

type JobCopyDestinationTableArgs added in v3.3.0

type JobCopyDestinationTableArgs struct {
	// The ID of the dataset containing this model.
	DatasetId pulumi.StringInput `pulumi:"datasetId"`
	// The ID of the project containing this model.
	ProjectId pulumi.StringInput `pulumi:"projectId"`
	// The ID of the table.
	TableId pulumi.StringInput `pulumi:"tableId"`
}

func (JobCopyDestinationTableArgs) ElementType added in v3.3.0

func (JobCopyDestinationTableArgs) ToJobCopyDestinationTableOutput added in v3.3.0

func (i JobCopyDestinationTableArgs) ToJobCopyDestinationTableOutput() JobCopyDestinationTableOutput

func (JobCopyDestinationTableArgs) ToJobCopyDestinationTableOutputWithContext added in v3.3.0

func (i JobCopyDestinationTableArgs) ToJobCopyDestinationTableOutputWithContext(ctx context.Context) JobCopyDestinationTableOutput

func (JobCopyDestinationTableArgs) ToJobCopyDestinationTablePtrOutput added in v3.3.0

func (i JobCopyDestinationTableArgs) ToJobCopyDestinationTablePtrOutput() JobCopyDestinationTablePtrOutput

func (JobCopyDestinationTableArgs) ToJobCopyDestinationTablePtrOutputWithContext added in v3.3.0

func (i JobCopyDestinationTableArgs) ToJobCopyDestinationTablePtrOutputWithContext(ctx context.Context) JobCopyDestinationTablePtrOutput

type JobCopyDestinationTableInput added in v3.3.0

type JobCopyDestinationTableInput interface {
	pulumi.Input

	ToJobCopyDestinationTableOutput() JobCopyDestinationTableOutput
	ToJobCopyDestinationTableOutputWithContext(context.Context) JobCopyDestinationTableOutput
}

JobCopyDestinationTableInput is an input type that accepts JobCopyDestinationTableArgs and JobCopyDestinationTableOutput values. You can construct a concrete instance of `JobCopyDestinationTableInput` via:

JobCopyDestinationTableArgs{...}

type JobCopyDestinationTableOutput added in v3.3.0

type JobCopyDestinationTableOutput struct{ *pulumi.OutputState }

func (JobCopyDestinationTableOutput) DatasetId added in v3.3.0

The ID of the dataset containing this model.

func (JobCopyDestinationTableOutput) ElementType added in v3.3.0

func (JobCopyDestinationTableOutput) ProjectId added in v3.3.0

The ID of the project containing this model.

func (JobCopyDestinationTableOutput) TableId added in v3.3.0

The ID of the table.

func (JobCopyDestinationTableOutput) ToJobCopyDestinationTableOutput added in v3.3.0

func (o JobCopyDestinationTableOutput) ToJobCopyDestinationTableOutput() JobCopyDestinationTableOutput

func (JobCopyDestinationTableOutput) ToJobCopyDestinationTableOutputWithContext added in v3.3.0

func (o JobCopyDestinationTableOutput) ToJobCopyDestinationTableOutputWithContext(ctx context.Context) JobCopyDestinationTableOutput

func (JobCopyDestinationTableOutput) ToJobCopyDestinationTablePtrOutput added in v3.3.0

func (o JobCopyDestinationTableOutput) ToJobCopyDestinationTablePtrOutput() JobCopyDestinationTablePtrOutput

func (JobCopyDestinationTableOutput) ToJobCopyDestinationTablePtrOutputWithContext added in v3.3.0

func (o JobCopyDestinationTableOutput) ToJobCopyDestinationTablePtrOutputWithContext(ctx context.Context) JobCopyDestinationTablePtrOutput

type JobCopyDestinationTablePtrInput added in v3.3.0

type JobCopyDestinationTablePtrInput interface {
	pulumi.Input

	ToJobCopyDestinationTablePtrOutput() JobCopyDestinationTablePtrOutput
	ToJobCopyDestinationTablePtrOutputWithContext(context.Context) JobCopyDestinationTablePtrOutput
}

JobCopyDestinationTablePtrInput is an input type that accepts JobCopyDestinationTableArgs, JobCopyDestinationTablePtr and JobCopyDestinationTablePtrOutput values. You can construct a concrete instance of `JobCopyDestinationTablePtrInput` via:

		 JobCopyDestinationTableArgs{...}

 or:

		 nil

func JobCopyDestinationTablePtr added in v3.3.0

func JobCopyDestinationTablePtr(v *JobCopyDestinationTableArgs) JobCopyDestinationTablePtrInput

type JobCopyDestinationTablePtrOutput added in v3.3.0

type JobCopyDestinationTablePtrOutput struct{ *pulumi.OutputState }

func (JobCopyDestinationTablePtrOutput) DatasetId added in v3.3.0

The ID of the dataset containing this model.

func (JobCopyDestinationTablePtrOutput) Elem added in v3.3.0

func (JobCopyDestinationTablePtrOutput) ElementType added in v3.3.0

func (JobCopyDestinationTablePtrOutput) ProjectId added in v3.3.0

The ID of the project containing this model.

func (JobCopyDestinationTablePtrOutput) TableId added in v3.3.0

The ID of the table.

func (JobCopyDestinationTablePtrOutput) ToJobCopyDestinationTablePtrOutput added in v3.3.0

func (o JobCopyDestinationTablePtrOutput) ToJobCopyDestinationTablePtrOutput() JobCopyDestinationTablePtrOutput

func (JobCopyDestinationTablePtrOutput) ToJobCopyDestinationTablePtrOutputWithContext added in v3.3.0

func (o JobCopyDestinationTablePtrOutput) ToJobCopyDestinationTablePtrOutputWithContext(ctx context.Context) JobCopyDestinationTablePtrOutput

type JobCopyInput added in v3.3.0

type JobCopyInput interface {
	pulumi.Input

	ToJobCopyOutput() JobCopyOutput
	ToJobCopyOutputWithContext(context.Context) JobCopyOutput
}

JobCopyInput is an input type that accepts JobCopyArgs and JobCopyOutput values. You can construct a concrete instance of `JobCopyInput` via:

JobCopyArgs{...}

type JobCopyOutput added in v3.3.0

type JobCopyOutput struct{ *pulumi.OutputState }

func (JobCopyOutput) CreateDisposition added in v3.3.0

func (o JobCopyOutput) CreateDisposition() pulumi.StringPtrOutput

Specifies whether the job is allowed to create new tables. The following values are supported: CREATE_IF_NEEDED: If the table does not exist, BigQuery creates the table. CREATE_NEVER: The table must already exist. If it does not, a 'notFound' error is returned in the job result. Creation, truncation and append actions occur as one atomic update upon job completion

func (JobCopyOutput) DestinationEncryptionConfiguration added in v3.3.0

func (o JobCopyOutput) DestinationEncryptionConfiguration() JobCopyDestinationEncryptionConfigurationPtrOutput

Custom encryption configuration (e.g., Cloud KMS keys) Structure is documented below.

func (JobCopyOutput) DestinationTable added in v3.3.0

func (o JobCopyOutput) DestinationTable() JobCopyDestinationTablePtrOutput

The destination table. Structure is documented below.

func (JobCopyOutput) ElementType added in v3.3.0

func (JobCopyOutput) ElementType() reflect.Type

func (JobCopyOutput) SourceTables added in v3.3.0

Source tables to copy. Structure is documented below.

func (JobCopyOutput) ToJobCopyOutput added in v3.3.0

func (o JobCopyOutput) ToJobCopyOutput() JobCopyOutput

func (JobCopyOutput) ToJobCopyOutputWithContext added in v3.3.0

func (o JobCopyOutput) ToJobCopyOutputWithContext(ctx context.Context) JobCopyOutput

func (JobCopyOutput) ToJobCopyPtrOutput added in v3.3.0

func (o JobCopyOutput) ToJobCopyPtrOutput() JobCopyPtrOutput

func (JobCopyOutput) ToJobCopyPtrOutputWithContext added in v3.3.0

func (o JobCopyOutput) ToJobCopyPtrOutputWithContext(ctx context.Context) JobCopyPtrOutput

func (JobCopyOutput) WriteDisposition added in v3.3.0

func (o JobCopyOutput) WriteDisposition() pulumi.StringPtrOutput

Specifies the action that occurs if the destination table already exists. The following values are supported: WRITE_TRUNCATE: If the table already exists, BigQuery overwrites the table data and uses the schema from the query result. WRITE_APPEND: If the table already exists, BigQuery appends the data to the table. WRITE_EMPTY: If the table already exists and contains data, a 'duplicate' error is returned in the job result. Each action is atomic and only occurs if BigQuery is able to complete the job successfully. Creation, truncation and append actions occur as one atomic update upon job completion.

type JobCopyPtrInput added in v3.3.0

type JobCopyPtrInput interface {
	pulumi.Input

	ToJobCopyPtrOutput() JobCopyPtrOutput
	ToJobCopyPtrOutputWithContext(context.Context) JobCopyPtrOutput
}

JobCopyPtrInput is an input type that accepts JobCopyArgs, JobCopyPtr and JobCopyPtrOutput values. You can construct a concrete instance of `JobCopyPtrInput` via:

		 JobCopyArgs{...}

 or:

		 nil

func JobCopyPtr added in v3.3.0

func JobCopyPtr(v *JobCopyArgs) JobCopyPtrInput

type JobCopyPtrOutput added in v3.3.0

type JobCopyPtrOutput struct{ *pulumi.OutputState }

func (JobCopyPtrOutput) CreateDisposition added in v3.3.0

func (o JobCopyPtrOutput) CreateDisposition() pulumi.StringPtrOutput

Specifies whether the job is allowed to create new tables. The following values are supported: CREATE_IF_NEEDED: If the table does not exist, BigQuery creates the table. CREATE_NEVER: The table must already exist. If it does not, a 'notFound' error is returned in the job result. Creation, truncation and append actions occur as one atomic update upon job completion

func (JobCopyPtrOutput) DestinationEncryptionConfiguration added in v3.3.0

func (o JobCopyPtrOutput) DestinationEncryptionConfiguration() JobCopyDestinationEncryptionConfigurationPtrOutput

Custom encryption configuration (e.g., Cloud KMS keys) Structure is documented below.

func (JobCopyPtrOutput) DestinationTable added in v3.3.0

The destination table. Structure is documented below.

func (JobCopyPtrOutput) Elem added in v3.3.0

func (JobCopyPtrOutput) ElementType added in v3.3.0

func (JobCopyPtrOutput) ElementType() reflect.Type

func (JobCopyPtrOutput) SourceTables added in v3.3.0

Source tables to copy. Structure is documented below.

func (JobCopyPtrOutput) ToJobCopyPtrOutput added in v3.3.0

func (o JobCopyPtrOutput) ToJobCopyPtrOutput() JobCopyPtrOutput

func (JobCopyPtrOutput) ToJobCopyPtrOutputWithContext added in v3.3.0

func (o JobCopyPtrOutput) ToJobCopyPtrOutputWithContext(ctx context.Context) JobCopyPtrOutput

func (JobCopyPtrOutput) WriteDisposition added in v3.3.0

func (o JobCopyPtrOutput) WriteDisposition() pulumi.StringPtrOutput

Specifies the action that occurs if the destination table already exists. The following values are supported: WRITE_TRUNCATE: If the table already exists, BigQuery overwrites the table data and uses the schema from the query result. WRITE_APPEND: If the table already exists, BigQuery appends the data to the table. WRITE_EMPTY: If the table already exists and contains data, a 'duplicate' error is returned in the job result. Each action is atomic and only occurs if BigQuery is able to complete the job successfully. Creation, truncation and append actions occur as one atomic update upon job completion.

type JobCopySourceTable added in v3.3.0

type JobCopySourceTable struct {
	// The ID of the dataset containing this model.
	DatasetId string `pulumi:"datasetId"`
	// The ID of the project containing this model.
	ProjectId string `pulumi:"projectId"`
	// The ID of the table.
	TableId string `pulumi:"tableId"`
}

type JobCopySourceTableArgs added in v3.3.0

type JobCopySourceTableArgs struct {
	// The ID of the dataset containing this model.
	DatasetId pulumi.StringInput `pulumi:"datasetId"`
	// The ID of the project containing this model.
	ProjectId pulumi.StringInput `pulumi:"projectId"`
	// The ID of the table.
	TableId pulumi.StringInput `pulumi:"tableId"`
}

func (JobCopySourceTableArgs) ElementType added in v3.3.0

func (JobCopySourceTableArgs) ElementType() reflect.Type

func (JobCopySourceTableArgs) ToJobCopySourceTableOutput added in v3.3.0

func (i JobCopySourceTableArgs) ToJobCopySourceTableOutput() JobCopySourceTableOutput

func (JobCopySourceTableArgs) ToJobCopySourceTableOutputWithContext added in v3.3.0

func (i JobCopySourceTableArgs) ToJobCopySourceTableOutputWithContext(ctx context.Context) JobCopySourceTableOutput

type JobCopySourceTableArray added in v3.3.0

type JobCopySourceTableArray []JobCopySourceTableInput

func (JobCopySourceTableArray) ElementType added in v3.3.0

func (JobCopySourceTableArray) ElementType() reflect.Type

func (JobCopySourceTableArray) ToJobCopySourceTableArrayOutput added in v3.3.0

func (i JobCopySourceTableArray) ToJobCopySourceTableArrayOutput() JobCopySourceTableArrayOutput

func (JobCopySourceTableArray) ToJobCopySourceTableArrayOutputWithContext added in v3.3.0

func (i JobCopySourceTableArray) ToJobCopySourceTableArrayOutputWithContext(ctx context.Context) JobCopySourceTableArrayOutput

type JobCopySourceTableArrayInput added in v3.3.0

type JobCopySourceTableArrayInput interface {
	pulumi.Input

	ToJobCopySourceTableArrayOutput() JobCopySourceTableArrayOutput
	ToJobCopySourceTableArrayOutputWithContext(context.Context) JobCopySourceTableArrayOutput
}

JobCopySourceTableArrayInput is an input type that accepts JobCopySourceTableArray and JobCopySourceTableArrayOutput values. You can construct a concrete instance of `JobCopySourceTableArrayInput` via:

JobCopySourceTableArray{ JobCopySourceTableArgs{...} }

type JobCopySourceTableArrayOutput added in v3.3.0

type JobCopySourceTableArrayOutput struct{ *pulumi.OutputState }

func (JobCopySourceTableArrayOutput) ElementType added in v3.3.0

func (JobCopySourceTableArrayOutput) Index added in v3.3.0

func (JobCopySourceTableArrayOutput) ToJobCopySourceTableArrayOutput added in v3.3.0

func (o JobCopySourceTableArrayOutput) ToJobCopySourceTableArrayOutput() JobCopySourceTableArrayOutput

func (JobCopySourceTableArrayOutput) ToJobCopySourceTableArrayOutputWithContext added in v3.3.0

func (o JobCopySourceTableArrayOutput) ToJobCopySourceTableArrayOutputWithContext(ctx context.Context) JobCopySourceTableArrayOutput

type JobCopySourceTableInput added in v3.3.0

type JobCopySourceTableInput interface {
	pulumi.Input

	ToJobCopySourceTableOutput() JobCopySourceTableOutput
	ToJobCopySourceTableOutputWithContext(context.Context) JobCopySourceTableOutput
}

JobCopySourceTableInput is an input type that accepts JobCopySourceTableArgs and JobCopySourceTableOutput values. You can construct a concrete instance of `JobCopySourceTableInput` via:

JobCopySourceTableArgs{...}

type JobCopySourceTableOutput added in v3.3.0

type JobCopySourceTableOutput struct{ *pulumi.OutputState }

func (JobCopySourceTableOutput) DatasetId added in v3.3.0

The ID of the dataset containing this model.

func (JobCopySourceTableOutput) ElementType added in v3.3.0

func (JobCopySourceTableOutput) ElementType() reflect.Type

func (JobCopySourceTableOutput) ProjectId added in v3.3.0

The ID of the project containing this model.

func (JobCopySourceTableOutput) TableId added in v3.3.0

The ID of the table.

func (JobCopySourceTableOutput) ToJobCopySourceTableOutput added in v3.3.0

func (o JobCopySourceTableOutput) ToJobCopySourceTableOutput() JobCopySourceTableOutput

func (JobCopySourceTableOutput) ToJobCopySourceTableOutputWithContext added in v3.3.0

func (o JobCopySourceTableOutput) ToJobCopySourceTableOutputWithContext(ctx context.Context) JobCopySourceTableOutput

type JobExtract added in v3.3.0

type JobExtract struct {
	// The compression type to use for exported files. Possible values include GZIP, DEFLATE, SNAPPY, and NONE.
	// The default value is NONE. DEFLATE and SNAPPY are only supported for Avro.
	Compression *string `pulumi:"compression"`
	// The exported file format. Possible values include CSV, NEWLINE_DELIMITED_JSON and AVRO for tables and SAVED_MODEL for models.
	// The default value for tables is CSV. Tables with nested or repeated fields cannot be exported as CSV.
	// The default value for models is SAVED_MODEL.
	DestinationFormat *string `pulumi:"destinationFormat"`
	// A list of fully-qualified Google Cloud Storage URIs where the extracted table should be written.
	DestinationUris []string `pulumi:"destinationUris"`
	// When extracting data in CSV format, this defines the delimiter to use between fields in the exported data.
	// Default is ','
	FieldDelimiter *string `pulumi:"fieldDelimiter"`
	// Whether to print out a header row in the results. Default is true.
	PrintHeader *bool `pulumi:"printHeader"`
	// A reference to the model being exported.  Structure is documented below.
	SourceModel *JobExtractSourceModel `pulumi:"sourceModel"`
	// A reference to the table being exported.  Structure is documented below.
	SourceTable *JobExtractSourceTable `pulumi:"sourceTable"`
	// Whether to use logical types when extracting to AVRO format.
	UseAvroLogicalTypes *bool `pulumi:"useAvroLogicalTypes"`
}

type JobExtractArgs added in v3.3.0

type JobExtractArgs struct {
	// The compression type to use for exported files. Possible values include GZIP, DEFLATE, SNAPPY, and NONE.
	// The default value is NONE. DEFLATE and SNAPPY are only supported for Avro.
	Compression pulumi.StringPtrInput `pulumi:"compression"`
	// The exported file format. Possible values include CSV, NEWLINE_DELIMITED_JSON and AVRO for tables and SAVED_MODEL for models.
	// The default value for tables is CSV. Tables with nested or repeated fields cannot be exported as CSV.
	// The default value for models is SAVED_MODEL.
	DestinationFormat pulumi.StringPtrInput `pulumi:"destinationFormat"`
	// A list of fully-qualified Google Cloud Storage URIs where the extracted table should be written.
	DestinationUris pulumi.StringArrayInput `pulumi:"destinationUris"`
	// When extracting data in CSV format, this defines the delimiter to use between fields in the exported data.
	// Default is ','
	FieldDelimiter pulumi.StringPtrInput `pulumi:"fieldDelimiter"`
	// Whether to print out a header row in the results. Default is true.
	PrintHeader pulumi.BoolPtrInput `pulumi:"printHeader"`
	// A reference to the model being exported.  Structure is documented below.
	SourceModel JobExtractSourceModelPtrInput `pulumi:"sourceModel"`
	// A reference to the table being exported.  Structure is documented below.
	SourceTable JobExtractSourceTablePtrInput `pulumi:"sourceTable"`
	// Whether to use logical types when extracting to AVRO format.
	UseAvroLogicalTypes pulumi.BoolPtrInput `pulumi:"useAvroLogicalTypes"`
}

func (JobExtractArgs) ElementType added in v3.3.0

func (JobExtractArgs) ElementType() reflect.Type

func (JobExtractArgs) ToJobExtractOutput added in v3.3.0

func (i JobExtractArgs) ToJobExtractOutput() JobExtractOutput

func (JobExtractArgs) ToJobExtractOutputWithContext added in v3.3.0

func (i JobExtractArgs) ToJobExtractOutputWithContext(ctx context.Context) JobExtractOutput

func (JobExtractArgs) ToJobExtractPtrOutput added in v3.3.0

func (i JobExtractArgs) ToJobExtractPtrOutput() JobExtractPtrOutput

func (JobExtractArgs) ToJobExtractPtrOutputWithContext added in v3.3.0

func (i JobExtractArgs) ToJobExtractPtrOutputWithContext(ctx context.Context) JobExtractPtrOutput

type JobExtractInput added in v3.3.0

type JobExtractInput interface {
	pulumi.Input

	ToJobExtractOutput() JobExtractOutput
	ToJobExtractOutputWithContext(context.Context) JobExtractOutput
}

JobExtractInput is an input type that accepts JobExtractArgs and JobExtractOutput values. You can construct a concrete instance of `JobExtractInput` via:

JobExtractArgs{...}

type JobExtractOutput added in v3.3.0

type JobExtractOutput struct{ *pulumi.OutputState }

func (JobExtractOutput) Compression added in v3.3.0

func (o JobExtractOutput) Compression() pulumi.StringPtrOutput

The compression type to use for exported files. Possible values include GZIP, DEFLATE, SNAPPY, and NONE. The default value is NONE. DEFLATE and SNAPPY are only supported for Avro.

func (JobExtractOutput) DestinationFormat added in v3.3.0

func (o JobExtractOutput) DestinationFormat() pulumi.StringPtrOutput

The exported file format. Possible values include CSV, NEWLINE_DELIMITED_JSON and AVRO for tables and SAVED_MODEL for models. The default value for tables is CSV. Tables with nested or repeated fields cannot be exported as CSV. The default value for models is SAVED_MODEL.

func (JobExtractOutput) DestinationUris added in v3.3.0

func (o JobExtractOutput) DestinationUris() pulumi.StringArrayOutput

A list of fully-qualified Google Cloud Storage URIs where the extracted table should be written.

func (JobExtractOutput) ElementType added in v3.3.0

func (JobExtractOutput) ElementType() reflect.Type

func (JobExtractOutput) FieldDelimiter added in v3.3.0

func (o JobExtractOutput) FieldDelimiter() pulumi.StringPtrOutput

When extracting data in CSV format, this defines the delimiter to use between fields in the exported data. Default is ','

func (JobExtractOutput) PrintHeader added in v3.3.0

func (o JobExtractOutput) PrintHeader() pulumi.BoolPtrOutput

Whether to print out a header row in the results. Default is true.

func (JobExtractOutput) SourceModel added in v3.3.0

A reference to the model being exported. Structure is documented below.

func (JobExtractOutput) SourceTable added in v3.3.0

A reference to the table being exported. Structure is documented below.

func (JobExtractOutput) ToJobExtractOutput added in v3.3.0

func (o JobExtractOutput) ToJobExtractOutput() JobExtractOutput

func (JobExtractOutput) ToJobExtractOutputWithContext added in v3.3.0

func (o JobExtractOutput) ToJobExtractOutputWithContext(ctx context.Context) JobExtractOutput

func (JobExtractOutput) ToJobExtractPtrOutput added in v3.3.0

func (o JobExtractOutput) ToJobExtractPtrOutput() JobExtractPtrOutput

func (JobExtractOutput) ToJobExtractPtrOutputWithContext added in v3.3.0

func (o JobExtractOutput) ToJobExtractPtrOutputWithContext(ctx context.Context) JobExtractPtrOutput

func (JobExtractOutput) UseAvroLogicalTypes added in v3.3.0

func (o JobExtractOutput) UseAvroLogicalTypes() pulumi.BoolPtrOutput

Whether to use logical types when extracting to AVRO format.

type JobExtractPtrInput added in v3.3.0

type JobExtractPtrInput interface {
	pulumi.Input

	ToJobExtractPtrOutput() JobExtractPtrOutput
	ToJobExtractPtrOutputWithContext(context.Context) JobExtractPtrOutput
}

JobExtractPtrInput is an input type that accepts JobExtractArgs, JobExtractPtr and JobExtractPtrOutput values. You can construct a concrete instance of `JobExtractPtrInput` via:

		 JobExtractArgs{...}

 or:

		 nil

func JobExtractPtr added in v3.3.0

func JobExtractPtr(v *JobExtractArgs) JobExtractPtrInput

type JobExtractPtrOutput added in v3.3.0

type JobExtractPtrOutput struct{ *pulumi.OutputState }

func (JobExtractPtrOutput) Compression added in v3.3.0

func (o JobExtractPtrOutput) Compression() pulumi.StringPtrOutput

The compression type to use for exported files. Possible values include GZIP, DEFLATE, SNAPPY, and NONE. The default value is NONE. DEFLATE and SNAPPY are only supported for Avro.

func (JobExtractPtrOutput) DestinationFormat added in v3.3.0

func (o JobExtractPtrOutput) DestinationFormat() pulumi.StringPtrOutput

The exported file format. Possible values include CSV, NEWLINE_DELIMITED_JSON and AVRO for tables and SAVED_MODEL for models. The default value for tables is CSV. Tables with nested or repeated fields cannot be exported as CSV. The default value for models is SAVED_MODEL.

func (JobExtractPtrOutput) DestinationUris added in v3.3.0

func (o JobExtractPtrOutput) DestinationUris() pulumi.StringArrayOutput

A list of fully-qualified Google Cloud Storage URIs where the extracted table should be written.

func (JobExtractPtrOutput) Elem added in v3.3.0

func (JobExtractPtrOutput) ElementType added in v3.3.0

func (JobExtractPtrOutput) ElementType() reflect.Type

func (JobExtractPtrOutput) FieldDelimiter added in v3.3.0

func (o JobExtractPtrOutput) FieldDelimiter() pulumi.StringPtrOutput

When extracting data in CSV format, this defines the delimiter to use between fields in the exported data. Default is ','

func (JobExtractPtrOutput) PrintHeader added in v3.3.0

func (o JobExtractPtrOutput) PrintHeader() pulumi.BoolPtrOutput

Whether to print out a header row in the results. Default is true.

func (JobExtractPtrOutput) SourceModel added in v3.3.0

A reference to the model being exported. Structure is documented below.

func (JobExtractPtrOutput) SourceTable added in v3.3.0

A reference to the table being exported. Structure is documented below.

func (JobExtractPtrOutput) ToJobExtractPtrOutput added in v3.3.0

func (o JobExtractPtrOutput) ToJobExtractPtrOutput() JobExtractPtrOutput

func (JobExtractPtrOutput) ToJobExtractPtrOutputWithContext added in v3.3.0

func (o JobExtractPtrOutput) ToJobExtractPtrOutputWithContext(ctx context.Context) JobExtractPtrOutput

func (JobExtractPtrOutput) UseAvroLogicalTypes added in v3.3.0

func (o JobExtractPtrOutput) UseAvroLogicalTypes() pulumi.BoolPtrOutput

Whether to use logical types when extracting to AVRO format.

type JobExtractSourceModel added in v3.3.0

type JobExtractSourceModel struct {
	// The ID of the dataset containing this model.
	DatasetId string `pulumi:"datasetId"`
	// The ID of the model.
	ModelId string `pulumi:"modelId"`
	// The ID of the project containing this model.
	ProjectId string `pulumi:"projectId"`
}

type JobExtractSourceModelArgs added in v3.3.0

type JobExtractSourceModelArgs struct {
	// The ID of the dataset containing this model.
	DatasetId pulumi.StringInput `pulumi:"datasetId"`
	// The ID of the model.
	ModelId pulumi.StringInput `pulumi:"modelId"`
	// The ID of the project containing this model.
	ProjectId pulumi.StringInput `pulumi:"projectId"`
}

func (JobExtractSourceModelArgs) ElementType added in v3.3.0

func (JobExtractSourceModelArgs) ElementType() reflect.Type

func (JobExtractSourceModelArgs) ToJobExtractSourceModelOutput added in v3.3.0

func (i JobExtractSourceModelArgs) ToJobExtractSourceModelOutput() JobExtractSourceModelOutput

func (JobExtractSourceModelArgs) ToJobExtractSourceModelOutputWithContext added in v3.3.0

func (i JobExtractSourceModelArgs) ToJobExtractSourceModelOutputWithContext(ctx context.Context) JobExtractSourceModelOutput

func (JobExtractSourceModelArgs) ToJobExtractSourceModelPtrOutput added in v3.3.0

func (i JobExtractSourceModelArgs) ToJobExtractSourceModelPtrOutput() JobExtractSourceModelPtrOutput

func (JobExtractSourceModelArgs) ToJobExtractSourceModelPtrOutputWithContext added in v3.3.0

func (i JobExtractSourceModelArgs) ToJobExtractSourceModelPtrOutputWithContext(ctx context.Context) JobExtractSourceModelPtrOutput

type JobExtractSourceModelInput added in v3.3.0

type JobExtractSourceModelInput interface {
	pulumi.Input

	ToJobExtractSourceModelOutput() JobExtractSourceModelOutput
	ToJobExtractSourceModelOutputWithContext(context.Context) JobExtractSourceModelOutput
}

JobExtractSourceModelInput is an input type that accepts JobExtractSourceModelArgs and JobExtractSourceModelOutput values. You can construct a concrete instance of `JobExtractSourceModelInput` via:

JobExtractSourceModelArgs{...}

type JobExtractSourceModelOutput added in v3.3.0

type JobExtractSourceModelOutput struct{ *pulumi.OutputState }

func (JobExtractSourceModelOutput) DatasetId added in v3.3.0

The ID of the dataset containing this model.

func (JobExtractSourceModelOutput) ElementType added in v3.3.0

func (JobExtractSourceModelOutput) ModelId added in v3.3.0

The ID of the model.

func (JobExtractSourceModelOutput) ProjectId added in v3.3.0

The ID of the project containing this model.

func (JobExtractSourceModelOutput) ToJobExtractSourceModelOutput added in v3.3.0

func (o JobExtractSourceModelOutput) ToJobExtractSourceModelOutput() JobExtractSourceModelOutput

func (JobExtractSourceModelOutput) ToJobExtractSourceModelOutputWithContext added in v3.3.0

func (o JobExtractSourceModelOutput) ToJobExtractSourceModelOutputWithContext(ctx context.Context) JobExtractSourceModelOutput

func (JobExtractSourceModelOutput) ToJobExtractSourceModelPtrOutput added in v3.3.0

func (o JobExtractSourceModelOutput) ToJobExtractSourceModelPtrOutput() JobExtractSourceModelPtrOutput

func (JobExtractSourceModelOutput) ToJobExtractSourceModelPtrOutputWithContext added in v3.3.0

func (o JobExtractSourceModelOutput) ToJobExtractSourceModelPtrOutputWithContext(ctx context.Context) JobExtractSourceModelPtrOutput

type JobExtractSourceModelPtrInput added in v3.3.0

type JobExtractSourceModelPtrInput interface {
	pulumi.Input

	ToJobExtractSourceModelPtrOutput() JobExtractSourceModelPtrOutput
	ToJobExtractSourceModelPtrOutputWithContext(context.Context) JobExtractSourceModelPtrOutput
}

JobExtractSourceModelPtrInput is an input type that accepts JobExtractSourceModelArgs, JobExtractSourceModelPtr and JobExtractSourceModelPtrOutput values. You can construct a concrete instance of `JobExtractSourceModelPtrInput` via:

		 JobExtractSourceModelArgs{...}

 or:

		 nil

func JobExtractSourceModelPtr added in v3.3.0

func JobExtractSourceModelPtr(v *JobExtractSourceModelArgs) JobExtractSourceModelPtrInput

type JobExtractSourceModelPtrOutput added in v3.3.0

type JobExtractSourceModelPtrOutput struct{ *pulumi.OutputState }

func (JobExtractSourceModelPtrOutput) DatasetId added in v3.3.0

The ID of the dataset containing this model.

func (JobExtractSourceModelPtrOutput) Elem added in v3.3.0

func (JobExtractSourceModelPtrOutput) ElementType added in v3.3.0

func (JobExtractSourceModelPtrOutput) ModelId added in v3.3.0

The ID of the model.

func (JobExtractSourceModelPtrOutput) ProjectId added in v3.3.0

The ID of the project containing this model.

func (JobExtractSourceModelPtrOutput) ToJobExtractSourceModelPtrOutput added in v3.3.0

func (o JobExtractSourceModelPtrOutput) ToJobExtractSourceModelPtrOutput() JobExtractSourceModelPtrOutput

func (JobExtractSourceModelPtrOutput) ToJobExtractSourceModelPtrOutputWithContext added in v3.3.0

func (o JobExtractSourceModelPtrOutput) ToJobExtractSourceModelPtrOutputWithContext(ctx context.Context) JobExtractSourceModelPtrOutput

type JobExtractSourceTable added in v3.3.0

type JobExtractSourceTable struct {
	// The ID of the dataset containing this model.
	DatasetId string `pulumi:"datasetId"`
	// The ID of the project containing this model.
	ProjectId string `pulumi:"projectId"`
	// The ID of the table.
	TableId string `pulumi:"tableId"`
}

type JobExtractSourceTableArgs added in v3.3.0

type JobExtractSourceTableArgs struct {
	// The ID of the dataset containing this model.
	DatasetId pulumi.StringInput `pulumi:"datasetId"`
	// The ID of the project containing this model.
	ProjectId pulumi.StringInput `pulumi:"projectId"`
	// The ID of the table.
	TableId pulumi.StringInput `pulumi:"tableId"`
}

func (JobExtractSourceTableArgs) ElementType added in v3.3.0

func (JobExtractSourceTableArgs) ElementType() reflect.Type

func (JobExtractSourceTableArgs) ToJobExtractSourceTableOutput added in v3.3.0

func (i JobExtractSourceTableArgs) ToJobExtractSourceTableOutput() JobExtractSourceTableOutput

func (JobExtractSourceTableArgs) ToJobExtractSourceTableOutputWithContext added in v3.3.0

func (i JobExtractSourceTableArgs) ToJobExtractSourceTableOutputWithContext(ctx context.Context) JobExtractSourceTableOutput

func (JobExtractSourceTableArgs) ToJobExtractSourceTablePtrOutput added in v3.3.0

func (i JobExtractSourceTableArgs) ToJobExtractSourceTablePtrOutput() JobExtractSourceTablePtrOutput

func (JobExtractSourceTableArgs) ToJobExtractSourceTablePtrOutputWithContext added in v3.3.0

func (i JobExtractSourceTableArgs) ToJobExtractSourceTablePtrOutputWithContext(ctx context.Context) JobExtractSourceTablePtrOutput

type JobExtractSourceTableInput added in v3.3.0

type JobExtractSourceTableInput interface {
	pulumi.Input

	ToJobExtractSourceTableOutput() JobExtractSourceTableOutput
	ToJobExtractSourceTableOutputWithContext(context.Context) JobExtractSourceTableOutput
}

JobExtractSourceTableInput is an input type that accepts JobExtractSourceTableArgs and JobExtractSourceTableOutput values. You can construct a concrete instance of `JobExtractSourceTableInput` via:

JobExtractSourceTableArgs{...}

type JobExtractSourceTableOutput added in v3.3.0

type JobExtractSourceTableOutput struct{ *pulumi.OutputState }

func (JobExtractSourceTableOutput) DatasetId added in v3.3.0

The ID of the dataset containing this model.

func (JobExtractSourceTableOutput) ElementType added in v3.3.0

func (JobExtractSourceTableOutput) ProjectId added in v3.3.0

The ID of the project containing this model.

func (JobExtractSourceTableOutput) TableId added in v3.3.0

The ID of the table.

func (JobExtractSourceTableOutput) ToJobExtractSourceTableOutput added in v3.3.0

func (o JobExtractSourceTableOutput) ToJobExtractSourceTableOutput() JobExtractSourceTableOutput

func (JobExtractSourceTableOutput) ToJobExtractSourceTableOutputWithContext added in v3.3.0

func (o JobExtractSourceTableOutput) ToJobExtractSourceTableOutputWithContext(ctx context.Context) JobExtractSourceTableOutput

func (JobExtractSourceTableOutput) ToJobExtractSourceTablePtrOutput added in v3.3.0

func (o JobExtractSourceTableOutput) ToJobExtractSourceTablePtrOutput() JobExtractSourceTablePtrOutput

func (JobExtractSourceTableOutput) ToJobExtractSourceTablePtrOutputWithContext added in v3.3.0

func (o JobExtractSourceTableOutput) ToJobExtractSourceTablePtrOutputWithContext(ctx context.Context) JobExtractSourceTablePtrOutput

type JobExtractSourceTablePtrInput added in v3.3.0

type JobExtractSourceTablePtrInput interface {
	pulumi.Input

	ToJobExtractSourceTablePtrOutput() JobExtractSourceTablePtrOutput
	ToJobExtractSourceTablePtrOutputWithContext(context.Context) JobExtractSourceTablePtrOutput
}

JobExtractSourceTablePtrInput is an input type that accepts JobExtractSourceTableArgs, JobExtractSourceTablePtr and JobExtractSourceTablePtrOutput values. You can construct a concrete instance of `JobExtractSourceTablePtrInput` via:

		 JobExtractSourceTableArgs{...}

 or:

		 nil

func JobExtractSourceTablePtr added in v3.3.0

func JobExtractSourceTablePtr(v *JobExtractSourceTableArgs) JobExtractSourceTablePtrInput

type JobExtractSourceTablePtrOutput added in v3.3.0

type JobExtractSourceTablePtrOutput struct{ *pulumi.OutputState }

func (JobExtractSourceTablePtrOutput) DatasetId added in v3.3.0

The ID of the dataset containing this model.

func (JobExtractSourceTablePtrOutput) Elem added in v3.3.0

func (JobExtractSourceTablePtrOutput) ElementType added in v3.3.0

func (JobExtractSourceTablePtrOutput) ProjectId added in v3.3.0

The ID of the project containing this model.

func (JobExtractSourceTablePtrOutput) TableId added in v3.3.0

The ID of the table.

func (JobExtractSourceTablePtrOutput) ToJobExtractSourceTablePtrOutput added in v3.3.0

func (o JobExtractSourceTablePtrOutput) ToJobExtractSourceTablePtrOutput() JobExtractSourceTablePtrOutput

func (JobExtractSourceTablePtrOutput) ToJobExtractSourceTablePtrOutputWithContext added in v3.3.0

func (o JobExtractSourceTablePtrOutput) ToJobExtractSourceTablePtrOutputWithContext(ctx context.Context) JobExtractSourceTablePtrOutput

type JobLoad added in v3.3.0

type JobLoad struct {
	// Accept rows that are missing trailing optional columns. The missing values are treated as nulls.
	// If false, records with missing trailing columns are treated as bad records, and if there are too many bad records,
	// an invalid error is returned in the job result. The default value is false. Only applicable to CSV, ignored for other formats.
	AllowJaggedRows *bool `pulumi:"allowJaggedRows"`
	// Indicates if BigQuery should allow quoted data sections that contain newline characters in a CSV file.
	// The default value is false.
	AllowQuotedNewlines *bool `pulumi:"allowQuotedNewlines"`
	// Indicates if we should automatically infer the options and schema for CSV and JSON sources.
	Autodetect *bool `pulumi:"autodetect"`
	// Specifies whether the job is allowed to create new tables. The following values are supported:
	// CREATE_IF_NEEDED: If the table does not exist, BigQuery creates the table.
	// CREATE_NEVER: The table must already exist. If it does not, a 'notFound' error is returned in the job result.
	// Creation, truncation and append actions occur as one atomic update upon job completion
	CreateDisposition *string `pulumi:"createDisposition"`
	// Custom encryption configuration (e.g., Cloud KMS keys)  Structure is documented below.
	DestinationEncryptionConfiguration *JobLoadDestinationEncryptionConfiguration `pulumi:"destinationEncryptionConfiguration"`
	// The destination table.  Structure is documented below.
	DestinationTable JobLoadDestinationTable `pulumi:"destinationTable"`
	// The character encoding of the data. The supported values are UTF-8 or ISO-8859-1.
	// The default value is UTF-8. BigQuery decodes the data after the raw, binary data
	// has been split using the values of the quote and fieldDelimiter properties.
	Encoding *string `pulumi:"encoding"`
	// When extracting data in CSV format, this defines the delimiter to use between fields in the exported data.
	// Default is ','
	FieldDelimiter *string `pulumi:"fieldDelimiter"`
	// Indicates if BigQuery should allow extra values that are not represented in the table schema.
	// If true, the extra values are ignored. If false, records with extra columns are treated as bad records,
	// and if there are too many bad records, an invalid error is returned in the job result.
	// The default value is false. The sourceFormat property determines what BigQuery treats as an extra value:
	// CSV: Trailing columns
	// JSON: Named values that don't match any column names
	IgnoreUnknownValues *bool `pulumi:"ignoreUnknownValues"`
	// The maximum number of bad records that BigQuery can ignore when running the job. If the number of bad records exceeds this value,
	// an invalid error is returned in the job result. The default value is 0, which requires that all records are valid.
	MaxBadRecords *int `pulumi:"maxBadRecords"`
	// Specifies a string that represents a null value in a CSV file. The default value is the empty string. If you set this
	// property to a custom value, BigQuery throws an error if an
	// empty string is present for all data types except for STRING and BYTE. For STRING and BYTE columns, BigQuery interprets the empty string as
	// an empty value.
	NullMarker *string `pulumi:"nullMarker"`
	// If sourceFormat is set to "DATASTORE_BACKUP", indicates which entity properties to load into BigQuery from a Cloud Datastore backup.
	// Property names are case sensitive and must be top-level properties. If no properties are specified, BigQuery loads all properties.
	// If any named property isn't found in the Cloud Datastore backup, an invalid error is returned in the job result.
	ProjectionFields []string `pulumi:"projectionFields"`
	// The value that is used to quote data sections in a CSV file. BigQuery converts the string to ISO-8859-1 encoding,
	// and then uses the first byte of the encoded string to split the data in its raw, binary state.
	// The default value is a double-quote ('"'). If your data does not contain quoted sections, set the property value to an empty string.
	// If your data contains quoted newline characters, you must also set the allowQuotedNewlines property to true.
	Quote *string `pulumi:"quote"`
	// Allows the schema of the destination table to be updated as a side effect of the load job if a schema is autodetected or
	// supplied in the job configuration. Schema update options are supported in two cases: when writeDisposition is WRITE_APPEND;
	// when writeDisposition is WRITE_TRUNCATE and the destination table is a partition of a table, specified by partition decorators.
	// For normal tables, WRITE_TRUNCATE will always overwrite the schema. One or more of the following values are specified:
	// ALLOW_FIELD_ADDITION: allow adding a nullable field to the schema.
	// ALLOW_FIELD_RELAXATION: allow relaxing a required field in the original schema to nullable.
	SchemaUpdateOptions []string `pulumi:"schemaUpdateOptions"`
	// The number of rows at the top of a CSV file that BigQuery will skip when loading the data.
	// The default value is 0. This property is useful if you have header rows in the file that should be skipped.
	// When autodetect is on, the behavior is the following:
	// skipLeadingRows unspecified - Autodetect tries to detect headers in the first row. If they are not detected,
	// the row is read as data. Otherwise data is read starting from the second row.
	// skipLeadingRows is 0 - Instructs autodetect that there are no headers and data should be read starting from the first row.
	// skipLeadingRows = N > 0 - Autodetect skips N-1 rows and tries to detect headers in row N. If headers are not detected,
	// row N is just skipped. Otherwise row N is used to extract column names for the detected schema.
	SkipLeadingRows *int `pulumi:"skipLeadingRows"`
	// The format of the data files. For CSV files, specify "CSV". For datastore backups, specify "DATASTORE_BACKUP".
	// For newline-delimited JSON, specify "NEWLINE_DELIMITED_JSON". For Avro, specify "AVRO". For parquet, specify "PARQUET".
	// For orc, specify "ORC". The default value is CSV.
	SourceFormat *string `pulumi:"sourceFormat"`
	// The fully-qualified URIs that point to your data in Google Cloud.
	// For Google Cloud Storage URIs: Each URI can contain one '*' wildcard character
	// and it must come after the 'bucket' name. Size limits related to load jobs apply
	// to external data sources. For Google Cloud Bigtable URIs: Exactly one URI can be
	// specified and it has be a fully specified and valid HTTPS URL for a Google Cloud Bigtable table.
	// For Google Cloud Datastore backups: Exactly one URI can be specified. Also, the '*' wildcard character is not allowed.
	SourceUris []string `pulumi:"sourceUris"`
	// Time-based partitioning specification for the destination table.  Structure is documented below.
	TimePartitioning *JobLoadTimePartitioning `pulumi:"timePartitioning"`
	// Specifies the action that occurs if the destination table already exists. The following values are supported:
	// WRITE_TRUNCATE: If the table already exists, BigQuery overwrites the table data and uses the schema from the query result.
	// WRITE_APPEND: If the table already exists, BigQuery appends the data to the table.
	// WRITE_EMPTY: If the table already exists and contains data, a 'duplicate' error is returned in the job result.
	// Each action is atomic and only occurs if BigQuery is able to complete the job successfully.
	// Creation, truncation and append actions occur as one atomic update upon job completion.
	WriteDisposition *string `pulumi:"writeDisposition"`
}

type JobLoadArgs added in v3.3.0

type JobLoadArgs struct {
	// Accept rows that are missing trailing optional columns. The missing values are treated as nulls.
	// If false, records with missing trailing columns are treated as bad records, and if there are too many bad records,
	// an invalid error is returned in the job result. The default value is false. Only applicable to CSV, ignored for other formats.
	AllowJaggedRows pulumi.BoolPtrInput `pulumi:"allowJaggedRows"`
	// Indicates if BigQuery should allow quoted data sections that contain newline characters in a CSV file.
	// The default value is false.
	AllowQuotedNewlines pulumi.BoolPtrInput `pulumi:"allowQuotedNewlines"`
	// Indicates if we should automatically infer the options and schema for CSV and JSON sources.
	Autodetect pulumi.BoolPtrInput `pulumi:"autodetect"`
	// Specifies whether the job is allowed to create new tables. The following values are supported:
	// CREATE_IF_NEEDED: If the table does not exist, BigQuery creates the table.
	// CREATE_NEVER: The table must already exist. If it does not, a 'notFound' error is returned in the job result.
	// Creation, truncation and append actions occur as one atomic update upon job completion
	CreateDisposition pulumi.StringPtrInput `pulumi:"createDisposition"`
	// Custom encryption configuration (e.g., Cloud KMS keys)  Structure is documented below.
	DestinationEncryptionConfiguration JobLoadDestinationEncryptionConfigurationPtrInput `pulumi:"destinationEncryptionConfiguration"`
	// The destination table.  Structure is documented below.
	DestinationTable JobLoadDestinationTableInput `pulumi:"destinationTable"`
	// The character encoding of the data. The supported values are UTF-8 or ISO-8859-1.
	// The default value is UTF-8. BigQuery decodes the data after the raw, binary data
	// has been split using the values of the quote and fieldDelimiter properties.
	Encoding pulumi.StringPtrInput `pulumi:"encoding"`
	// When extracting data in CSV format, this defines the delimiter to use between fields in the exported data.
	// Default is ','
	FieldDelimiter pulumi.StringPtrInput `pulumi:"fieldDelimiter"`
	// Indicates if BigQuery should allow extra values that are not represented in the table schema.
	// If true, the extra values are ignored. If false, records with extra columns are treated as bad records,
	// and if there are too many bad records, an invalid error is returned in the job result.
	// The default value is false. The sourceFormat property determines what BigQuery treats as an extra value:
	// CSV: Trailing columns
	// JSON: Named values that don't match any column names
	IgnoreUnknownValues pulumi.BoolPtrInput `pulumi:"ignoreUnknownValues"`
	// The maximum number of bad records that BigQuery can ignore when running the job. If the number of bad records exceeds this value,
	// an invalid error is returned in the job result. The default value is 0, which requires that all records are valid.
	MaxBadRecords pulumi.IntPtrInput `pulumi:"maxBadRecords"`
	// Specifies a string that represents a null value in a CSV file. The default value is the empty string. If you set this
	// property to a custom value, BigQuery throws an error if an
	// empty string is present for all data types except for STRING and BYTE. For STRING and BYTE columns, BigQuery interprets the empty string as
	// an empty value.
	NullMarker pulumi.StringPtrInput `pulumi:"nullMarker"`
	// If sourceFormat is set to "DATASTORE_BACKUP", indicates which entity properties to load into BigQuery from a Cloud Datastore backup.
	// Property names are case sensitive and must be top-level properties. If no properties are specified, BigQuery loads all properties.
	// If any named property isn't found in the Cloud Datastore backup, an invalid error is returned in the job result.
	ProjectionFields pulumi.StringArrayInput `pulumi:"projectionFields"`
	// The value that is used to quote data sections in a CSV file. BigQuery converts the string to ISO-8859-1 encoding,
	// and then uses the first byte of the encoded string to split the data in its raw, binary state.
	// The default value is a double-quote ('"'). If your data does not contain quoted sections, set the property value to an empty string.
	// If your data contains quoted newline characters, you must also set the allowQuotedNewlines property to true.
	Quote pulumi.StringPtrInput `pulumi:"quote"`
	// Allows the schema of the destination table to be updated as a side effect of the load job if a schema is autodetected or
	// supplied in the job configuration. Schema update options are supported in two cases: when writeDisposition is WRITE_APPEND;
	// when writeDisposition is WRITE_TRUNCATE and the destination table is a partition of a table, specified by partition decorators.
	// For normal tables, WRITE_TRUNCATE will always overwrite the schema. One or more of the following values are specified:
	// ALLOW_FIELD_ADDITION: allow adding a nullable field to the schema.
	// ALLOW_FIELD_RELAXATION: allow relaxing a required field in the original schema to nullable.
	SchemaUpdateOptions pulumi.StringArrayInput `pulumi:"schemaUpdateOptions"`
	// The number of rows at the top of a CSV file that BigQuery will skip when loading the data.
	// The default value is 0. This property is useful if you have header rows in the file that should be skipped.
	// When autodetect is on, the behavior is the following:
	// skipLeadingRows unspecified - Autodetect tries to detect headers in the first row. If they are not detected,
	// the row is read as data. Otherwise data is read starting from the second row.
	// skipLeadingRows is 0 - Instructs autodetect that there are no headers and data should be read starting from the first row.
	// skipLeadingRows = N > 0 - Autodetect skips N-1 rows and tries to detect headers in row N. If headers are not detected,
	// row N is just skipped. Otherwise row N is used to extract column names for the detected schema.
	SkipLeadingRows pulumi.IntPtrInput `pulumi:"skipLeadingRows"`
	// The format of the data files. For CSV files, specify "CSV". For datastore backups, specify "DATASTORE_BACKUP".
	// For newline-delimited JSON, specify "NEWLINE_DELIMITED_JSON". For Avro, specify "AVRO". For parquet, specify "PARQUET".
	// For orc, specify "ORC". The default value is CSV.
	SourceFormat pulumi.StringPtrInput `pulumi:"sourceFormat"`
	// The fully-qualified URIs that point to your data in Google Cloud.
	// For Google Cloud Storage URIs: Each URI can contain one '*' wildcard character
	// and it must come after the 'bucket' name. Size limits related to load jobs apply
	// to external data sources. For Google Cloud Bigtable URIs: Exactly one URI can be
	// specified and it has be a fully specified and valid HTTPS URL for a Google Cloud Bigtable table.
	// For Google Cloud Datastore backups: Exactly one URI can be specified. Also, the '*' wildcard character is not allowed.
	SourceUris pulumi.StringArrayInput `pulumi:"sourceUris"`
	// Time-based partitioning specification for the destination table.  Structure is documented below.
	TimePartitioning JobLoadTimePartitioningPtrInput `pulumi:"timePartitioning"`
	// Specifies the action that occurs if the destination table already exists. The following values are supported:
	// WRITE_TRUNCATE: If the table already exists, BigQuery overwrites the table data and uses the schema from the query result.
	// WRITE_APPEND: If the table already exists, BigQuery appends the data to the table.
	// WRITE_EMPTY: If the table already exists and contains data, a 'duplicate' error is returned in the job result.
	// Each action is atomic and only occurs if BigQuery is able to complete the job successfully.
	// Creation, truncation and append actions occur as one atomic update upon job completion.
	WriteDisposition pulumi.StringPtrInput `pulumi:"writeDisposition"`
}

func (JobLoadArgs) ElementType added in v3.3.0

func (JobLoadArgs) ElementType() reflect.Type

func (JobLoadArgs) ToJobLoadOutput added in v3.3.0

func (i JobLoadArgs) ToJobLoadOutput() JobLoadOutput

func (JobLoadArgs) ToJobLoadOutputWithContext added in v3.3.0

func (i JobLoadArgs) ToJobLoadOutputWithContext(ctx context.Context) JobLoadOutput

func (JobLoadArgs) ToJobLoadPtrOutput added in v3.3.0

func (i JobLoadArgs) ToJobLoadPtrOutput() JobLoadPtrOutput

func (JobLoadArgs) ToJobLoadPtrOutputWithContext added in v3.3.0

func (i JobLoadArgs) ToJobLoadPtrOutputWithContext(ctx context.Context) JobLoadPtrOutput

type JobLoadDestinationEncryptionConfiguration added in v3.3.0

type JobLoadDestinationEncryptionConfiguration struct {
	// Describes the Cloud KMS encryption key that will be used to protect destination BigQuery table.
	// The BigQuery Service Account associated with your project requires access to this encryption key.
	KmsKeyName string `pulumi:"kmsKeyName"`
}

type JobLoadDestinationEncryptionConfigurationArgs added in v3.3.0

type JobLoadDestinationEncryptionConfigurationArgs struct {
	// Describes the Cloud KMS encryption key that will be used to protect destination BigQuery table.
	// The BigQuery Service Account associated with your project requires access to this encryption key.
	KmsKeyName pulumi.StringInput `pulumi:"kmsKeyName"`
}

func (JobLoadDestinationEncryptionConfigurationArgs) ElementType added in v3.3.0

func (JobLoadDestinationEncryptionConfigurationArgs) ToJobLoadDestinationEncryptionConfigurationOutput added in v3.3.0

func (i JobLoadDestinationEncryptionConfigurationArgs) ToJobLoadDestinationEncryptionConfigurationOutput() JobLoadDestinationEncryptionConfigurationOutput

func (JobLoadDestinationEncryptionConfigurationArgs) ToJobLoadDestinationEncryptionConfigurationOutputWithContext added in v3.3.0

func (i JobLoadDestinationEncryptionConfigurationArgs) ToJobLoadDestinationEncryptionConfigurationOutputWithContext(ctx context.Context) JobLoadDestinationEncryptionConfigurationOutput

func (JobLoadDestinationEncryptionConfigurationArgs) ToJobLoadDestinationEncryptionConfigurationPtrOutput added in v3.3.0

func (i JobLoadDestinationEncryptionConfigurationArgs) ToJobLoadDestinationEncryptionConfigurationPtrOutput() JobLoadDestinationEncryptionConfigurationPtrOutput

func (JobLoadDestinationEncryptionConfigurationArgs) ToJobLoadDestinationEncryptionConfigurationPtrOutputWithContext added in v3.3.0

func (i JobLoadDestinationEncryptionConfigurationArgs) ToJobLoadDestinationEncryptionConfigurationPtrOutputWithContext(ctx context.Context) JobLoadDestinationEncryptionConfigurationPtrOutput

type JobLoadDestinationEncryptionConfigurationInput added in v3.3.0

type JobLoadDestinationEncryptionConfigurationInput interface {
	pulumi.Input

	ToJobLoadDestinationEncryptionConfigurationOutput() JobLoadDestinationEncryptionConfigurationOutput
	ToJobLoadDestinationEncryptionConfigurationOutputWithContext(context.Context) JobLoadDestinationEncryptionConfigurationOutput
}

JobLoadDestinationEncryptionConfigurationInput is an input type that accepts JobLoadDestinationEncryptionConfigurationArgs and JobLoadDestinationEncryptionConfigurationOutput values. You can construct a concrete instance of `JobLoadDestinationEncryptionConfigurationInput` via:

JobLoadDestinationEncryptionConfigurationArgs{...}

type JobLoadDestinationEncryptionConfigurationOutput added in v3.3.0

type JobLoadDestinationEncryptionConfigurationOutput struct{ *pulumi.OutputState }

func (JobLoadDestinationEncryptionConfigurationOutput) ElementType added in v3.3.0

func (JobLoadDestinationEncryptionConfigurationOutput) KmsKeyName added in v3.3.0

Describes the Cloud KMS encryption key that will be used to protect destination BigQuery table. The BigQuery Service Account associated with your project requires access to this encryption key.

func (JobLoadDestinationEncryptionConfigurationOutput) ToJobLoadDestinationEncryptionConfigurationOutput added in v3.3.0

func (o JobLoadDestinationEncryptionConfigurationOutput) ToJobLoadDestinationEncryptionConfigurationOutput() JobLoadDestinationEncryptionConfigurationOutput

func (JobLoadDestinationEncryptionConfigurationOutput) ToJobLoadDestinationEncryptionConfigurationOutputWithContext added in v3.3.0

func (o JobLoadDestinationEncryptionConfigurationOutput) ToJobLoadDestinationEncryptionConfigurationOutputWithContext(ctx context.Context) JobLoadDestinationEncryptionConfigurationOutput

func (JobLoadDestinationEncryptionConfigurationOutput) ToJobLoadDestinationEncryptionConfigurationPtrOutput added in v3.3.0

func (o JobLoadDestinationEncryptionConfigurationOutput) ToJobLoadDestinationEncryptionConfigurationPtrOutput() JobLoadDestinationEncryptionConfigurationPtrOutput

func (JobLoadDestinationEncryptionConfigurationOutput) ToJobLoadDestinationEncryptionConfigurationPtrOutputWithContext added in v3.3.0

func (o JobLoadDestinationEncryptionConfigurationOutput) ToJobLoadDestinationEncryptionConfigurationPtrOutputWithContext(ctx context.Context) JobLoadDestinationEncryptionConfigurationPtrOutput

type JobLoadDestinationEncryptionConfigurationPtrInput added in v3.3.0

type JobLoadDestinationEncryptionConfigurationPtrInput interface {
	pulumi.Input

	ToJobLoadDestinationEncryptionConfigurationPtrOutput() JobLoadDestinationEncryptionConfigurationPtrOutput
	ToJobLoadDestinationEncryptionConfigurationPtrOutputWithContext(context.Context) JobLoadDestinationEncryptionConfigurationPtrOutput
}

JobLoadDestinationEncryptionConfigurationPtrInput is an input type that accepts JobLoadDestinationEncryptionConfigurationArgs, JobLoadDestinationEncryptionConfigurationPtr and JobLoadDestinationEncryptionConfigurationPtrOutput values. You can construct a concrete instance of `JobLoadDestinationEncryptionConfigurationPtrInput` via:

		 JobLoadDestinationEncryptionConfigurationArgs{...}

 or:

		 nil

type JobLoadDestinationEncryptionConfigurationPtrOutput added in v3.3.0

type JobLoadDestinationEncryptionConfigurationPtrOutput struct{ *pulumi.OutputState }

func (JobLoadDestinationEncryptionConfigurationPtrOutput) Elem added in v3.3.0

func (JobLoadDestinationEncryptionConfigurationPtrOutput) ElementType added in v3.3.0

func (JobLoadDestinationEncryptionConfigurationPtrOutput) KmsKeyName added in v3.3.0

Describes the Cloud KMS encryption key that will be used to protect destination BigQuery table. The BigQuery Service Account associated with your project requires access to this encryption key.

func (JobLoadDestinationEncryptionConfigurationPtrOutput) ToJobLoadDestinationEncryptionConfigurationPtrOutput added in v3.3.0

func (o JobLoadDestinationEncryptionConfigurationPtrOutput) ToJobLoadDestinationEncryptionConfigurationPtrOutput() JobLoadDestinationEncryptionConfigurationPtrOutput

func (JobLoadDestinationEncryptionConfigurationPtrOutput) ToJobLoadDestinationEncryptionConfigurationPtrOutputWithContext added in v3.3.0

func (o JobLoadDestinationEncryptionConfigurationPtrOutput) ToJobLoadDestinationEncryptionConfigurationPtrOutputWithContext(ctx context.Context) JobLoadDestinationEncryptionConfigurationPtrOutput

type JobLoadDestinationTable added in v3.3.0

type JobLoadDestinationTable struct {
	// The ID of the dataset containing this model.
	DatasetId string `pulumi:"datasetId"`
	// The ID of the project containing this model.
	ProjectId string `pulumi:"projectId"`
	// The ID of the table.
	TableId string `pulumi:"tableId"`
}

type JobLoadDestinationTableArgs added in v3.3.0

type JobLoadDestinationTableArgs struct {
	// The ID of the dataset containing this model.
	DatasetId pulumi.StringInput `pulumi:"datasetId"`
	// The ID of the project containing this model.
	ProjectId pulumi.StringInput `pulumi:"projectId"`
	// The ID of the table.
	TableId pulumi.StringInput `pulumi:"tableId"`
}

func (JobLoadDestinationTableArgs) ElementType added in v3.3.0

func (JobLoadDestinationTableArgs) ToJobLoadDestinationTableOutput added in v3.3.0

func (i JobLoadDestinationTableArgs) ToJobLoadDestinationTableOutput() JobLoadDestinationTableOutput

func (JobLoadDestinationTableArgs) ToJobLoadDestinationTableOutputWithContext added in v3.3.0

func (i JobLoadDestinationTableArgs) ToJobLoadDestinationTableOutputWithContext(ctx context.Context) JobLoadDestinationTableOutput

func (JobLoadDestinationTableArgs) ToJobLoadDestinationTablePtrOutput added in v3.3.0

func (i JobLoadDestinationTableArgs) ToJobLoadDestinationTablePtrOutput() JobLoadDestinationTablePtrOutput

func (JobLoadDestinationTableArgs) ToJobLoadDestinationTablePtrOutputWithContext added in v3.3.0

func (i JobLoadDestinationTableArgs) ToJobLoadDestinationTablePtrOutputWithContext(ctx context.Context) JobLoadDestinationTablePtrOutput

type JobLoadDestinationTableInput added in v3.3.0

type JobLoadDestinationTableInput interface {
	pulumi.Input

	ToJobLoadDestinationTableOutput() JobLoadDestinationTableOutput
	ToJobLoadDestinationTableOutputWithContext(context.Context) JobLoadDestinationTableOutput
}

JobLoadDestinationTableInput is an input type that accepts JobLoadDestinationTableArgs and JobLoadDestinationTableOutput values. You can construct a concrete instance of `JobLoadDestinationTableInput` via:

JobLoadDestinationTableArgs{...}

type JobLoadDestinationTableOutput added in v3.3.0

type JobLoadDestinationTableOutput struct{ *pulumi.OutputState }

func (JobLoadDestinationTableOutput) DatasetId added in v3.3.0

The ID of the dataset containing this model.

func (JobLoadDestinationTableOutput) ElementType added in v3.3.0

func (JobLoadDestinationTableOutput) ProjectId added in v3.3.0

The ID of the project containing this model.

func (JobLoadDestinationTableOutput) TableId added in v3.3.0

The ID of the table.

func (JobLoadDestinationTableOutput) ToJobLoadDestinationTableOutput added in v3.3.0

func (o JobLoadDestinationTableOutput) ToJobLoadDestinationTableOutput() JobLoadDestinationTableOutput

func (JobLoadDestinationTableOutput) ToJobLoadDestinationTableOutputWithContext added in v3.3.0

func (o JobLoadDestinationTableOutput) ToJobLoadDestinationTableOutputWithContext(ctx context.Context) JobLoadDestinationTableOutput

func (JobLoadDestinationTableOutput) ToJobLoadDestinationTablePtrOutput added in v3.3.0

func (o JobLoadDestinationTableOutput) ToJobLoadDestinationTablePtrOutput() JobLoadDestinationTablePtrOutput

func (JobLoadDestinationTableOutput) ToJobLoadDestinationTablePtrOutputWithContext added in v3.3.0

func (o JobLoadDestinationTableOutput) ToJobLoadDestinationTablePtrOutputWithContext(ctx context.Context) JobLoadDestinationTablePtrOutput

type JobLoadDestinationTablePtrInput added in v3.3.0

type JobLoadDestinationTablePtrInput interface {
	pulumi.Input

	ToJobLoadDestinationTablePtrOutput() JobLoadDestinationTablePtrOutput
	ToJobLoadDestinationTablePtrOutputWithContext(context.Context) JobLoadDestinationTablePtrOutput
}

JobLoadDestinationTablePtrInput is an input type that accepts JobLoadDestinationTableArgs, JobLoadDestinationTablePtr and JobLoadDestinationTablePtrOutput values. You can construct a concrete instance of `JobLoadDestinationTablePtrInput` via:

		 JobLoadDestinationTableArgs{...}

 or:

		 nil

func JobLoadDestinationTablePtr added in v3.3.0

func JobLoadDestinationTablePtr(v *JobLoadDestinationTableArgs) JobLoadDestinationTablePtrInput

type JobLoadDestinationTablePtrOutput added in v3.3.0

type JobLoadDestinationTablePtrOutput struct{ *pulumi.OutputState }

func (JobLoadDestinationTablePtrOutput) DatasetId added in v3.3.0

The ID of the dataset containing this model.

func (JobLoadDestinationTablePtrOutput) Elem added in v3.3.0

func (JobLoadDestinationTablePtrOutput) ElementType added in v3.3.0

func (JobLoadDestinationTablePtrOutput) ProjectId added in v3.3.0

The ID of the project containing this model.

func (JobLoadDestinationTablePtrOutput) TableId added in v3.3.0

The ID of the table.

func (JobLoadDestinationTablePtrOutput) ToJobLoadDestinationTablePtrOutput added in v3.3.0

func (o JobLoadDestinationTablePtrOutput) ToJobLoadDestinationTablePtrOutput() JobLoadDestinationTablePtrOutput

func (JobLoadDestinationTablePtrOutput) ToJobLoadDestinationTablePtrOutputWithContext added in v3.3.0

func (o JobLoadDestinationTablePtrOutput) ToJobLoadDestinationTablePtrOutputWithContext(ctx context.Context) JobLoadDestinationTablePtrOutput

type JobLoadInput added in v3.3.0

type JobLoadInput interface {
	pulumi.Input

	ToJobLoadOutput() JobLoadOutput
	ToJobLoadOutputWithContext(context.Context) JobLoadOutput
}

JobLoadInput is an input type that accepts JobLoadArgs and JobLoadOutput values. You can construct a concrete instance of `JobLoadInput` via:

JobLoadArgs{...}

type JobLoadOutput added in v3.3.0

type JobLoadOutput struct{ *pulumi.OutputState }

func (JobLoadOutput) AllowJaggedRows added in v3.3.0

func (o JobLoadOutput) AllowJaggedRows() pulumi.BoolPtrOutput

Accept rows that are missing trailing optional columns. The missing values are treated as nulls. If false, records with missing trailing columns are treated as bad records, and if there are too many bad records, an invalid error is returned in the job result. The default value is false. Only applicable to CSV, ignored for other formats.

func (JobLoadOutput) AllowQuotedNewlines added in v3.3.0

func (o JobLoadOutput) AllowQuotedNewlines() pulumi.BoolPtrOutput

Indicates if BigQuery should allow quoted data sections that contain newline characters in a CSV file. The default value is false.

func (JobLoadOutput) Autodetect added in v3.3.0

func (o JobLoadOutput) Autodetect() pulumi.BoolPtrOutput

Indicates if we should automatically infer the options and schema for CSV and JSON sources.

func (JobLoadOutput) CreateDisposition added in v3.3.0

func (o JobLoadOutput) CreateDisposition() pulumi.StringPtrOutput

Specifies whether the job is allowed to create new tables. The following values are supported: CREATE_IF_NEEDED: If the table does not exist, BigQuery creates the table. CREATE_NEVER: The table must already exist. If it does not, a 'notFound' error is returned in the job result. Creation, truncation and append actions occur as one atomic update upon job completion

func (JobLoadOutput) DestinationEncryptionConfiguration added in v3.3.0

func (o JobLoadOutput) DestinationEncryptionConfiguration() JobLoadDestinationEncryptionConfigurationPtrOutput

Custom encryption configuration (e.g., Cloud KMS keys) Structure is documented below.

func (JobLoadOutput) DestinationTable added in v3.3.0

func (o JobLoadOutput) DestinationTable() JobLoadDestinationTableOutput

The destination table. Structure is documented below.

func (JobLoadOutput) ElementType added in v3.3.0

func (JobLoadOutput) ElementType() reflect.Type

func (JobLoadOutput) Encoding added in v3.3.0

func (o JobLoadOutput) Encoding() pulumi.StringPtrOutput

The character encoding of the data. The supported values are UTF-8 or ISO-8859-1. The default value is UTF-8. BigQuery decodes the data after the raw, binary data has been split using the values of the quote and fieldDelimiter properties.

func (JobLoadOutput) FieldDelimiter added in v3.3.0

func (o JobLoadOutput) FieldDelimiter() pulumi.StringPtrOutput

When extracting data in CSV format, this defines the delimiter to use between fields in the exported data. Default is ','

func (JobLoadOutput) IgnoreUnknownValues added in v3.3.0

func (o JobLoadOutput) IgnoreUnknownValues() pulumi.BoolPtrOutput

Indicates if BigQuery should allow extra values that are not represented in the table schema. If true, the extra values are ignored. If false, records with extra columns are treated as bad records, and if there are too many bad records, an invalid error is returned in the job result. The default value is false. The sourceFormat property determines what BigQuery treats as an extra value: CSV: Trailing columns JSON: Named values that don't match any column names

func (JobLoadOutput) MaxBadRecords added in v3.3.0

func (o JobLoadOutput) MaxBadRecords() pulumi.IntPtrOutput

The maximum number of bad records that BigQuery can ignore when running the job. If the number of bad records exceeds this value, an invalid error is returned in the job result. The default value is 0, which requires that all records are valid.

func (JobLoadOutput) NullMarker added in v3.3.0

func (o JobLoadOutput) NullMarker() pulumi.StringPtrOutput

Specifies a string that represents a null value in a CSV file. The default value is the empty string. If you set this property to a custom value, BigQuery throws an error if an empty string is present for all data types except for STRING and BYTE. For STRING and BYTE columns, BigQuery interprets the empty string as an empty value.

func (JobLoadOutput) ProjectionFields added in v3.3.0

func (o JobLoadOutput) ProjectionFields() pulumi.StringArrayOutput

If sourceFormat is set to "DATASTORE_BACKUP", indicates which entity properties to load into BigQuery from a Cloud Datastore backup. Property names are case sensitive and must be top-level properties. If no properties are specified, BigQuery loads all properties. If any named property isn't found in the Cloud Datastore backup, an invalid error is returned in the job result.

func (JobLoadOutput) Quote added in v3.3.0

The value that is used to quote data sections in a CSV file. BigQuery converts the string to ISO-8859-1 encoding, and then uses the first byte of the encoded string to split the data in its raw, binary state. The default value is a double-quote ('"'). If your data does not contain quoted sections, set the property value to an empty string. If your data contains quoted newline characters, you must also set the allowQuotedNewlines property to true.

func (JobLoadOutput) SchemaUpdateOptions added in v3.3.0

func (o JobLoadOutput) SchemaUpdateOptions() pulumi.StringArrayOutput

Allows the schema of the destination table to be updated as a side effect of the load job if a schema is autodetected or supplied in the job configuration. Schema update options are supported in two cases: when writeDisposition is WRITE_APPEND; when writeDisposition is WRITE_TRUNCATE and the destination table is a partition of a table, specified by partition decorators. For normal tables, WRITE_TRUNCATE will always overwrite the schema. One or more of the following values are specified: ALLOW_FIELD_ADDITION: allow adding a nullable field to the schema. ALLOW_FIELD_RELAXATION: allow relaxing a required field in the original schema to nullable.

func (JobLoadOutput) SkipLeadingRows added in v3.3.0

func (o JobLoadOutput) SkipLeadingRows() pulumi.IntPtrOutput

The number of rows at the top of a CSV file that BigQuery will skip when loading the data. The default value is 0. This property is useful if you have header rows in the file that should be skipped. When autodetect is on, the behavior is the following: skipLeadingRows unspecified - Autodetect tries to detect headers in the first row. If they are not detected, the row is read as data. Otherwise data is read starting from the second row. skipLeadingRows is 0 - Instructs autodetect that there are no headers and data should be read starting from the first row. skipLeadingRows = N > 0 - Autodetect skips N-1 rows and tries to detect headers in row N. If headers are not detected, row N is just skipped. Otherwise row N is used to extract column names for the detected schema.

func (JobLoadOutput) SourceFormat added in v3.3.0

func (o JobLoadOutput) SourceFormat() pulumi.StringPtrOutput

The format of the data files. For CSV files, specify "CSV". For datastore backups, specify "DATASTORE_BACKUP". For newline-delimited JSON, specify "NEWLINE_DELIMITED_JSON". For Avro, specify "AVRO". For parquet, specify "PARQUET". For orc, specify "ORC". The default value is CSV.

func (JobLoadOutput) SourceUris added in v3.3.0

func (o JobLoadOutput) SourceUris() pulumi.StringArrayOutput

The fully-qualified URIs that point to your data in Google Cloud. For Google Cloud Storage URIs: Each URI can contain one '*' wildcard character and it must come after the 'bucket' name. Size limits related to load jobs apply to external data sources. For Google Cloud Bigtable URIs: Exactly one URI can be specified and it has be a fully specified and valid HTTPS URL for a Google Cloud Bigtable table. For Google Cloud Datastore backups: Exactly one URI can be specified. Also, the '*' wildcard character is not allowed.

func (JobLoadOutput) TimePartitioning added in v3.3.0

func (o JobLoadOutput) TimePartitioning() JobLoadTimePartitioningPtrOutput

Time-based partitioning specification for the destination table. Structure is documented below.

func (JobLoadOutput) ToJobLoadOutput added in v3.3.0

func (o JobLoadOutput) ToJobLoadOutput() JobLoadOutput

func (JobLoadOutput) ToJobLoadOutputWithContext added in v3.3.0

func (o JobLoadOutput) ToJobLoadOutputWithContext(ctx context.Context) JobLoadOutput

func (JobLoadOutput) ToJobLoadPtrOutput added in v3.3.0

func (o JobLoadOutput) ToJobLoadPtrOutput() JobLoadPtrOutput

func (JobLoadOutput) ToJobLoadPtrOutputWithContext added in v3.3.0

func (o JobLoadOutput) ToJobLoadPtrOutputWithContext(ctx context.Context) JobLoadPtrOutput

func (JobLoadOutput) WriteDisposition added in v3.3.0

func (o JobLoadOutput) WriteDisposition() pulumi.StringPtrOutput

Specifies the action that occurs if the destination table already exists. The following values are supported: WRITE_TRUNCATE: If the table already exists, BigQuery overwrites the table data and uses the schema from the query result. WRITE_APPEND: If the table already exists, BigQuery appends the data to the table. WRITE_EMPTY: If the table already exists and contains data, a 'duplicate' error is returned in the job result. Each action is atomic and only occurs if BigQuery is able to complete the job successfully. Creation, truncation and append actions occur as one atomic update upon job completion.

type JobLoadPtrInput added in v3.3.0

type JobLoadPtrInput interface {
	pulumi.Input

	ToJobLoadPtrOutput() JobLoadPtrOutput
	ToJobLoadPtrOutputWithContext(context.Context) JobLoadPtrOutput
}

JobLoadPtrInput is an input type that accepts JobLoadArgs, JobLoadPtr and JobLoadPtrOutput values. You can construct a concrete instance of `JobLoadPtrInput` via:

		 JobLoadArgs{...}

 or:

		 nil

func JobLoadPtr added in v3.3.0

func JobLoadPtr(v *JobLoadArgs) JobLoadPtrInput

type JobLoadPtrOutput added in v3.3.0

type JobLoadPtrOutput struct{ *pulumi.OutputState }

func (JobLoadPtrOutput) AllowJaggedRows added in v3.3.0

func (o JobLoadPtrOutput) AllowJaggedRows() pulumi.BoolPtrOutput

Accept rows that are missing trailing optional columns. The missing values are treated as nulls. If false, records with missing trailing columns are treated as bad records, and if there are too many bad records, an invalid error is returned in the job result. The default value is false. Only applicable to CSV, ignored for other formats.

func (JobLoadPtrOutput) AllowQuotedNewlines added in v3.3.0

func (o JobLoadPtrOutput) AllowQuotedNewlines() pulumi.BoolPtrOutput

Indicates if BigQuery should allow quoted data sections that contain newline characters in a CSV file. The default value is false.

func (JobLoadPtrOutput) Autodetect added in v3.3.0

func (o JobLoadPtrOutput) Autodetect() pulumi.BoolPtrOutput

Indicates if we should automatically infer the options and schema for CSV and JSON sources.

func (JobLoadPtrOutput) CreateDisposition added in v3.3.0

func (o JobLoadPtrOutput) CreateDisposition() pulumi.StringPtrOutput

Specifies whether the job is allowed to create new tables. The following values are supported: CREATE_IF_NEEDED: If the table does not exist, BigQuery creates the table. CREATE_NEVER: The table must already exist. If it does not, a 'notFound' error is returned in the job result. Creation, truncation and append actions occur as one atomic update upon job completion

func (JobLoadPtrOutput) DestinationEncryptionConfiguration added in v3.3.0

func (o JobLoadPtrOutput) DestinationEncryptionConfiguration() JobLoadDestinationEncryptionConfigurationPtrOutput

Custom encryption configuration (e.g., Cloud KMS keys) Structure is documented below.

func (JobLoadPtrOutput) DestinationTable added in v3.3.0

The destination table. Structure is documented below.

func (JobLoadPtrOutput) Elem added in v3.3.0

func (JobLoadPtrOutput) ElementType added in v3.3.0

func (JobLoadPtrOutput) ElementType() reflect.Type

func (JobLoadPtrOutput) Encoding added in v3.3.0

The character encoding of the data. The supported values are UTF-8 or ISO-8859-1. The default value is UTF-8. BigQuery decodes the data after the raw, binary data has been split using the values of the quote and fieldDelimiter properties.

func (JobLoadPtrOutput) FieldDelimiter added in v3.3.0

func (o JobLoadPtrOutput) FieldDelimiter() pulumi.StringPtrOutput

When extracting data in CSV format, this defines the delimiter to use between fields in the exported data. Default is ','

func (JobLoadPtrOutput) IgnoreUnknownValues added in v3.3.0

func (o JobLoadPtrOutput) IgnoreUnknownValues() pulumi.BoolPtrOutput

Indicates if BigQuery should allow extra values that are not represented in the table schema. If true, the extra values are ignored. If false, records with extra columns are treated as bad records, and if there are too many bad records, an invalid error is returned in the job result. The default value is false. The sourceFormat property determines what BigQuery treats as an extra value: CSV: Trailing columns JSON: Named values that don't match any column names

func (JobLoadPtrOutput) MaxBadRecords added in v3.3.0

func (o JobLoadPtrOutput) MaxBadRecords() pulumi.IntPtrOutput

The maximum number of bad records that BigQuery can ignore when running the job. If the number of bad records exceeds this value, an invalid error is returned in the job result. The default value is 0, which requires that all records are valid.

func (JobLoadPtrOutput) NullMarker added in v3.3.0

func (o JobLoadPtrOutput) NullMarker() pulumi.StringPtrOutput

Specifies a string that represents a null value in a CSV file. The default value is the empty string. If you set this property to a custom value, BigQuery throws an error if an empty string is present for all data types except for STRING and BYTE. For STRING and BYTE columns, BigQuery interprets the empty string as an empty value.

func (JobLoadPtrOutput) ProjectionFields added in v3.3.0

func (o JobLoadPtrOutput) ProjectionFields() pulumi.StringArrayOutput

If sourceFormat is set to "DATASTORE_BACKUP", indicates which entity properties to load into BigQuery from a Cloud Datastore backup. Property names are case sensitive and must be top-level properties. If no properties are specified, BigQuery loads all properties. If any named property isn't found in the Cloud Datastore backup, an invalid error is returned in the job result.

func (JobLoadPtrOutput) Quote added in v3.3.0

The value that is used to quote data sections in a CSV file. BigQuery converts the string to ISO-8859-1 encoding, and then uses the first byte of the encoded string to split the data in its raw, binary state. The default value is a double-quote ('"'). If your data does not contain quoted sections, set the property value to an empty string. If your data contains quoted newline characters, you must also set the allowQuotedNewlines property to true.

func (JobLoadPtrOutput) SchemaUpdateOptions added in v3.3.0

func (o JobLoadPtrOutput) SchemaUpdateOptions() pulumi.StringArrayOutput

Allows the schema of the destination table to be updated as a side effect of the load job if a schema is autodetected or supplied in the job configuration. Schema update options are supported in two cases: when writeDisposition is WRITE_APPEND; when writeDisposition is WRITE_TRUNCATE and the destination table is a partition of a table, specified by partition decorators. For normal tables, WRITE_TRUNCATE will always overwrite the schema. One or more of the following values are specified: ALLOW_FIELD_ADDITION: allow adding a nullable field to the schema. ALLOW_FIELD_RELAXATION: allow relaxing a required field in the original schema to nullable.

func (JobLoadPtrOutput) SkipLeadingRows added in v3.3.0

func (o JobLoadPtrOutput) SkipLeadingRows() pulumi.IntPtrOutput

The number of rows at the top of a CSV file that BigQuery will skip when loading the data. The default value is 0. This property is useful if you have header rows in the file that should be skipped. When autodetect is on, the behavior is the following: skipLeadingRows unspecified - Autodetect tries to detect headers in the first row. If they are not detected, the row is read as data. Otherwise data is read starting from the second row. skipLeadingRows is 0 - Instructs autodetect that there are no headers and data should be read starting from the first row. skipLeadingRows = N > 0 - Autodetect skips N-1 rows and tries to detect headers in row N. If headers are not detected, row N is just skipped. Otherwise row N is used to extract column names for the detected schema.

func (JobLoadPtrOutput) SourceFormat added in v3.3.0

func (o JobLoadPtrOutput) SourceFormat() pulumi.StringPtrOutput

The format of the data files. For CSV files, specify "CSV". For datastore backups, specify "DATASTORE_BACKUP". For newline-delimited JSON, specify "NEWLINE_DELIMITED_JSON". For Avro, specify "AVRO". For parquet, specify "PARQUET". For orc, specify "ORC". The default value is CSV.

func (JobLoadPtrOutput) SourceUris added in v3.3.0

func (o JobLoadPtrOutput) SourceUris() pulumi.StringArrayOutput

The fully-qualified URIs that point to your data in Google Cloud. For Google Cloud Storage URIs: Each URI can contain one '*' wildcard character and it must come after the 'bucket' name. Size limits related to load jobs apply to external data sources. For Google Cloud Bigtable URIs: Exactly one URI can be specified and it has be a fully specified and valid HTTPS URL for a Google Cloud Bigtable table. For Google Cloud Datastore backups: Exactly one URI can be specified. Also, the '*' wildcard character is not allowed.

func (JobLoadPtrOutput) TimePartitioning added in v3.3.0

Time-based partitioning specification for the destination table. Structure is documented below.

func (JobLoadPtrOutput) ToJobLoadPtrOutput added in v3.3.0

func (o JobLoadPtrOutput) ToJobLoadPtrOutput() JobLoadPtrOutput

func (JobLoadPtrOutput) ToJobLoadPtrOutputWithContext added in v3.3.0

func (o JobLoadPtrOutput) ToJobLoadPtrOutputWithContext(ctx context.Context) JobLoadPtrOutput

func (JobLoadPtrOutput) WriteDisposition added in v3.3.0

func (o JobLoadPtrOutput) WriteDisposition() pulumi.StringPtrOutput

Specifies the action that occurs if the destination table already exists. The following values are supported: WRITE_TRUNCATE: If the table already exists, BigQuery overwrites the table data and uses the schema from the query result. WRITE_APPEND: If the table already exists, BigQuery appends the data to the table. WRITE_EMPTY: If the table already exists and contains data, a 'duplicate' error is returned in the job result. Each action is atomic and only occurs if BigQuery is able to complete the job successfully. Creation, truncation and append actions occur as one atomic update upon job completion.

type JobLoadTimePartitioning added in v3.3.0

type JobLoadTimePartitioning struct {
	// Number of milliseconds for which to keep the storage for a partition. A wrapper is used here because 0 is an invalid value.
	ExpirationMs *string `pulumi:"expirationMs"`
	// If not set, the table is partitioned by pseudo column '_PARTITIONTIME'; if set, the table is partitioned by this field.
	// The field must be a top-level TIMESTAMP or DATE field. Its mode must be NULLABLE or REQUIRED.
	// A wrapper is used here because an empty string is an invalid value.
	Field *string `pulumi:"field"`
	// The only type supported is DAY, which will generate one partition per day. Providing an empty string used to cause an error,
	// but in OnePlatform the field will be treated as unset.
	Type string `pulumi:"type"`
}

type JobLoadTimePartitioningArgs added in v3.3.0

type JobLoadTimePartitioningArgs struct {
	// Number of milliseconds for which to keep the storage for a partition. A wrapper is used here because 0 is an invalid value.
	ExpirationMs pulumi.StringPtrInput `pulumi:"expirationMs"`
	// If not set, the table is partitioned by pseudo column '_PARTITIONTIME'; if set, the table is partitioned by this field.
	// The field must be a top-level TIMESTAMP or DATE field. Its mode must be NULLABLE or REQUIRED.
	// A wrapper is used here because an empty string is an invalid value.
	Field pulumi.StringPtrInput `pulumi:"field"`
	// The only type supported is DAY, which will generate one partition per day. Providing an empty string used to cause an error,
	// but in OnePlatform the field will be treated as unset.
	Type pulumi.StringInput `pulumi:"type"`
}

func (JobLoadTimePartitioningArgs) ElementType added in v3.3.0

func (JobLoadTimePartitioningArgs) ToJobLoadTimePartitioningOutput added in v3.3.0

func (i JobLoadTimePartitioningArgs) ToJobLoadTimePartitioningOutput() JobLoadTimePartitioningOutput

func (JobLoadTimePartitioningArgs) ToJobLoadTimePartitioningOutputWithContext added in v3.3.0

func (i JobLoadTimePartitioningArgs) ToJobLoadTimePartitioningOutputWithContext(ctx context.Context) JobLoadTimePartitioningOutput

func (JobLoadTimePartitioningArgs) ToJobLoadTimePartitioningPtrOutput added in v3.3.0

func (i JobLoadTimePartitioningArgs) ToJobLoadTimePartitioningPtrOutput() JobLoadTimePartitioningPtrOutput

func (JobLoadTimePartitioningArgs) ToJobLoadTimePartitioningPtrOutputWithContext added in v3.3.0

func (i JobLoadTimePartitioningArgs) ToJobLoadTimePartitioningPtrOutputWithContext(ctx context.Context) JobLoadTimePartitioningPtrOutput

type JobLoadTimePartitioningInput added in v3.3.0

type JobLoadTimePartitioningInput interface {
	pulumi.Input

	ToJobLoadTimePartitioningOutput() JobLoadTimePartitioningOutput
	ToJobLoadTimePartitioningOutputWithContext(context.Context) JobLoadTimePartitioningOutput
}

JobLoadTimePartitioningInput is an input type that accepts JobLoadTimePartitioningArgs and JobLoadTimePartitioningOutput values. You can construct a concrete instance of `JobLoadTimePartitioningInput` via:

JobLoadTimePartitioningArgs{...}

type JobLoadTimePartitioningOutput added in v3.3.0

type JobLoadTimePartitioningOutput struct{ *pulumi.OutputState }

func (JobLoadTimePartitioningOutput) ElementType added in v3.3.0

func (JobLoadTimePartitioningOutput) ExpirationMs added in v3.3.0

Number of milliseconds for which to keep the storage for a partition. A wrapper is used here because 0 is an invalid value.

func (JobLoadTimePartitioningOutput) Field added in v3.3.0

If not set, the table is partitioned by pseudo column '_PARTITIONTIME'; if set, the table is partitioned by this field. The field must be a top-level TIMESTAMP or DATE field. Its mode must be NULLABLE or REQUIRED. A wrapper is used here because an empty string is an invalid value.

func (JobLoadTimePartitioningOutput) ToJobLoadTimePartitioningOutput added in v3.3.0

func (o JobLoadTimePartitioningOutput) ToJobLoadTimePartitioningOutput() JobLoadTimePartitioningOutput

func (JobLoadTimePartitioningOutput) ToJobLoadTimePartitioningOutputWithContext added in v3.3.0

func (o JobLoadTimePartitioningOutput) ToJobLoadTimePartitioningOutputWithContext(ctx context.Context) JobLoadTimePartitioningOutput

func (JobLoadTimePartitioningOutput) ToJobLoadTimePartitioningPtrOutput added in v3.3.0

func (o JobLoadTimePartitioningOutput) ToJobLoadTimePartitioningPtrOutput() JobLoadTimePartitioningPtrOutput

func (JobLoadTimePartitioningOutput) ToJobLoadTimePartitioningPtrOutputWithContext added in v3.3.0

func (o JobLoadTimePartitioningOutput) ToJobLoadTimePartitioningPtrOutputWithContext(ctx context.Context) JobLoadTimePartitioningPtrOutput

func (JobLoadTimePartitioningOutput) Type added in v3.3.0

The only type supported is DAY, which will generate one partition per day. Providing an empty string used to cause an error, but in OnePlatform the field will be treated as unset.

type JobLoadTimePartitioningPtrInput added in v3.3.0

type JobLoadTimePartitioningPtrInput interface {
	pulumi.Input

	ToJobLoadTimePartitioningPtrOutput() JobLoadTimePartitioningPtrOutput
	ToJobLoadTimePartitioningPtrOutputWithContext(context.Context) JobLoadTimePartitioningPtrOutput
}

JobLoadTimePartitioningPtrInput is an input type that accepts JobLoadTimePartitioningArgs, JobLoadTimePartitioningPtr and JobLoadTimePartitioningPtrOutput values. You can construct a concrete instance of `JobLoadTimePartitioningPtrInput` via:

		 JobLoadTimePartitioningArgs{...}

 or:

		 nil

func JobLoadTimePartitioningPtr added in v3.3.0

func JobLoadTimePartitioningPtr(v *JobLoadTimePartitioningArgs) JobLoadTimePartitioningPtrInput

type JobLoadTimePartitioningPtrOutput added in v3.3.0

type JobLoadTimePartitioningPtrOutput struct{ *pulumi.OutputState }

func (JobLoadTimePartitioningPtrOutput) Elem added in v3.3.0

func (JobLoadTimePartitioningPtrOutput) ElementType added in v3.3.0

func (JobLoadTimePartitioningPtrOutput) ExpirationMs added in v3.3.0

Number of milliseconds for which to keep the storage for a partition. A wrapper is used here because 0 is an invalid value.

func (JobLoadTimePartitioningPtrOutput) Field added in v3.3.0

If not set, the table is partitioned by pseudo column '_PARTITIONTIME'; if set, the table is partitioned by this field. The field must be a top-level TIMESTAMP or DATE field. Its mode must be NULLABLE or REQUIRED. A wrapper is used here because an empty string is an invalid value.

func (JobLoadTimePartitioningPtrOutput) ToJobLoadTimePartitioningPtrOutput added in v3.3.0

func (o JobLoadTimePartitioningPtrOutput) ToJobLoadTimePartitioningPtrOutput() JobLoadTimePartitioningPtrOutput

func (JobLoadTimePartitioningPtrOutput) ToJobLoadTimePartitioningPtrOutputWithContext added in v3.3.0

func (o JobLoadTimePartitioningPtrOutput) ToJobLoadTimePartitioningPtrOutputWithContext(ctx context.Context) JobLoadTimePartitioningPtrOutput

func (JobLoadTimePartitioningPtrOutput) Type added in v3.3.0

The only type supported is DAY, which will generate one partition per day. Providing an empty string used to cause an error, but in OnePlatform the field will be treated as unset.

type JobQuery added in v3.3.0

type JobQuery struct {
	// If true and query uses legacy SQL dialect, allows the query to produce arbitrarily large result tables at a slight cost in performance.
	// Requires destinationTable to be set. For standard SQL queries, this flag is ignored and large results are always allowed.
	// However, you must still set destinationTable when result size exceeds the allowed maximum response size.
	AllowLargeResults *bool `pulumi:"allowLargeResults"`
	// Specifies whether the job is allowed to create new tables. The following values are supported:
	// CREATE_IF_NEEDED: If the table does not exist, BigQuery creates the table.
	// CREATE_NEVER: The table must already exist. If it does not, a 'notFound' error is returned in the job result.
	// Creation, truncation and append actions occur as one atomic update upon job completion
	CreateDisposition *string `pulumi:"createDisposition"`
	// Specifies the default dataset to use for unqualified table names in the query. Note that this does not alter behavior of unqualified dataset names.  Structure is documented below.
	DefaultDataset *JobQueryDefaultDataset `pulumi:"defaultDataset"`
	// Custom encryption configuration (e.g., Cloud KMS keys)  Structure is documented below.
	DestinationEncryptionConfiguration *JobQueryDestinationEncryptionConfiguration `pulumi:"destinationEncryptionConfiguration"`
	// The destination table.  Structure is documented below.
	DestinationTable *JobQueryDestinationTable `pulumi:"destinationTable"`
	// If true and query uses legacy SQL dialect, flattens all nested and repeated fields in the query results.
	// allowLargeResults must be true if this is set to false. For standard SQL queries, this flag is ignored and results are never flattened.
	FlattenResults *bool `pulumi:"flattenResults"`
	// Limits the billing tier for this job. Queries that have resource usage beyond this tier will fail (without incurring a charge).
	// If unspecified, this will be set to your project default.
	MaximumBillingTier *int `pulumi:"maximumBillingTier"`
	// Limits the bytes billed for this job. Queries that will have bytes billed beyond this limit will fail (without incurring a charge).
	// If unspecified, this will be set to your project default.
	MaximumBytesBilled *string `pulumi:"maximumBytesBilled"`
	// Standard SQL only. Set to POSITIONAL to use positional (?) query parameters or to NAMED to use named (@myparam) query parameters in this query.
	ParameterMode *string `pulumi:"parameterMode"`
	// Specifies a priority for the query.
	Priority *string `pulumi:"priority"`
	// Configures a query job.  Structure is documented below.
	Query string `pulumi:"query"`
	// Allows the schema of the destination table to be updated as a side effect of the load job if a schema is autodetected or
	// supplied in the job configuration. Schema update options are supported in two cases: when writeDisposition is WRITE_APPEND;
	// when writeDisposition is WRITE_TRUNCATE and the destination table is a partition of a table, specified by partition decorators.
	// For normal tables, WRITE_TRUNCATE will always overwrite the schema. One or more of the following values are specified:
	// ALLOW_FIELD_ADDITION: allow adding a nullable field to the schema.
	// ALLOW_FIELD_RELAXATION: allow relaxing a required field in the original schema to nullable.
	SchemaUpdateOptions []string `pulumi:"schemaUpdateOptions"`
	// Options controlling the execution of scripts.  Structure is documented below.
	ScriptOptions *JobQueryScriptOptions `pulumi:"scriptOptions"`
	// Specifies whether to use BigQuery's legacy SQL dialect for this query. The default value is true.
	// If set to false, the query will use BigQuery's standard SQL.
	UseLegacySql *bool `pulumi:"useLegacySql"`
	// Whether to look for the result in the query cache. The query cache is a best-effort cache that will be flushed whenever
	// tables in the query are modified. Moreover, the query cache is only available when a query does not have a destination table specified.
	// The default value is true.
	UseQueryCache *bool `pulumi:"useQueryCache"`
	// Describes user-defined function resources used in the query.  Structure is documented below.
	UserDefinedFunctionResources []JobQueryUserDefinedFunctionResource `pulumi:"userDefinedFunctionResources"`
	// Specifies the action that occurs if the destination table already exists. The following values are supported:
	// WRITE_TRUNCATE: If the table already exists, BigQuery overwrites the table data and uses the schema from the query result.
	// WRITE_APPEND: If the table already exists, BigQuery appends the data to the table.
	// WRITE_EMPTY: If the table already exists and contains data, a 'duplicate' error is returned in the job result.
	// Each action is atomic and only occurs if BigQuery is able to complete the job successfully.
	// Creation, truncation and append actions occur as one atomic update upon job completion.
	WriteDisposition *string `pulumi:"writeDisposition"`
}

type JobQueryArgs added in v3.3.0

type JobQueryArgs struct {
	// If true and query uses legacy SQL dialect, allows the query to produce arbitrarily large result tables at a slight cost in performance.
	// Requires destinationTable to be set. For standard SQL queries, this flag is ignored and large results are always allowed.
	// However, you must still set destinationTable when result size exceeds the allowed maximum response size.
	AllowLargeResults pulumi.BoolPtrInput `pulumi:"allowLargeResults"`
	// Specifies whether the job is allowed to create new tables. The following values are supported:
	// CREATE_IF_NEEDED: If the table does not exist, BigQuery creates the table.
	// CREATE_NEVER: The table must already exist. If it does not, a 'notFound' error is returned in the job result.
	// Creation, truncation and append actions occur as one atomic update upon job completion
	CreateDisposition pulumi.StringPtrInput `pulumi:"createDisposition"`
	// Specifies the default dataset to use for unqualified table names in the query. Note that this does not alter behavior of unqualified dataset names.  Structure is documented below.
	DefaultDataset JobQueryDefaultDatasetPtrInput `pulumi:"defaultDataset"`
	// Custom encryption configuration (e.g., Cloud KMS keys)  Structure is documented below.
	DestinationEncryptionConfiguration JobQueryDestinationEncryptionConfigurationPtrInput `pulumi:"destinationEncryptionConfiguration"`
	// The destination table.  Structure is documented below.
	DestinationTable JobQueryDestinationTablePtrInput `pulumi:"destinationTable"`
	// If true and query uses legacy SQL dialect, flattens all nested and repeated fields in the query results.
	// allowLargeResults must be true if this is set to false. For standard SQL queries, this flag is ignored and results are never flattened.
	FlattenResults pulumi.BoolPtrInput `pulumi:"flattenResults"`
	// Limits the billing tier for this job. Queries that have resource usage beyond this tier will fail (without incurring a charge).
	// If unspecified, this will be set to your project default.
	MaximumBillingTier pulumi.IntPtrInput `pulumi:"maximumBillingTier"`
	// Limits the bytes billed for this job. Queries that will have bytes billed beyond this limit will fail (without incurring a charge).
	// If unspecified, this will be set to your project default.
	MaximumBytesBilled pulumi.StringPtrInput `pulumi:"maximumBytesBilled"`
	// Standard SQL only. Set to POSITIONAL to use positional (?) query parameters or to NAMED to use named (@myparam) query parameters in this query.
	ParameterMode pulumi.StringPtrInput `pulumi:"parameterMode"`
	// Specifies a priority for the query.
	Priority pulumi.StringPtrInput `pulumi:"priority"`
	// Configures a query job.  Structure is documented below.
	Query pulumi.StringInput `pulumi:"query"`
	// Allows the schema of the destination table to be updated as a side effect of the load job if a schema is autodetected or
	// supplied in the job configuration. Schema update options are supported in two cases: when writeDisposition is WRITE_APPEND;
	// when writeDisposition is WRITE_TRUNCATE and the destination table is a partition of a table, specified by partition decorators.
	// For normal tables, WRITE_TRUNCATE will always overwrite the schema. One or more of the following values are specified:
	// ALLOW_FIELD_ADDITION: allow adding a nullable field to the schema.
	// ALLOW_FIELD_RELAXATION: allow relaxing a required field in the original schema to nullable.
	SchemaUpdateOptions pulumi.StringArrayInput `pulumi:"schemaUpdateOptions"`
	// Options controlling the execution of scripts.  Structure is documented below.
	ScriptOptions JobQueryScriptOptionsPtrInput `pulumi:"scriptOptions"`
	// Specifies whether to use BigQuery's legacy SQL dialect for this query. The default value is true.
	// If set to false, the query will use BigQuery's standard SQL.
	UseLegacySql pulumi.BoolPtrInput `pulumi:"useLegacySql"`
	// Whether to look for the result in the query cache. The query cache is a best-effort cache that will be flushed whenever
	// tables in the query are modified. Moreover, the query cache is only available when a query does not have a destination table specified.
	// The default value is true.
	UseQueryCache pulumi.BoolPtrInput `pulumi:"useQueryCache"`
	// Describes user-defined function resources used in the query.  Structure is documented below.
	UserDefinedFunctionResources JobQueryUserDefinedFunctionResourceArrayInput `pulumi:"userDefinedFunctionResources"`
	// Specifies the action that occurs if the destination table already exists. The following values are supported:
	// WRITE_TRUNCATE: If the table already exists, BigQuery overwrites the table data and uses the schema from the query result.
	// WRITE_APPEND: If the table already exists, BigQuery appends the data to the table.
	// WRITE_EMPTY: If the table already exists and contains data, a 'duplicate' error is returned in the job result.
	// Each action is atomic and only occurs if BigQuery is able to complete the job successfully.
	// Creation, truncation and append actions occur as one atomic update upon job completion.
	WriteDisposition pulumi.StringPtrInput `pulumi:"writeDisposition"`
}

func (JobQueryArgs) ElementType added in v3.3.0

func (JobQueryArgs) ElementType() reflect.Type

func (JobQueryArgs) ToJobQueryOutput added in v3.3.0

func (i JobQueryArgs) ToJobQueryOutput() JobQueryOutput

func (JobQueryArgs) ToJobQueryOutputWithContext added in v3.3.0

func (i JobQueryArgs) ToJobQueryOutputWithContext(ctx context.Context) JobQueryOutput

func (JobQueryArgs) ToJobQueryPtrOutput added in v3.3.0

func (i JobQueryArgs) ToJobQueryPtrOutput() JobQueryPtrOutput

func (JobQueryArgs) ToJobQueryPtrOutputWithContext added in v3.3.0

func (i JobQueryArgs) ToJobQueryPtrOutputWithContext(ctx context.Context) JobQueryPtrOutput

type JobQueryDefaultDataset added in v3.3.0

type JobQueryDefaultDataset struct {
	// The ID of the dataset containing this model.
	DatasetId string `pulumi:"datasetId"`
	// The ID of the project containing this model.
	ProjectId *string `pulumi:"projectId"`
}

type JobQueryDefaultDatasetArgs added in v3.3.0

type JobQueryDefaultDatasetArgs struct {
	// The ID of the dataset containing this model.
	DatasetId pulumi.StringInput `pulumi:"datasetId"`
	// The ID of the project containing this model.
	ProjectId pulumi.StringPtrInput `pulumi:"projectId"`
}

func (JobQueryDefaultDatasetArgs) ElementType added in v3.3.0

func (JobQueryDefaultDatasetArgs) ElementType() reflect.Type

func (JobQueryDefaultDatasetArgs) ToJobQueryDefaultDatasetOutput added in v3.3.0

func (i JobQueryDefaultDatasetArgs) ToJobQueryDefaultDatasetOutput() JobQueryDefaultDatasetOutput

func (JobQueryDefaultDatasetArgs) ToJobQueryDefaultDatasetOutputWithContext added in v3.3.0

func (i JobQueryDefaultDatasetArgs) ToJobQueryDefaultDatasetOutputWithContext(ctx context.Context) JobQueryDefaultDatasetOutput

func (JobQueryDefaultDatasetArgs) ToJobQueryDefaultDatasetPtrOutput added in v3.3.0

func (i JobQueryDefaultDatasetArgs) ToJobQueryDefaultDatasetPtrOutput() JobQueryDefaultDatasetPtrOutput

func (JobQueryDefaultDatasetArgs) ToJobQueryDefaultDatasetPtrOutputWithContext added in v3.3.0

func (i JobQueryDefaultDatasetArgs) ToJobQueryDefaultDatasetPtrOutputWithContext(ctx context.Context) JobQueryDefaultDatasetPtrOutput

type JobQueryDefaultDatasetInput added in v3.3.0

type JobQueryDefaultDatasetInput interface {
	pulumi.Input

	ToJobQueryDefaultDatasetOutput() JobQueryDefaultDatasetOutput
	ToJobQueryDefaultDatasetOutputWithContext(context.Context) JobQueryDefaultDatasetOutput
}

JobQueryDefaultDatasetInput is an input type that accepts JobQueryDefaultDatasetArgs and JobQueryDefaultDatasetOutput values. You can construct a concrete instance of `JobQueryDefaultDatasetInput` via:

JobQueryDefaultDatasetArgs{...}

type JobQueryDefaultDatasetOutput added in v3.3.0

type JobQueryDefaultDatasetOutput struct{ *pulumi.OutputState }

func (JobQueryDefaultDatasetOutput) DatasetId added in v3.3.0

The ID of the dataset containing this model.

func (JobQueryDefaultDatasetOutput) ElementType added in v3.3.0

func (JobQueryDefaultDatasetOutput) ProjectId added in v3.3.0

The ID of the project containing this model.

func (JobQueryDefaultDatasetOutput) ToJobQueryDefaultDatasetOutput added in v3.3.0

func (o JobQueryDefaultDatasetOutput) ToJobQueryDefaultDatasetOutput() JobQueryDefaultDatasetOutput

func (JobQueryDefaultDatasetOutput) ToJobQueryDefaultDatasetOutputWithContext added in v3.3.0

func (o JobQueryDefaultDatasetOutput) ToJobQueryDefaultDatasetOutputWithContext(ctx context.Context) JobQueryDefaultDatasetOutput

func (JobQueryDefaultDatasetOutput) ToJobQueryDefaultDatasetPtrOutput added in v3.3.0

func (o JobQueryDefaultDatasetOutput) ToJobQueryDefaultDatasetPtrOutput() JobQueryDefaultDatasetPtrOutput

func (JobQueryDefaultDatasetOutput) ToJobQueryDefaultDatasetPtrOutputWithContext added in v3.3.0

func (o JobQueryDefaultDatasetOutput) ToJobQueryDefaultDatasetPtrOutputWithContext(ctx context.Context) JobQueryDefaultDatasetPtrOutput

type JobQueryDefaultDatasetPtrInput added in v3.3.0

type JobQueryDefaultDatasetPtrInput interface {
	pulumi.Input

	ToJobQueryDefaultDatasetPtrOutput() JobQueryDefaultDatasetPtrOutput
	ToJobQueryDefaultDatasetPtrOutputWithContext(context.Context) JobQueryDefaultDatasetPtrOutput
}

JobQueryDefaultDatasetPtrInput is an input type that accepts JobQueryDefaultDatasetArgs, JobQueryDefaultDatasetPtr and JobQueryDefaultDatasetPtrOutput values. You can construct a concrete instance of `JobQueryDefaultDatasetPtrInput` via:

		 JobQueryDefaultDatasetArgs{...}

 or:

		 nil

func JobQueryDefaultDatasetPtr added in v3.3.0

func JobQueryDefaultDatasetPtr(v *JobQueryDefaultDatasetArgs) JobQueryDefaultDatasetPtrInput

type JobQueryDefaultDatasetPtrOutput added in v3.3.0

type JobQueryDefaultDatasetPtrOutput struct{ *pulumi.OutputState }

func (JobQueryDefaultDatasetPtrOutput) DatasetId added in v3.3.0

The ID of the dataset containing this model.

func (JobQueryDefaultDatasetPtrOutput) Elem added in v3.3.0

func (JobQueryDefaultDatasetPtrOutput) ElementType added in v3.3.0

func (JobQueryDefaultDatasetPtrOutput) ProjectId added in v3.3.0

The ID of the project containing this model.

func (JobQueryDefaultDatasetPtrOutput) ToJobQueryDefaultDatasetPtrOutput added in v3.3.0

func (o JobQueryDefaultDatasetPtrOutput) ToJobQueryDefaultDatasetPtrOutput() JobQueryDefaultDatasetPtrOutput

func (JobQueryDefaultDatasetPtrOutput) ToJobQueryDefaultDatasetPtrOutputWithContext added in v3.3.0

func (o JobQueryDefaultDatasetPtrOutput) ToJobQueryDefaultDatasetPtrOutputWithContext(ctx context.Context) JobQueryDefaultDatasetPtrOutput

type JobQueryDestinationEncryptionConfiguration added in v3.3.0

type JobQueryDestinationEncryptionConfiguration struct {
	// Describes the Cloud KMS encryption key that will be used to protect destination BigQuery table.
	// The BigQuery Service Account associated with your project requires access to this encryption key.
	KmsKeyName string `pulumi:"kmsKeyName"`
}

type JobQueryDestinationEncryptionConfigurationArgs added in v3.3.0

type JobQueryDestinationEncryptionConfigurationArgs struct {
	// Describes the Cloud KMS encryption key that will be used to protect destination BigQuery table.
	// The BigQuery Service Account associated with your project requires access to this encryption key.
	KmsKeyName pulumi.StringInput `pulumi:"kmsKeyName"`
}

func (JobQueryDestinationEncryptionConfigurationArgs) ElementType added in v3.3.0

func (JobQueryDestinationEncryptionConfigurationArgs) ToJobQueryDestinationEncryptionConfigurationOutput added in v3.3.0

func (i JobQueryDestinationEncryptionConfigurationArgs) ToJobQueryDestinationEncryptionConfigurationOutput() JobQueryDestinationEncryptionConfigurationOutput

func (JobQueryDestinationEncryptionConfigurationArgs) ToJobQueryDestinationEncryptionConfigurationOutputWithContext added in v3.3.0

func (i JobQueryDestinationEncryptionConfigurationArgs) ToJobQueryDestinationEncryptionConfigurationOutputWithContext(ctx context.Context) JobQueryDestinationEncryptionConfigurationOutput

func (JobQueryDestinationEncryptionConfigurationArgs) ToJobQueryDestinationEncryptionConfigurationPtrOutput added in v3.3.0

func (i JobQueryDestinationEncryptionConfigurationArgs) ToJobQueryDestinationEncryptionConfigurationPtrOutput() JobQueryDestinationEncryptionConfigurationPtrOutput

func (JobQueryDestinationEncryptionConfigurationArgs) ToJobQueryDestinationEncryptionConfigurationPtrOutputWithContext added in v3.3.0

func (i JobQueryDestinationEncryptionConfigurationArgs) ToJobQueryDestinationEncryptionConfigurationPtrOutputWithContext(ctx context.Context) JobQueryDestinationEncryptionConfigurationPtrOutput

type JobQueryDestinationEncryptionConfigurationInput added in v3.3.0

type JobQueryDestinationEncryptionConfigurationInput interface {
	pulumi.Input

	ToJobQueryDestinationEncryptionConfigurationOutput() JobQueryDestinationEncryptionConfigurationOutput
	ToJobQueryDestinationEncryptionConfigurationOutputWithContext(context.Context) JobQueryDestinationEncryptionConfigurationOutput
}

JobQueryDestinationEncryptionConfigurationInput is an input type that accepts JobQueryDestinationEncryptionConfigurationArgs and JobQueryDestinationEncryptionConfigurationOutput values. You can construct a concrete instance of `JobQueryDestinationEncryptionConfigurationInput` via:

JobQueryDestinationEncryptionConfigurationArgs{...}

type JobQueryDestinationEncryptionConfigurationOutput added in v3.3.0

type JobQueryDestinationEncryptionConfigurationOutput struct{ *pulumi.OutputState }

func (JobQueryDestinationEncryptionConfigurationOutput) ElementType added in v3.3.0

func (JobQueryDestinationEncryptionConfigurationOutput) KmsKeyName added in v3.3.0

Describes the Cloud KMS encryption key that will be used to protect destination BigQuery table. The BigQuery Service Account associated with your project requires access to this encryption key.

func (JobQueryDestinationEncryptionConfigurationOutput) ToJobQueryDestinationEncryptionConfigurationOutput added in v3.3.0

func (o JobQueryDestinationEncryptionConfigurationOutput) ToJobQueryDestinationEncryptionConfigurationOutput() JobQueryDestinationEncryptionConfigurationOutput

func (JobQueryDestinationEncryptionConfigurationOutput) ToJobQueryDestinationEncryptionConfigurationOutputWithContext added in v3.3.0

func (o JobQueryDestinationEncryptionConfigurationOutput) ToJobQueryDestinationEncryptionConfigurationOutputWithContext(ctx context.Context) JobQueryDestinationEncryptionConfigurationOutput

func (JobQueryDestinationEncryptionConfigurationOutput) ToJobQueryDestinationEncryptionConfigurationPtrOutput added in v3.3.0

func (o JobQueryDestinationEncryptionConfigurationOutput) ToJobQueryDestinationEncryptionConfigurationPtrOutput() JobQueryDestinationEncryptionConfigurationPtrOutput

func (JobQueryDestinationEncryptionConfigurationOutput) ToJobQueryDestinationEncryptionConfigurationPtrOutputWithContext added in v3.3.0

func (o JobQueryDestinationEncryptionConfigurationOutput) ToJobQueryDestinationEncryptionConfigurationPtrOutputWithContext(ctx context.Context) JobQueryDestinationEncryptionConfigurationPtrOutput

type JobQueryDestinationEncryptionConfigurationPtrInput added in v3.3.0

type JobQueryDestinationEncryptionConfigurationPtrInput interface {
	pulumi.Input

	ToJobQueryDestinationEncryptionConfigurationPtrOutput() JobQueryDestinationEncryptionConfigurationPtrOutput
	ToJobQueryDestinationEncryptionConfigurationPtrOutputWithContext(context.Context) JobQueryDestinationEncryptionConfigurationPtrOutput
}

JobQueryDestinationEncryptionConfigurationPtrInput is an input type that accepts JobQueryDestinationEncryptionConfigurationArgs, JobQueryDestinationEncryptionConfigurationPtr and JobQueryDestinationEncryptionConfigurationPtrOutput values. You can construct a concrete instance of `JobQueryDestinationEncryptionConfigurationPtrInput` via:

		 JobQueryDestinationEncryptionConfigurationArgs{...}

 or:

		 nil

type JobQueryDestinationEncryptionConfigurationPtrOutput added in v3.3.0

type JobQueryDestinationEncryptionConfigurationPtrOutput struct{ *pulumi.OutputState }

func (JobQueryDestinationEncryptionConfigurationPtrOutput) Elem added in v3.3.0

func (JobQueryDestinationEncryptionConfigurationPtrOutput) ElementType added in v3.3.0

func (JobQueryDestinationEncryptionConfigurationPtrOutput) KmsKeyName added in v3.3.0

Describes the Cloud KMS encryption key that will be used to protect destination BigQuery table. The BigQuery Service Account associated with your project requires access to this encryption key.

func (JobQueryDestinationEncryptionConfigurationPtrOutput) ToJobQueryDestinationEncryptionConfigurationPtrOutput added in v3.3.0

func (o JobQueryDestinationEncryptionConfigurationPtrOutput) ToJobQueryDestinationEncryptionConfigurationPtrOutput() JobQueryDestinationEncryptionConfigurationPtrOutput

func (JobQueryDestinationEncryptionConfigurationPtrOutput) ToJobQueryDestinationEncryptionConfigurationPtrOutputWithContext added in v3.3.0

func (o JobQueryDestinationEncryptionConfigurationPtrOutput) ToJobQueryDestinationEncryptionConfigurationPtrOutputWithContext(ctx context.Context) JobQueryDestinationEncryptionConfigurationPtrOutput

type JobQueryDestinationTable added in v3.3.0

type JobQueryDestinationTable struct {
	// The ID of the dataset containing this model.
	DatasetId string `pulumi:"datasetId"`
	// The ID of the project containing this model.
	ProjectId string `pulumi:"projectId"`
	// The ID of the table.
	TableId string `pulumi:"tableId"`
}

type JobQueryDestinationTableArgs added in v3.3.0

type JobQueryDestinationTableArgs struct {
	// The ID of the dataset containing this model.
	DatasetId pulumi.StringInput `pulumi:"datasetId"`
	// The ID of the project containing this model.
	ProjectId pulumi.StringInput `pulumi:"projectId"`
	// The ID of the table.
	TableId pulumi.StringInput `pulumi:"tableId"`
}

func (JobQueryDestinationTableArgs) ElementType added in v3.3.0

func (JobQueryDestinationTableArgs) ToJobQueryDestinationTableOutput added in v3.3.0

func (i JobQueryDestinationTableArgs) ToJobQueryDestinationTableOutput() JobQueryDestinationTableOutput

func (JobQueryDestinationTableArgs) ToJobQueryDestinationTableOutputWithContext added in v3.3.0

func (i JobQueryDestinationTableArgs) ToJobQueryDestinationTableOutputWithContext(ctx context.Context) JobQueryDestinationTableOutput

func (JobQueryDestinationTableArgs) ToJobQueryDestinationTablePtrOutput added in v3.3.0

func (i JobQueryDestinationTableArgs) ToJobQueryDestinationTablePtrOutput() JobQueryDestinationTablePtrOutput

func (JobQueryDestinationTableArgs) ToJobQueryDestinationTablePtrOutputWithContext added in v3.3.0

func (i JobQueryDestinationTableArgs) ToJobQueryDestinationTablePtrOutputWithContext(ctx context.Context) JobQueryDestinationTablePtrOutput

type JobQueryDestinationTableInput added in v3.3.0

type JobQueryDestinationTableInput interface {
	pulumi.Input

	ToJobQueryDestinationTableOutput() JobQueryDestinationTableOutput
	ToJobQueryDestinationTableOutputWithContext(context.Context) JobQueryDestinationTableOutput
}

JobQueryDestinationTableInput is an input type that accepts JobQueryDestinationTableArgs and JobQueryDestinationTableOutput values. You can construct a concrete instance of `JobQueryDestinationTableInput` via:

JobQueryDestinationTableArgs{...}

type JobQueryDestinationTableOutput added in v3.3.0

type JobQueryDestinationTableOutput struct{ *pulumi.OutputState }

func (JobQueryDestinationTableOutput) DatasetId added in v3.3.0

The ID of the dataset containing this model.

func (JobQueryDestinationTableOutput) ElementType added in v3.3.0

func (JobQueryDestinationTableOutput) ProjectId added in v3.3.0

The ID of the project containing this model.

func (JobQueryDestinationTableOutput) TableId added in v3.3.0

The ID of the table.

func (JobQueryDestinationTableOutput) ToJobQueryDestinationTableOutput added in v3.3.0

func (o JobQueryDestinationTableOutput) ToJobQueryDestinationTableOutput() JobQueryDestinationTableOutput

func (JobQueryDestinationTableOutput) ToJobQueryDestinationTableOutputWithContext added in v3.3.0

func (o JobQueryDestinationTableOutput) ToJobQueryDestinationTableOutputWithContext(ctx context.Context) JobQueryDestinationTableOutput

func (JobQueryDestinationTableOutput) ToJobQueryDestinationTablePtrOutput added in v3.3.0

func (o JobQueryDestinationTableOutput) ToJobQueryDestinationTablePtrOutput() JobQueryDestinationTablePtrOutput

func (JobQueryDestinationTableOutput) ToJobQueryDestinationTablePtrOutputWithContext added in v3.3.0

func (o JobQueryDestinationTableOutput) ToJobQueryDestinationTablePtrOutputWithContext(ctx context.Context) JobQueryDestinationTablePtrOutput

type JobQueryDestinationTablePtrInput added in v3.3.0

type JobQueryDestinationTablePtrInput interface {
	pulumi.Input

	ToJobQueryDestinationTablePtrOutput() JobQueryDestinationTablePtrOutput
	ToJobQueryDestinationTablePtrOutputWithContext(context.Context) JobQueryDestinationTablePtrOutput
}

JobQueryDestinationTablePtrInput is an input type that accepts JobQueryDestinationTableArgs, JobQueryDestinationTablePtr and JobQueryDestinationTablePtrOutput values. You can construct a concrete instance of `JobQueryDestinationTablePtrInput` via:

		 JobQueryDestinationTableArgs{...}

 or:

		 nil

func JobQueryDestinationTablePtr added in v3.3.0

func JobQueryDestinationTablePtr(v *JobQueryDestinationTableArgs) JobQueryDestinationTablePtrInput

type JobQueryDestinationTablePtrOutput added in v3.3.0

type JobQueryDestinationTablePtrOutput struct{ *pulumi.OutputState }

func (JobQueryDestinationTablePtrOutput) DatasetId added in v3.3.0

The ID of the dataset containing this model.

func (JobQueryDestinationTablePtrOutput) Elem added in v3.3.0

func (JobQueryDestinationTablePtrOutput) ElementType added in v3.3.0

func (JobQueryDestinationTablePtrOutput) ProjectId added in v3.3.0

The ID of the project containing this model.

func (JobQueryDestinationTablePtrOutput) TableId added in v3.3.0

The ID of the table.

func (JobQueryDestinationTablePtrOutput) ToJobQueryDestinationTablePtrOutput added in v3.3.0

func (o JobQueryDestinationTablePtrOutput) ToJobQueryDestinationTablePtrOutput() JobQueryDestinationTablePtrOutput

func (JobQueryDestinationTablePtrOutput) ToJobQueryDestinationTablePtrOutputWithContext added in v3.3.0

func (o JobQueryDestinationTablePtrOutput) ToJobQueryDestinationTablePtrOutputWithContext(ctx context.Context) JobQueryDestinationTablePtrOutput

type JobQueryInput added in v3.3.0

type JobQueryInput interface {
	pulumi.Input

	ToJobQueryOutput() JobQueryOutput
	ToJobQueryOutputWithContext(context.Context) JobQueryOutput
}

JobQueryInput is an input type that accepts JobQueryArgs and JobQueryOutput values. You can construct a concrete instance of `JobQueryInput` via:

JobQueryArgs{...}

type JobQueryOutput added in v3.3.0

type JobQueryOutput struct{ *pulumi.OutputState }

func (JobQueryOutput) AllowLargeResults added in v3.3.0

func (o JobQueryOutput) AllowLargeResults() pulumi.BoolPtrOutput

If true and query uses legacy SQL dialect, allows the query to produce arbitrarily large result tables at a slight cost in performance. Requires destinationTable to be set. For standard SQL queries, this flag is ignored and large results are always allowed. However, you must still set destinationTable when result size exceeds the allowed maximum response size.

func (JobQueryOutput) CreateDisposition added in v3.3.0

func (o JobQueryOutput) CreateDisposition() pulumi.StringPtrOutput

Specifies whether the job is allowed to create new tables. The following values are supported: CREATE_IF_NEEDED: If the table does not exist, BigQuery creates the table. CREATE_NEVER: The table must already exist. If it does not, a 'notFound' error is returned in the job result. Creation, truncation and append actions occur as one atomic update upon job completion

func (JobQueryOutput) DefaultDataset added in v3.3.0

Specifies the default dataset to use for unqualified table names in the query. Note that this does not alter behavior of unqualified dataset names. Structure is documented below.

func (JobQueryOutput) DestinationEncryptionConfiguration added in v3.3.0

func (o JobQueryOutput) DestinationEncryptionConfiguration() JobQueryDestinationEncryptionConfigurationPtrOutput

Custom encryption configuration (e.g., Cloud KMS keys) Structure is documented below.

func (JobQueryOutput) DestinationTable added in v3.3.0

The destination table. Structure is documented below.

func (JobQueryOutput) ElementType added in v3.3.0

func (JobQueryOutput) ElementType() reflect.Type

func (JobQueryOutput) FlattenResults added in v3.3.0

func (o JobQueryOutput) FlattenResults() pulumi.BoolPtrOutput

If true and query uses legacy SQL dialect, flattens all nested and repeated fields in the query results. allowLargeResults must be true if this is set to false. For standard SQL queries, this flag is ignored and results are never flattened.

func (JobQueryOutput) MaximumBillingTier added in v3.3.0

func (o JobQueryOutput) MaximumBillingTier() pulumi.IntPtrOutput

Limits the billing tier for this job. Queries that have resource usage beyond this tier will fail (without incurring a charge). If unspecified, this will be set to your project default.

func (JobQueryOutput) MaximumBytesBilled added in v3.3.0

func (o JobQueryOutput) MaximumBytesBilled() pulumi.StringPtrOutput

Limits the bytes billed for this job. Queries that will have bytes billed beyond this limit will fail (without incurring a charge). If unspecified, this will be set to your project default.

func (JobQueryOutput) ParameterMode added in v3.3.0

func (o JobQueryOutput) ParameterMode() pulumi.StringPtrOutput

Standard SQL only. Set to POSITIONAL to use positional (?) query parameters or to NAMED to use named (@myparam) query parameters in this query.

func (JobQueryOutput) Priority added in v3.3.0

func (o JobQueryOutput) Priority() pulumi.StringPtrOutput

Specifies a priority for the query.

func (JobQueryOutput) Query added in v3.3.0

Configures a query job. Structure is documented below.

func (JobQueryOutput) SchemaUpdateOptions added in v3.3.0

func (o JobQueryOutput) SchemaUpdateOptions() pulumi.StringArrayOutput

Allows the schema of the destination table to be updated as a side effect of the load job if a schema is autodetected or supplied in the job configuration. Schema update options are supported in two cases: when writeDisposition is WRITE_APPEND; when writeDisposition is WRITE_TRUNCATE and the destination table is a partition of a table, specified by partition decorators. For normal tables, WRITE_TRUNCATE will always overwrite the schema. One or more of the following values are specified: ALLOW_FIELD_ADDITION: allow adding a nullable field to the schema. ALLOW_FIELD_RELAXATION: allow relaxing a required field in the original schema to nullable.

func (JobQueryOutput) ScriptOptions added in v3.3.0

Options controlling the execution of scripts. Structure is documented below.

func (JobQueryOutput) ToJobQueryOutput added in v3.3.0

func (o JobQueryOutput) ToJobQueryOutput() JobQueryOutput

func (JobQueryOutput) ToJobQueryOutputWithContext added in v3.3.0

func (o JobQueryOutput) ToJobQueryOutputWithContext(ctx context.Context) JobQueryOutput

func (JobQueryOutput) ToJobQueryPtrOutput added in v3.3.0

func (o JobQueryOutput) ToJobQueryPtrOutput() JobQueryPtrOutput

func (JobQueryOutput) ToJobQueryPtrOutputWithContext added in v3.3.0

func (o JobQueryOutput) ToJobQueryPtrOutputWithContext(ctx context.Context) JobQueryPtrOutput

func (JobQueryOutput) UseLegacySql added in v3.3.0

func (o JobQueryOutput) UseLegacySql() pulumi.BoolPtrOutput

Specifies whether to use BigQuery's legacy SQL dialect for this query. The default value is true. If set to false, the query will use BigQuery's standard SQL.

func (JobQueryOutput) UseQueryCache added in v3.3.0

func (o JobQueryOutput) UseQueryCache() pulumi.BoolPtrOutput

Whether to look for the result in the query cache. The query cache is a best-effort cache that will be flushed whenever tables in the query are modified. Moreover, the query cache is only available when a query does not have a destination table specified. The default value is true.

func (JobQueryOutput) UserDefinedFunctionResources added in v3.3.0

func (o JobQueryOutput) UserDefinedFunctionResources() JobQueryUserDefinedFunctionResourceArrayOutput

Describes user-defined function resources used in the query. Structure is documented below.

func (JobQueryOutput) WriteDisposition added in v3.3.0

func (o JobQueryOutput) WriteDisposition() pulumi.StringPtrOutput

Specifies the action that occurs if the destination table already exists. The following values are supported: WRITE_TRUNCATE: If the table already exists, BigQuery overwrites the table data and uses the schema from the query result. WRITE_APPEND: If the table already exists, BigQuery appends the data to the table. WRITE_EMPTY: If the table already exists and contains data, a 'duplicate' error is returned in the job result. Each action is atomic and only occurs if BigQuery is able to complete the job successfully. Creation, truncation and append actions occur as one atomic update upon job completion.

type JobQueryPtrInput added in v3.3.0

type JobQueryPtrInput interface {
	pulumi.Input

	ToJobQueryPtrOutput() JobQueryPtrOutput
	ToJobQueryPtrOutputWithContext(context.Context) JobQueryPtrOutput
}

JobQueryPtrInput is an input type that accepts JobQueryArgs, JobQueryPtr and JobQueryPtrOutput values. You can construct a concrete instance of `JobQueryPtrInput` via:

		 JobQueryArgs{...}

 or:

		 nil

func JobQueryPtr added in v3.3.0

func JobQueryPtr(v *JobQueryArgs) JobQueryPtrInput

type JobQueryPtrOutput added in v3.3.0

type JobQueryPtrOutput struct{ *pulumi.OutputState }

func (JobQueryPtrOutput) AllowLargeResults added in v3.3.0

func (o JobQueryPtrOutput) AllowLargeResults() pulumi.BoolPtrOutput

If true and query uses legacy SQL dialect, allows the query to produce arbitrarily large result tables at a slight cost in performance. Requires destinationTable to be set. For standard SQL queries, this flag is ignored and large results are always allowed. However, you must still set destinationTable when result size exceeds the allowed maximum response size.

func (JobQueryPtrOutput) CreateDisposition added in v3.3.0

func (o JobQueryPtrOutput) CreateDisposition() pulumi.StringPtrOutput

Specifies whether the job is allowed to create new tables. The following values are supported: CREATE_IF_NEEDED: If the table does not exist, BigQuery creates the table. CREATE_NEVER: The table must already exist. If it does not, a 'notFound' error is returned in the job result. Creation, truncation and append actions occur as one atomic update upon job completion

func (JobQueryPtrOutput) DefaultDataset added in v3.3.0

Specifies the default dataset to use for unqualified table names in the query. Note that this does not alter behavior of unqualified dataset names. Structure is documented below.

func (JobQueryPtrOutput) DestinationEncryptionConfiguration added in v3.3.0

func (o JobQueryPtrOutput) DestinationEncryptionConfiguration() JobQueryDestinationEncryptionConfigurationPtrOutput

Custom encryption configuration (e.g., Cloud KMS keys) Structure is documented below.

func (JobQueryPtrOutput) DestinationTable added in v3.3.0

The destination table. Structure is documented below.

func (JobQueryPtrOutput) Elem added in v3.3.0

func (JobQueryPtrOutput) ElementType added in v3.3.0

func (JobQueryPtrOutput) ElementType() reflect.Type

func (JobQueryPtrOutput) FlattenResults added in v3.3.0

func (o JobQueryPtrOutput) FlattenResults() pulumi.BoolPtrOutput

If true and query uses legacy SQL dialect, flattens all nested and repeated fields in the query results. allowLargeResults must be true if this is set to false. For standard SQL queries, this flag is ignored and results are never flattened.

func (JobQueryPtrOutput) MaximumBillingTier added in v3.3.0

func (o JobQueryPtrOutput) MaximumBillingTier() pulumi.IntPtrOutput

Limits the billing tier for this job. Queries that have resource usage beyond this tier will fail (without incurring a charge). If unspecified, this will be set to your project default.

func (JobQueryPtrOutput) MaximumBytesBilled added in v3.3.0

func (o JobQueryPtrOutput) MaximumBytesBilled() pulumi.StringPtrOutput

Limits the bytes billed for this job. Queries that will have bytes billed beyond this limit will fail (without incurring a charge). If unspecified, this will be set to your project default.

func (JobQueryPtrOutput) ParameterMode added in v3.3.0

func (o JobQueryPtrOutput) ParameterMode() pulumi.StringPtrOutput

Standard SQL only. Set to POSITIONAL to use positional (?) query parameters or to NAMED to use named (@myparam) query parameters in this query.

func (JobQueryPtrOutput) Priority added in v3.3.0

Specifies a priority for the query.

func (JobQueryPtrOutput) Query added in v3.3.0

Configures a query job. Structure is documented below.

func (JobQueryPtrOutput) SchemaUpdateOptions added in v3.3.0

func (o JobQueryPtrOutput) SchemaUpdateOptions() pulumi.StringArrayOutput

Allows the schema of the destination table to be updated as a side effect of the load job if a schema is autodetected or supplied in the job configuration. Schema update options are supported in two cases: when writeDisposition is WRITE_APPEND; when writeDisposition is WRITE_TRUNCATE and the destination table is a partition of a table, specified by partition decorators. For normal tables, WRITE_TRUNCATE will always overwrite the schema. One or more of the following values are specified: ALLOW_FIELD_ADDITION: allow adding a nullable field to the schema. ALLOW_FIELD_RELAXATION: allow relaxing a required field in the original schema to nullable.

func (JobQueryPtrOutput) ScriptOptions added in v3.3.0

Options controlling the execution of scripts. Structure is documented below.

func (JobQueryPtrOutput) ToJobQueryPtrOutput added in v3.3.0

func (o JobQueryPtrOutput) ToJobQueryPtrOutput() JobQueryPtrOutput

func (JobQueryPtrOutput) ToJobQueryPtrOutputWithContext added in v3.3.0

func (o JobQueryPtrOutput) ToJobQueryPtrOutputWithContext(ctx context.Context) JobQueryPtrOutput

func (JobQueryPtrOutput) UseLegacySql added in v3.3.0

func (o JobQueryPtrOutput) UseLegacySql() pulumi.BoolPtrOutput

Specifies whether to use BigQuery's legacy SQL dialect for this query. The default value is true. If set to false, the query will use BigQuery's standard SQL.

func (JobQueryPtrOutput) UseQueryCache added in v3.3.0

func (o JobQueryPtrOutput) UseQueryCache() pulumi.BoolPtrOutput

Whether to look for the result in the query cache. The query cache is a best-effort cache that will be flushed whenever tables in the query are modified. Moreover, the query cache is only available when a query does not have a destination table specified. The default value is true.

func (JobQueryPtrOutput) UserDefinedFunctionResources added in v3.3.0

func (o JobQueryPtrOutput) UserDefinedFunctionResources() JobQueryUserDefinedFunctionResourceArrayOutput

Describes user-defined function resources used in the query. Structure is documented below.

func (JobQueryPtrOutput) WriteDisposition added in v3.3.0

func (o JobQueryPtrOutput) WriteDisposition() pulumi.StringPtrOutput

Specifies the action that occurs if the destination table already exists. The following values are supported: WRITE_TRUNCATE: If the table already exists, BigQuery overwrites the table data and uses the schema from the query result. WRITE_APPEND: If the table already exists, BigQuery appends the data to the table. WRITE_EMPTY: If the table already exists and contains data, a 'duplicate' error is returned in the job result. Each action is atomic and only occurs if BigQuery is able to complete the job successfully. Creation, truncation and append actions occur as one atomic update upon job completion.

type JobQueryScriptOptions added in v3.3.0

type JobQueryScriptOptions struct {
	// Determines which statement in the script represents the "key result",
	// used to populate the schema and query results of the script job.
	KeyResultStatement *string `pulumi:"keyResultStatement"`
	// Limit on the number of bytes billed per statement. Exceeding this budget results in an error.
	StatementByteBudget *string `pulumi:"statementByteBudget"`
	// Timeout period for each statement in a script.
	StatementTimeoutMs *string `pulumi:"statementTimeoutMs"`
}

type JobQueryScriptOptionsArgs added in v3.3.0

type JobQueryScriptOptionsArgs struct {
	// Determines which statement in the script represents the "key result",
	// used to populate the schema and query results of the script job.
	KeyResultStatement pulumi.StringPtrInput `pulumi:"keyResultStatement"`
	// Limit on the number of bytes billed per statement. Exceeding this budget results in an error.
	StatementByteBudget pulumi.StringPtrInput `pulumi:"statementByteBudget"`
	// Timeout period for each statement in a script.
	StatementTimeoutMs pulumi.StringPtrInput `pulumi:"statementTimeoutMs"`
}

func (JobQueryScriptOptionsArgs) ElementType added in v3.3.0

func (JobQueryScriptOptionsArgs) ElementType() reflect.Type

func (JobQueryScriptOptionsArgs) ToJobQueryScriptOptionsOutput added in v3.3.0

func (i JobQueryScriptOptionsArgs) ToJobQueryScriptOptionsOutput() JobQueryScriptOptionsOutput

func (JobQueryScriptOptionsArgs) ToJobQueryScriptOptionsOutputWithContext added in v3.3.0

func (i JobQueryScriptOptionsArgs) ToJobQueryScriptOptionsOutputWithContext(ctx context.Context) JobQueryScriptOptionsOutput

func (JobQueryScriptOptionsArgs) ToJobQueryScriptOptionsPtrOutput added in v3.3.0

func (i JobQueryScriptOptionsArgs) ToJobQueryScriptOptionsPtrOutput() JobQueryScriptOptionsPtrOutput

func (JobQueryScriptOptionsArgs) ToJobQueryScriptOptionsPtrOutputWithContext added in v3.3.0

func (i JobQueryScriptOptionsArgs) ToJobQueryScriptOptionsPtrOutputWithContext(ctx context.Context) JobQueryScriptOptionsPtrOutput

type JobQueryScriptOptionsInput added in v3.3.0

type JobQueryScriptOptionsInput interface {
	pulumi.Input

	ToJobQueryScriptOptionsOutput() JobQueryScriptOptionsOutput
	ToJobQueryScriptOptionsOutputWithContext(context.Context) JobQueryScriptOptionsOutput
}

JobQueryScriptOptionsInput is an input type that accepts JobQueryScriptOptionsArgs and JobQueryScriptOptionsOutput values. You can construct a concrete instance of `JobQueryScriptOptionsInput` via:

JobQueryScriptOptionsArgs{...}

type JobQueryScriptOptionsOutput added in v3.3.0

type JobQueryScriptOptionsOutput struct{ *pulumi.OutputState }

func (JobQueryScriptOptionsOutput) ElementType added in v3.3.0

func (JobQueryScriptOptionsOutput) KeyResultStatement added in v3.3.0

func (o JobQueryScriptOptionsOutput) KeyResultStatement() pulumi.StringPtrOutput

Determines which statement in the script represents the "key result", used to populate the schema and query results of the script job.

func (JobQueryScriptOptionsOutput) StatementByteBudget added in v3.3.0

func (o JobQueryScriptOptionsOutput) StatementByteBudget() pulumi.StringPtrOutput

Limit on the number of bytes billed per statement. Exceeding this budget results in an error.

func (JobQueryScriptOptionsOutput) StatementTimeoutMs added in v3.3.0

func (o JobQueryScriptOptionsOutput) StatementTimeoutMs() pulumi.StringPtrOutput

Timeout period for each statement in a script.

func (JobQueryScriptOptionsOutput) ToJobQueryScriptOptionsOutput added in v3.3.0

func (o JobQueryScriptOptionsOutput) ToJobQueryScriptOptionsOutput() JobQueryScriptOptionsOutput

func (JobQueryScriptOptionsOutput) ToJobQueryScriptOptionsOutputWithContext added in v3.3.0

func (o JobQueryScriptOptionsOutput) ToJobQueryScriptOptionsOutputWithContext(ctx context.Context) JobQueryScriptOptionsOutput

func (JobQueryScriptOptionsOutput) ToJobQueryScriptOptionsPtrOutput added in v3.3.0

func (o JobQueryScriptOptionsOutput) ToJobQueryScriptOptionsPtrOutput() JobQueryScriptOptionsPtrOutput

func (JobQueryScriptOptionsOutput) ToJobQueryScriptOptionsPtrOutputWithContext added in v3.3.0

func (o JobQueryScriptOptionsOutput) ToJobQueryScriptOptionsPtrOutputWithContext(ctx context.Context) JobQueryScriptOptionsPtrOutput

type JobQueryScriptOptionsPtrInput added in v3.3.0

type JobQueryScriptOptionsPtrInput interface {
	pulumi.Input

	ToJobQueryScriptOptionsPtrOutput() JobQueryScriptOptionsPtrOutput
	ToJobQueryScriptOptionsPtrOutputWithContext(context.Context) JobQueryScriptOptionsPtrOutput
}

JobQueryScriptOptionsPtrInput is an input type that accepts JobQueryScriptOptionsArgs, JobQueryScriptOptionsPtr and JobQueryScriptOptionsPtrOutput values. You can construct a concrete instance of `JobQueryScriptOptionsPtrInput` via:

		 JobQueryScriptOptionsArgs{...}

 or:

		 nil

func JobQueryScriptOptionsPtr added in v3.3.0

func JobQueryScriptOptionsPtr(v *JobQueryScriptOptionsArgs) JobQueryScriptOptionsPtrInput

type JobQueryScriptOptionsPtrOutput added in v3.3.0

type JobQueryScriptOptionsPtrOutput struct{ *pulumi.OutputState }

func (JobQueryScriptOptionsPtrOutput) Elem added in v3.3.0

func (JobQueryScriptOptionsPtrOutput) ElementType added in v3.3.0

func (JobQueryScriptOptionsPtrOutput) KeyResultStatement added in v3.3.0

func (o JobQueryScriptOptionsPtrOutput) KeyResultStatement() pulumi.StringPtrOutput

Determines which statement in the script represents the "key result", used to populate the schema and query results of the script job.

func (JobQueryScriptOptionsPtrOutput) StatementByteBudget added in v3.3.0

func (o JobQueryScriptOptionsPtrOutput) StatementByteBudget() pulumi.StringPtrOutput

Limit on the number of bytes billed per statement. Exceeding this budget results in an error.

func (JobQueryScriptOptionsPtrOutput) StatementTimeoutMs added in v3.3.0

func (o JobQueryScriptOptionsPtrOutput) StatementTimeoutMs() pulumi.StringPtrOutput

Timeout period for each statement in a script.

func (JobQueryScriptOptionsPtrOutput) ToJobQueryScriptOptionsPtrOutput added in v3.3.0

func (o JobQueryScriptOptionsPtrOutput) ToJobQueryScriptOptionsPtrOutput() JobQueryScriptOptionsPtrOutput

func (JobQueryScriptOptionsPtrOutput) ToJobQueryScriptOptionsPtrOutputWithContext added in v3.3.0

func (o JobQueryScriptOptionsPtrOutput) ToJobQueryScriptOptionsPtrOutputWithContext(ctx context.Context) JobQueryScriptOptionsPtrOutput

type JobQueryUserDefinedFunctionResource added in v3.3.0

type JobQueryUserDefinedFunctionResource struct {
	// An inline resource that contains code for a user-defined function (UDF).
	// Providing a inline code resource is equivalent to providing a URI for a file containing the same code.
	InlineCode *string `pulumi:"inlineCode"`
	// A code resource to load from a Google Cloud Storage URI (gs://bucket/path).
	ResourceUri *string `pulumi:"resourceUri"`
}

type JobQueryUserDefinedFunctionResourceArgs added in v3.3.0

type JobQueryUserDefinedFunctionResourceArgs struct {
	// An inline resource that contains code for a user-defined function (UDF).
	// Providing a inline code resource is equivalent to providing a URI for a file containing the same code.
	InlineCode pulumi.StringPtrInput `pulumi:"inlineCode"`
	// A code resource to load from a Google Cloud Storage URI (gs://bucket/path).
	ResourceUri pulumi.StringPtrInput `pulumi:"resourceUri"`
}

func (JobQueryUserDefinedFunctionResourceArgs) ElementType added in v3.3.0

func (JobQueryUserDefinedFunctionResourceArgs) ToJobQueryUserDefinedFunctionResourceOutput added in v3.3.0

func (i JobQueryUserDefinedFunctionResourceArgs) ToJobQueryUserDefinedFunctionResourceOutput() JobQueryUserDefinedFunctionResourceOutput

func (JobQueryUserDefinedFunctionResourceArgs) ToJobQueryUserDefinedFunctionResourceOutputWithContext added in v3.3.0

func (i JobQueryUserDefinedFunctionResourceArgs) ToJobQueryUserDefinedFunctionResourceOutputWithContext(ctx context.Context) JobQueryUserDefinedFunctionResourceOutput

type JobQueryUserDefinedFunctionResourceArray added in v3.3.0

type JobQueryUserDefinedFunctionResourceArray []JobQueryUserDefinedFunctionResourceInput

func (JobQueryUserDefinedFunctionResourceArray) ElementType added in v3.3.0

func (JobQueryUserDefinedFunctionResourceArray) ToJobQueryUserDefinedFunctionResourceArrayOutput added in v3.3.0

func (i JobQueryUserDefinedFunctionResourceArray) ToJobQueryUserDefinedFunctionResourceArrayOutput() JobQueryUserDefinedFunctionResourceArrayOutput

func (JobQueryUserDefinedFunctionResourceArray) ToJobQueryUserDefinedFunctionResourceArrayOutputWithContext added in v3.3.0

func (i JobQueryUserDefinedFunctionResourceArray) ToJobQueryUserDefinedFunctionResourceArrayOutputWithContext(ctx context.Context) JobQueryUserDefinedFunctionResourceArrayOutput

type JobQueryUserDefinedFunctionResourceArrayInput added in v3.3.0

type JobQueryUserDefinedFunctionResourceArrayInput interface {
	pulumi.Input

	ToJobQueryUserDefinedFunctionResourceArrayOutput() JobQueryUserDefinedFunctionResourceArrayOutput
	ToJobQueryUserDefinedFunctionResourceArrayOutputWithContext(context.Context) JobQueryUserDefinedFunctionResourceArrayOutput
}

JobQueryUserDefinedFunctionResourceArrayInput is an input type that accepts JobQueryUserDefinedFunctionResourceArray and JobQueryUserDefinedFunctionResourceArrayOutput values. You can construct a concrete instance of `JobQueryUserDefinedFunctionResourceArrayInput` via:

JobQueryUserDefinedFunctionResourceArray{ JobQueryUserDefinedFunctionResourceArgs{...} }

type JobQueryUserDefinedFunctionResourceArrayOutput added in v3.3.0

type JobQueryUserDefinedFunctionResourceArrayOutput struct{ *pulumi.OutputState }

func (JobQueryUserDefinedFunctionResourceArrayOutput) ElementType added in v3.3.0

func (JobQueryUserDefinedFunctionResourceArrayOutput) Index added in v3.3.0

func (JobQueryUserDefinedFunctionResourceArrayOutput) ToJobQueryUserDefinedFunctionResourceArrayOutput added in v3.3.0

func (o JobQueryUserDefinedFunctionResourceArrayOutput) ToJobQueryUserDefinedFunctionResourceArrayOutput() JobQueryUserDefinedFunctionResourceArrayOutput

func (JobQueryUserDefinedFunctionResourceArrayOutput) ToJobQueryUserDefinedFunctionResourceArrayOutputWithContext added in v3.3.0

func (o JobQueryUserDefinedFunctionResourceArrayOutput) ToJobQueryUserDefinedFunctionResourceArrayOutputWithContext(ctx context.Context) JobQueryUserDefinedFunctionResourceArrayOutput

type JobQueryUserDefinedFunctionResourceInput added in v3.3.0

type JobQueryUserDefinedFunctionResourceInput interface {
	pulumi.Input

	ToJobQueryUserDefinedFunctionResourceOutput() JobQueryUserDefinedFunctionResourceOutput
	ToJobQueryUserDefinedFunctionResourceOutputWithContext(context.Context) JobQueryUserDefinedFunctionResourceOutput
}

JobQueryUserDefinedFunctionResourceInput is an input type that accepts JobQueryUserDefinedFunctionResourceArgs and JobQueryUserDefinedFunctionResourceOutput values. You can construct a concrete instance of `JobQueryUserDefinedFunctionResourceInput` via:

JobQueryUserDefinedFunctionResourceArgs{...}

type JobQueryUserDefinedFunctionResourceOutput added in v3.3.0

type JobQueryUserDefinedFunctionResourceOutput struct{ *pulumi.OutputState }

func (JobQueryUserDefinedFunctionResourceOutput) ElementType added in v3.3.0

func (JobQueryUserDefinedFunctionResourceOutput) InlineCode added in v3.3.0

An inline resource that contains code for a user-defined function (UDF). Providing a inline code resource is equivalent to providing a URI for a file containing the same code.

func (JobQueryUserDefinedFunctionResourceOutput) ResourceUri added in v3.3.0

A code resource to load from a Google Cloud Storage URI (gs://bucket/path).

func (JobQueryUserDefinedFunctionResourceOutput) ToJobQueryUserDefinedFunctionResourceOutput added in v3.3.0

func (o JobQueryUserDefinedFunctionResourceOutput) ToJobQueryUserDefinedFunctionResourceOutput() JobQueryUserDefinedFunctionResourceOutput

func (JobQueryUserDefinedFunctionResourceOutput) ToJobQueryUserDefinedFunctionResourceOutputWithContext added in v3.3.0

func (o JobQueryUserDefinedFunctionResourceOutput) ToJobQueryUserDefinedFunctionResourceOutputWithContext(ctx context.Context) JobQueryUserDefinedFunctionResourceOutput

type JobState added in v3.3.0

type JobState struct {
	// Copies a table.  Structure is documented below.
	Copy JobCopyPtrInput
	// Configures an extract job.  Structure is documented below.
	Extract JobExtractPtrInput
	// The ID of the job. The ID must contain only letters (a-z, A-Z), numbers (0-9), underscores (_), or dashes (-). The maximum length is 1,024 characters.
	JobId pulumi.StringPtrInput
	// Job timeout in milliseconds. If this time limit is exceeded, BigQuery may attempt to terminate the job.
	JobTimeoutMs pulumi.StringPtrInput
	// The type of the job.
	JobType pulumi.StringPtrInput
	// The labels associated with this job. You can use these to organize and group your jobs.
	Labels pulumi.StringMapInput
	// Configures a load job.  Structure is documented below.
	Load JobLoadPtrInput
	// The geographic location of the job. The default value is US.
	Location pulumi.StringPtrInput
	// The ID of the project in which the resource belongs.
	// If it is not provided, the provider project is used.
	Project pulumi.StringPtrInput
	// Configures a query job.  Structure is documented below.
	Query JobQueryPtrInput
	// Email address of the user who ran the job.
	UserEmail pulumi.StringPtrInput
}

func (JobState) ElementType added in v3.3.0

func (JobState) ElementType() reflect.Type

type Reservation

type Reservation struct {
	pulumi.CustomResourceState

	// If false, any query using this reservation will use idle slots from other reservations within
	// the same admin project. If true, a query using this reservation will execute with the slot
	// capacity specified above at most.
	IgnoreIdleSlots pulumi.BoolPtrOutput `pulumi:"ignoreIdleSlots"`
	// The geographic location where the transfer config should reside.
	// Examples: US, EU, asia-northeast1. The default value is US.
	Location pulumi.StringPtrOutput `pulumi:"location"`
	// The name of the reservation. This field must only contain alphanumeric characters or dash.
	Name pulumi.StringOutput `pulumi:"name"`
	// The ID of the project in which the resource belongs.
	// If it is not provided, the provider project is used.
	Project pulumi.StringOutput `pulumi:"project"`
	// Minimum slots available to this reservation. A slot is a unit of computational power in BigQuery, and serves as the
	// unit of parallelism. Queries using this reservation might use more slots during runtime if ignoreIdleSlots is set to false.
	SlotCapacity pulumi.IntOutput `pulumi:"slotCapacity"`
}

A reservation is a mechanism used to guarantee BigQuery slots to users.

To get more information about Reservation, see:

* [API documentation](https://cloud.google.com/bigquery/docs/reference/reservations/rest/v1beta1/projects.locations.reservations/create) * How-to Guides

func GetReservation

func GetReservation(ctx *pulumi.Context,
	name string, id pulumi.IDInput, state *ReservationState, opts ...pulumi.ResourceOption) (*Reservation, error)

GetReservation gets an existing Reservation resource's state with the given name, ID, and optional state properties that are used to uniquely qualify the lookup (nil if not required).

func NewReservation

func NewReservation(ctx *pulumi.Context,
	name string, args *ReservationArgs, opts ...pulumi.ResourceOption) (*Reservation, error)

NewReservation registers a new resource with the given unique name, arguments, and options.

type ReservationArgs

type ReservationArgs struct {
	// If false, any query using this reservation will use idle slots from other reservations within
	// the same admin project. If true, a query using this reservation will execute with the slot
	// capacity specified above at most.
	IgnoreIdleSlots pulumi.BoolPtrInput
	// The geographic location where the transfer config should reside.
	// Examples: US, EU, asia-northeast1. The default value is US.
	Location pulumi.StringPtrInput
	// The name of the reservation. This field must only contain alphanumeric characters or dash.
	Name pulumi.StringPtrInput
	// The ID of the project in which the resource belongs.
	// If it is not provided, the provider project is used.
	Project pulumi.StringPtrInput
	// Minimum slots available to this reservation. A slot is a unit of computational power in BigQuery, and serves as the
	// unit of parallelism. Queries using this reservation might use more slots during runtime if ignoreIdleSlots is set to false.
	SlotCapacity pulumi.IntInput
}

The set of arguments for constructing a Reservation resource.

func (ReservationArgs) ElementType

func (ReservationArgs) ElementType() reflect.Type

type ReservationState

type ReservationState struct {
	// If false, any query using this reservation will use idle slots from other reservations within
	// the same admin project. If true, a query using this reservation will execute with the slot
	// capacity specified above at most.
	IgnoreIdleSlots pulumi.BoolPtrInput
	// The geographic location where the transfer config should reside.
	// Examples: US, EU, asia-northeast1. The default value is US.
	Location pulumi.StringPtrInput
	// The name of the reservation. This field must only contain alphanumeric characters or dash.
	Name pulumi.StringPtrInput
	// The ID of the project in which the resource belongs.
	// If it is not provided, the provider project is used.
	Project pulumi.StringPtrInput
	// Minimum slots available to this reservation. A slot is a unit of computational power in BigQuery, and serves as the
	// unit of parallelism. Queries using this reservation might use more slots during runtime if ignoreIdleSlots is set to false.
	SlotCapacity pulumi.IntPtrInput
}

func (ReservationState) ElementType

func (ReservationState) ElementType() reflect.Type

type Table

type Table struct {
	pulumi.CustomResourceState

	// Specifies column names to use for data clustering.
	// Up to four top-level columns are allowed, and should be specified in
	// descending priority order.
	Clusterings pulumi.StringArrayOutput `pulumi:"clusterings"`
	// The time when this table was created, in milliseconds since the epoch.
	CreationTime pulumi.IntOutput `pulumi:"creationTime"`
	// The dataset ID to create the table in.
	// Changing this forces a new resource to be created.
	DatasetId pulumi.StringOutput `pulumi:"datasetId"`
	// The field description.
	Description pulumi.StringPtrOutput `pulumi:"description"`
	// Specifies how the table should be encrypted.
	// If left blank, the table will be encrypted with a Google-managed key; that process
	// is transparent to the user.  Structure is documented below.
	EncryptionConfiguration TableEncryptionConfigurationPtrOutput `pulumi:"encryptionConfiguration"`
	// A hash of the resource.
	Etag pulumi.StringOutput `pulumi:"etag"`
	// The time when this table expires, in
	// milliseconds since the epoch. If not present, the table will persist
	// indefinitely. Expired tables will be deleted and their storage
	// reclaimed.
	ExpirationTime pulumi.IntOutput `pulumi:"expirationTime"`
	// Describes the data format,
	// location, and other properties of a table stored outside of BigQuery.
	// By defining these properties, the data source can then be queried as
	// if it were a standard BigQuery table. Structure is documented below.
	ExternalDataConfiguration TableExternalDataConfigurationPtrOutput `pulumi:"externalDataConfiguration"`
	// A descriptive name for the table.
	FriendlyName pulumi.StringPtrOutput `pulumi:"friendlyName"`
	// A mapping of labels to assign to the resource.
	Labels pulumi.StringMapOutput `pulumi:"labels"`
	// The time when this table was last modified, in milliseconds since the epoch.
	LastModifiedTime pulumi.IntOutput `pulumi:"lastModifiedTime"`
	// The geographic location where the table resides. This value is inherited from the dataset.
	Location pulumi.StringOutput `pulumi:"location"`
	// The size of this table in bytes, excluding any data in the streaming buffer.
	NumBytes pulumi.IntOutput `pulumi:"numBytes"`
	// The number of bytes in the table that are considered "long-term storage".
	NumLongTermBytes pulumi.IntOutput `pulumi:"numLongTermBytes"`
	// The number of rows of data in this table, excluding any data in the streaming buffer.
	NumRows pulumi.IntOutput `pulumi:"numRows"`
	// The ID of the project in which the resource belongs. If it
	// is not provided, the provider project is used.
	Project pulumi.StringOutput `pulumi:"project"`
	// If specified, configures range-based
	// partitioning for this table. Structure is documented below.
	RangePartitioning TableRangePartitioningPtrOutput `pulumi:"rangePartitioning"`
	// A JSON schema for the table. Schema is required
	// for CSV and JSON formats and is disallowed for Google Cloud
	// Bigtable, Cloud Datastore backups, and Avro formats when using
	// external tables. For more information see the
	// [BigQuery API documentation](https://cloud.google.com/bigquery/docs/reference/rest/v2/tables#resource).
	// ~>**NOTE**: Because this field expects a JSON string, any changes to the
	// string will create a diff, even if the JSON itself hasn't changed.
	// If the API returns a different value for the same schema, e.g. it
	// switched the order of values or replaced `STRUCT` field type with `RECORD`
	// field type, we currently cannot suppress the recurring diff this causes.
	// As a workaround, we recommend using the schema as returned by the API.
	Schema pulumi.StringOutput `pulumi:"schema"`
	// The URI of the created resource.
	SelfLink pulumi.StringOutput `pulumi:"selfLink"`
	// A unique ID for the resource.
	// Changing this forces a new resource to be created.
	TableId pulumi.StringOutput `pulumi:"tableId"`
	// If specified, configures time-based
	// partitioning for this table. Structure is documented below.
	TimePartitioning TableTimePartitioningPtrOutput `pulumi:"timePartitioning"`
	// The only type supported is DAY, which will generate
	// one partition per day based on data loading time.
	Type pulumi.StringOutput `pulumi:"type"`
	// If specified, configures this table as a view.
	// Structure is documented below.
	View TableViewPtrOutput `pulumi:"view"`
}

Creates a table resource in a dataset for Google BigQuery. For more information see [the official documentation](https://cloud.google.com/bigquery/docs/) and [API](https://cloud.google.com/bigquery/docs/reference/rest/v2/tables).

func GetTable

func GetTable(ctx *pulumi.Context,
	name string, id pulumi.IDInput, state *TableState, opts ...pulumi.ResourceOption) (*Table, error)

GetTable gets an existing Table resource's state with the given name, ID, and optional state properties that are used to uniquely qualify the lookup (nil if not required).

func NewTable

func NewTable(ctx *pulumi.Context,
	name string, args *TableArgs, opts ...pulumi.ResourceOption) (*Table, error)

NewTable registers a new resource with the given unique name, arguments, and options.

type TableArgs

type TableArgs struct {
	// Specifies column names to use for data clustering.
	// Up to four top-level columns are allowed, and should be specified in
	// descending priority order.
	Clusterings pulumi.StringArrayInput
	// The dataset ID to create the table in.
	// Changing this forces a new resource to be created.
	DatasetId pulumi.StringInput
	// The field description.
	Description pulumi.StringPtrInput
	// Specifies how the table should be encrypted.
	// If left blank, the table will be encrypted with a Google-managed key; that process
	// is transparent to the user.  Structure is documented below.
	EncryptionConfiguration TableEncryptionConfigurationPtrInput
	// The time when this table expires, in
	// milliseconds since the epoch. If not present, the table will persist
	// indefinitely. Expired tables will be deleted and their storage
	// reclaimed.
	ExpirationTime pulumi.IntPtrInput
	// Describes the data format,
	// location, and other properties of a table stored outside of BigQuery.
	// By defining these properties, the data source can then be queried as
	// if it were a standard BigQuery table. Structure is documented below.
	ExternalDataConfiguration TableExternalDataConfigurationPtrInput
	// A descriptive name for the table.
	FriendlyName pulumi.StringPtrInput
	// A mapping of labels to assign to the resource.
	Labels pulumi.StringMapInput
	// The ID of the project in which the resource belongs. If it
	// is not provided, the provider project is used.
	Project pulumi.StringPtrInput
	// If specified, configures range-based
	// partitioning for this table. Structure is documented below.
	RangePartitioning TableRangePartitioningPtrInput
	// A JSON schema for the table. Schema is required
	// for CSV and JSON formats and is disallowed for Google Cloud
	// Bigtable, Cloud Datastore backups, and Avro formats when using
	// external tables. For more information see the
	// [BigQuery API documentation](https://cloud.google.com/bigquery/docs/reference/rest/v2/tables#resource).
	// ~>**NOTE**: Because this field expects a JSON string, any changes to the
	// string will create a diff, even if the JSON itself hasn't changed.
	// If the API returns a different value for the same schema, e.g. it
	// switched the order of values or replaced `STRUCT` field type with `RECORD`
	// field type, we currently cannot suppress the recurring diff this causes.
	// As a workaround, we recommend using the schema as returned by the API.
	Schema pulumi.StringPtrInput
	// A unique ID for the resource.
	// Changing this forces a new resource to be created.
	TableId pulumi.StringInput
	// If specified, configures time-based
	// partitioning for this table. Structure is documented below.
	TimePartitioning TableTimePartitioningPtrInput
	// If specified, configures this table as a view.
	// Structure is documented below.
	View TableViewPtrInput
}

The set of arguments for constructing a Table resource.

func (TableArgs) ElementType

func (TableArgs) ElementType() reflect.Type

type TableEncryptionConfiguration

type TableEncryptionConfiguration struct {
	// The self link or full name of a key which should be used to
	// encrypt this table.  Note that the default bigquery service account will need to have
	// encrypt/decrypt permissions on this key - you may want to see the
	// `bigquery.getDefaultServiceAccount` datasource and the
	// `kms.CryptoKeyIAMBinding` resource.
	KmsKeyName string `pulumi:"kmsKeyName"`
}

type TableEncryptionConfigurationArgs

type TableEncryptionConfigurationArgs struct {
	// The self link or full name of a key which should be used to
	// encrypt this table.  Note that the default bigquery service account will need to have
	// encrypt/decrypt permissions on this key - you may want to see the
	// `bigquery.getDefaultServiceAccount` datasource and the
	// `kms.CryptoKeyIAMBinding` resource.
	KmsKeyName pulumi.StringInput `pulumi:"kmsKeyName"`
}

func (TableEncryptionConfigurationArgs) ElementType

func (TableEncryptionConfigurationArgs) ToTableEncryptionConfigurationOutput

func (i TableEncryptionConfigurationArgs) ToTableEncryptionConfigurationOutput() TableEncryptionConfigurationOutput

func (TableEncryptionConfigurationArgs) ToTableEncryptionConfigurationOutputWithContext

func (i TableEncryptionConfigurationArgs) ToTableEncryptionConfigurationOutputWithContext(ctx context.Context) TableEncryptionConfigurationOutput

func (TableEncryptionConfigurationArgs) ToTableEncryptionConfigurationPtrOutput

func (i TableEncryptionConfigurationArgs) ToTableEncryptionConfigurationPtrOutput() TableEncryptionConfigurationPtrOutput

func (TableEncryptionConfigurationArgs) ToTableEncryptionConfigurationPtrOutputWithContext

func (i TableEncryptionConfigurationArgs) ToTableEncryptionConfigurationPtrOutputWithContext(ctx context.Context) TableEncryptionConfigurationPtrOutput

type TableEncryptionConfigurationInput

type TableEncryptionConfigurationInput interface {
	pulumi.Input

	ToTableEncryptionConfigurationOutput() TableEncryptionConfigurationOutput
	ToTableEncryptionConfigurationOutputWithContext(context.Context) TableEncryptionConfigurationOutput
}

TableEncryptionConfigurationInput is an input type that accepts TableEncryptionConfigurationArgs and TableEncryptionConfigurationOutput values. You can construct a concrete instance of `TableEncryptionConfigurationInput` via:

TableEncryptionConfigurationArgs{...}

type TableEncryptionConfigurationOutput

type TableEncryptionConfigurationOutput struct{ *pulumi.OutputState }

func (TableEncryptionConfigurationOutput) ElementType

func (TableEncryptionConfigurationOutput) KmsKeyName

The self link or full name of a key which should be used to encrypt this table. Note that the default bigquery service account will need to have encrypt/decrypt permissions on this key - you may want to see the `bigquery.getDefaultServiceAccount` datasource and the `kms.CryptoKeyIAMBinding` resource.

func (TableEncryptionConfigurationOutput) ToTableEncryptionConfigurationOutput

func (o TableEncryptionConfigurationOutput) ToTableEncryptionConfigurationOutput() TableEncryptionConfigurationOutput

func (TableEncryptionConfigurationOutput) ToTableEncryptionConfigurationOutputWithContext

func (o TableEncryptionConfigurationOutput) ToTableEncryptionConfigurationOutputWithContext(ctx context.Context) TableEncryptionConfigurationOutput

func (TableEncryptionConfigurationOutput) ToTableEncryptionConfigurationPtrOutput

func (o TableEncryptionConfigurationOutput) ToTableEncryptionConfigurationPtrOutput() TableEncryptionConfigurationPtrOutput

func (TableEncryptionConfigurationOutput) ToTableEncryptionConfigurationPtrOutputWithContext

func (o TableEncryptionConfigurationOutput) ToTableEncryptionConfigurationPtrOutputWithContext(ctx context.Context) TableEncryptionConfigurationPtrOutput

type TableEncryptionConfigurationPtrInput

type TableEncryptionConfigurationPtrInput interface {
	pulumi.Input

	ToTableEncryptionConfigurationPtrOutput() TableEncryptionConfigurationPtrOutput
	ToTableEncryptionConfigurationPtrOutputWithContext(context.Context) TableEncryptionConfigurationPtrOutput
}

TableEncryptionConfigurationPtrInput is an input type that accepts TableEncryptionConfigurationArgs, TableEncryptionConfigurationPtr and TableEncryptionConfigurationPtrOutput values. You can construct a concrete instance of `TableEncryptionConfigurationPtrInput` via:

		 TableEncryptionConfigurationArgs{...}

 or:

		 nil

type TableEncryptionConfigurationPtrOutput

type TableEncryptionConfigurationPtrOutput struct{ *pulumi.OutputState }

func (TableEncryptionConfigurationPtrOutput) Elem

func (TableEncryptionConfigurationPtrOutput) ElementType

func (TableEncryptionConfigurationPtrOutput) KmsKeyName

The self link or full name of a key which should be used to encrypt this table. Note that the default bigquery service account will need to have encrypt/decrypt permissions on this key - you may want to see the `bigquery.getDefaultServiceAccount` datasource and the `kms.CryptoKeyIAMBinding` resource.

func (TableEncryptionConfigurationPtrOutput) ToTableEncryptionConfigurationPtrOutput

func (o TableEncryptionConfigurationPtrOutput) ToTableEncryptionConfigurationPtrOutput() TableEncryptionConfigurationPtrOutput

func (TableEncryptionConfigurationPtrOutput) ToTableEncryptionConfigurationPtrOutputWithContext

func (o TableEncryptionConfigurationPtrOutput) ToTableEncryptionConfigurationPtrOutputWithContext(ctx context.Context) TableEncryptionConfigurationPtrOutput

type TableExternalDataConfiguration

type TableExternalDataConfiguration struct {
	// - Let BigQuery try to autodetect the schema
	// and format of the table.
	Autodetect bool `pulumi:"autodetect"`
	// The compression type of the data source.
	// Valid values are "NONE" or "GZIP".
	Compression *string `pulumi:"compression"`
	// Additional properties to set if
	// `sourceFormat` is set to "CSV". Structure is documented below.
	CsvOptions *TableExternalDataConfigurationCsvOptions `pulumi:"csvOptions"`
	// Additional options if
	// `sourceFormat` is set to "GOOGLE_SHEETS". Structure is
	// documented below.
	GoogleSheetsOptions *TableExternalDataConfigurationGoogleSheetsOptions `pulumi:"googleSheetsOptions"`
	// Indicates if BigQuery should
	// allow extra values that are not represented in the table schema.
	// If true, the extra values are ignored. If false, records with
	// extra columns are treated as bad records, and if there are too
	// many bad records, an invalid error is returned in the job result.
	// The default value is false.
	IgnoreUnknownValues *bool `pulumi:"ignoreUnknownValues"`
	// The maximum number of bad records that
	// BigQuery can ignore when reading data.
	MaxBadRecords *int `pulumi:"maxBadRecords"`
	// The data format. Supported values are:
	// "CSV", "GOOGLE_SHEETS", "NEWLINE_DELIMITED_JSON", "AVRO", "PARQUET",
	// and "DATSTORE_BACKUP". To use "GOOGLE_SHEETS"
	// the `scopes` must include
	// "https://www.googleapis.com/auth/drive.readonly".
	SourceFormat string `pulumi:"sourceFormat"`
	// A list of the fully-qualified URIs that point to
	// your data in Google Cloud.
	SourceUris []string `pulumi:"sourceUris"`
}

type TableExternalDataConfigurationArgs

type TableExternalDataConfigurationArgs struct {
	// - Let BigQuery try to autodetect the schema
	// and format of the table.
	Autodetect pulumi.BoolInput `pulumi:"autodetect"`
	// The compression type of the data source.
	// Valid values are "NONE" or "GZIP".
	Compression pulumi.StringPtrInput `pulumi:"compression"`
	// Additional properties to set if
	// `sourceFormat` is set to "CSV". Structure is documented below.
	CsvOptions TableExternalDataConfigurationCsvOptionsPtrInput `pulumi:"csvOptions"`
	// Additional options if
	// `sourceFormat` is set to "GOOGLE_SHEETS". Structure is
	// documented below.
	GoogleSheetsOptions TableExternalDataConfigurationGoogleSheetsOptionsPtrInput `pulumi:"googleSheetsOptions"`
	// Indicates if BigQuery should
	// allow extra values that are not represented in the table schema.
	// If true, the extra values are ignored. If false, records with
	// extra columns are treated as bad records, and if there are too
	// many bad records, an invalid error is returned in the job result.
	// The default value is false.
	IgnoreUnknownValues pulumi.BoolPtrInput `pulumi:"ignoreUnknownValues"`
	// The maximum number of bad records that
	// BigQuery can ignore when reading data.
	MaxBadRecords pulumi.IntPtrInput `pulumi:"maxBadRecords"`
	// The data format. Supported values are:
	// "CSV", "GOOGLE_SHEETS", "NEWLINE_DELIMITED_JSON", "AVRO", "PARQUET",
	// and "DATSTORE_BACKUP". To use "GOOGLE_SHEETS"
	// the `scopes` must include
	// "https://www.googleapis.com/auth/drive.readonly".
	SourceFormat pulumi.StringInput `pulumi:"sourceFormat"`
	// A list of the fully-qualified URIs that point to
	// your data in Google Cloud.
	SourceUris pulumi.StringArrayInput `pulumi:"sourceUris"`
}

func (TableExternalDataConfigurationArgs) ElementType

func (TableExternalDataConfigurationArgs) ToTableExternalDataConfigurationOutput

func (i TableExternalDataConfigurationArgs) ToTableExternalDataConfigurationOutput() TableExternalDataConfigurationOutput

func (TableExternalDataConfigurationArgs) ToTableExternalDataConfigurationOutputWithContext

func (i TableExternalDataConfigurationArgs) ToTableExternalDataConfigurationOutputWithContext(ctx context.Context) TableExternalDataConfigurationOutput

func (TableExternalDataConfigurationArgs) ToTableExternalDataConfigurationPtrOutput

func (i TableExternalDataConfigurationArgs) ToTableExternalDataConfigurationPtrOutput() TableExternalDataConfigurationPtrOutput

func (TableExternalDataConfigurationArgs) ToTableExternalDataConfigurationPtrOutputWithContext

func (i TableExternalDataConfigurationArgs) ToTableExternalDataConfigurationPtrOutputWithContext(ctx context.Context) TableExternalDataConfigurationPtrOutput

type TableExternalDataConfigurationCsvOptions

type TableExternalDataConfigurationCsvOptions struct {
	// Indicates if BigQuery should accept rows
	// that are missing trailing optional columns.
	AllowJaggedRows *bool `pulumi:"allowJaggedRows"`
	// Indicates if BigQuery should allow
	// quoted data sections that contain newline characters in a CSV file.
	// The default value is false.
	AllowQuotedNewlines *bool `pulumi:"allowQuotedNewlines"`
	// The character encoding of the data. The supported
	// values are UTF-8 or ISO-8859-1.
	Encoding *string `pulumi:"encoding"`
	// The separator for fields in a CSV file.
	FieldDelimiter *string `pulumi:"fieldDelimiter"`
	// The value that is used to quote data sections in a
	// CSV file. If your data does not contain quoted sections, set the
	// property value to an empty string. If your data contains quoted newline
	// characters, you must also set the `allowQuotedNewlines` property to true.
	// The API-side default is `"`, specified in the provider escaped as `\"`. Due to
	// limitations with default values, this value is required to be
	// explicitly set.
	Quote string `pulumi:"quote"`
	// The number of rows at the top of the sheet
	// that BigQuery will skip when reading the data. At least one of `range` or
	// `skipLeadingRows` must be set.
	SkipLeadingRows *int `pulumi:"skipLeadingRows"`
}

type TableExternalDataConfigurationCsvOptionsArgs

type TableExternalDataConfigurationCsvOptionsArgs struct {
	// Indicates if BigQuery should accept rows
	// that are missing trailing optional columns.
	AllowJaggedRows pulumi.BoolPtrInput `pulumi:"allowJaggedRows"`
	// Indicates if BigQuery should allow
	// quoted data sections that contain newline characters in a CSV file.
	// The default value is false.
	AllowQuotedNewlines pulumi.BoolPtrInput `pulumi:"allowQuotedNewlines"`
	// The character encoding of the data. The supported
	// values are UTF-8 or ISO-8859-1.
	Encoding pulumi.StringPtrInput `pulumi:"encoding"`
	// The separator for fields in a CSV file.
	FieldDelimiter pulumi.StringPtrInput `pulumi:"fieldDelimiter"`
	// The value that is used to quote data sections in a
	// CSV file. If your data does not contain quoted sections, set the
	// property value to an empty string. If your data contains quoted newline
	// characters, you must also set the `allowQuotedNewlines` property to true.
	// The API-side default is `"`, specified in the provider escaped as `\"`. Due to
	// limitations with default values, this value is required to be
	// explicitly set.
	Quote pulumi.StringInput `pulumi:"quote"`
	// The number of rows at the top of the sheet
	// that BigQuery will skip when reading the data. At least one of `range` or
	// `skipLeadingRows` must be set.
	SkipLeadingRows pulumi.IntPtrInput `pulumi:"skipLeadingRows"`
}

func (TableExternalDataConfigurationCsvOptionsArgs) ElementType

func (TableExternalDataConfigurationCsvOptionsArgs) ToTableExternalDataConfigurationCsvOptionsOutput

func (i TableExternalDataConfigurationCsvOptionsArgs) ToTableExternalDataConfigurationCsvOptionsOutput() TableExternalDataConfigurationCsvOptionsOutput

func (TableExternalDataConfigurationCsvOptionsArgs) ToTableExternalDataConfigurationCsvOptionsOutputWithContext

func (i TableExternalDataConfigurationCsvOptionsArgs) ToTableExternalDataConfigurationCsvOptionsOutputWithContext(ctx context.Context) TableExternalDataConfigurationCsvOptionsOutput

func (TableExternalDataConfigurationCsvOptionsArgs) ToTableExternalDataConfigurationCsvOptionsPtrOutput

func (i TableExternalDataConfigurationCsvOptionsArgs) ToTableExternalDataConfigurationCsvOptionsPtrOutput() TableExternalDataConfigurationCsvOptionsPtrOutput

func (TableExternalDataConfigurationCsvOptionsArgs) ToTableExternalDataConfigurationCsvOptionsPtrOutputWithContext

func (i TableExternalDataConfigurationCsvOptionsArgs) ToTableExternalDataConfigurationCsvOptionsPtrOutputWithContext(ctx context.Context) TableExternalDataConfigurationCsvOptionsPtrOutput

type TableExternalDataConfigurationCsvOptionsInput

type TableExternalDataConfigurationCsvOptionsInput interface {
	pulumi.Input

	ToTableExternalDataConfigurationCsvOptionsOutput() TableExternalDataConfigurationCsvOptionsOutput
	ToTableExternalDataConfigurationCsvOptionsOutputWithContext(context.Context) TableExternalDataConfigurationCsvOptionsOutput
}

TableExternalDataConfigurationCsvOptionsInput is an input type that accepts TableExternalDataConfigurationCsvOptionsArgs and TableExternalDataConfigurationCsvOptionsOutput values. You can construct a concrete instance of `TableExternalDataConfigurationCsvOptionsInput` via:

TableExternalDataConfigurationCsvOptionsArgs{...}

type TableExternalDataConfigurationCsvOptionsOutput

type TableExternalDataConfigurationCsvOptionsOutput struct{ *pulumi.OutputState }

func (TableExternalDataConfigurationCsvOptionsOutput) AllowJaggedRows

Indicates if BigQuery should accept rows that are missing trailing optional columns.

func (TableExternalDataConfigurationCsvOptionsOutput) AllowQuotedNewlines

Indicates if BigQuery should allow quoted data sections that contain newline characters in a CSV file. The default value is false.

func (TableExternalDataConfigurationCsvOptionsOutput) ElementType

func (TableExternalDataConfigurationCsvOptionsOutput) Encoding

The character encoding of the data. The supported values are UTF-8 or ISO-8859-1.

func (TableExternalDataConfigurationCsvOptionsOutput) FieldDelimiter

The separator for fields in a CSV file.

func (TableExternalDataConfigurationCsvOptionsOutput) Quote

The value that is used to quote data sections in a CSV file. If your data does not contain quoted sections, set the property value to an empty string. If your data contains quoted newline characters, you must also set the `allowQuotedNewlines` property to true. The API-side default is `"`, specified in the provider escaped as `\"`. Due to limitations with default values, this value is required to be explicitly set.

func (TableExternalDataConfigurationCsvOptionsOutput) SkipLeadingRows

The number of rows at the top of the sheet that BigQuery will skip when reading the data. At least one of `range` or `skipLeadingRows` must be set.

func (TableExternalDataConfigurationCsvOptionsOutput) ToTableExternalDataConfigurationCsvOptionsOutput

func (o TableExternalDataConfigurationCsvOptionsOutput) ToTableExternalDataConfigurationCsvOptionsOutput() TableExternalDataConfigurationCsvOptionsOutput

func (TableExternalDataConfigurationCsvOptionsOutput) ToTableExternalDataConfigurationCsvOptionsOutputWithContext

func (o TableExternalDataConfigurationCsvOptionsOutput) ToTableExternalDataConfigurationCsvOptionsOutputWithContext(ctx context.Context) TableExternalDataConfigurationCsvOptionsOutput

func (TableExternalDataConfigurationCsvOptionsOutput) ToTableExternalDataConfigurationCsvOptionsPtrOutput

func (o TableExternalDataConfigurationCsvOptionsOutput) ToTableExternalDataConfigurationCsvOptionsPtrOutput() TableExternalDataConfigurationCsvOptionsPtrOutput

func (TableExternalDataConfigurationCsvOptionsOutput) ToTableExternalDataConfigurationCsvOptionsPtrOutputWithContext

func (o TableExternalDataConfigurationCsvOptionsOutput) ToTableExternalDataConfigurationCsvOptionsPtrOutputWithContext(ctx context.Context) TableExternalDataConfigurationCsvOptionsPtrOutput

type TableExternalDataConfigurationCsvOptionsPtrInput

type TableExternalDataConfigurationCsvOptionsPtrInput interface {
	pulumi.Input

	ToTableExternalDataConfigurationCsvOptionsPtrOutput() TableExternalDataConfigurationCsvOptionsPtrOutput
	ToTableExternalDataConfigurationCsvOptionsPtrOutputWithContext(context.Context) TableExternalDataConfigurationCsvOptionsPtrOutput
}

TableExternalDataConfigurationCsvOptionsPtrInput is an input type that accepts TableExternalDataConfigurationCsvOptionsArgs, TableExternalDataConfigurationCsvOptionsPtr and TableExternalDataConfigurationCsvOptionsPtrOutput values. You can construct a concrete instance of `TableExternalDataConfigurationCsvOptionsPtrInput` via:

		 TableExternalDataConfigurationCsvOptionsArgs{...}

 or:

		 nil

type TableExternalDataConfigurationCsvOptionsPtrOutput

type TableExternalDataConfigurationCsvOptionsPtrOutput struct{ *pulumi.OutputState }

func (TableExternalDataConfigurationCsvOptionsPtrOutput) AllowJaggedRows

Indicates if BigQuery should accept rows that are missing trailing optional columns.

func (TableExternalDataConfigurationCsvOptionsPtrOutput) AllowQuotedNewlines

Indicates if BigQuery should allow quoted data sections that contain newline characters in a CSV file. The default value is false.

func (TableExternalDataConfigurationCsvOptionsPtrOutput) Elem

func (TableExternalDataConfigurationCsvOptionsPtrOutput) ElementType

func (TableExternalDataConfigurationCsvOptionsPtrOutput) Encoding

The character encoding of the data. The supported values are UTF-8 or ISO-8859-1.

func (TableExternalDataConfigurationCsvOptionsPtrOutput) FieldDelimiter

The separator for fields in a CSV file.

func (TableExternalDataConfigurationCsvOptionsPtrOutput) Quote

The value that is used to quote data sections in a CSV file. If your data does not contain quoted sections, set the property value to an empty string. If your data contains quoted newline characters, you must also set the `allowQuotedNewlines` property to true. The API-side default is `"`, specified in the provider escaped as `\"`. Due to limitations with default values, this value is required to be explicitly set.

func (TableExternalDataConfigurationCsvOptionsPtrOutput) SkipLeadingRows

The number of rows at the top of the sheet that BigQuery will skip when reading the data. At least one of `range` or `skipLeadingRows` must be set.

func (TableExternalDataConfigurationCsvOptionsPtrOutput) ToTableExternalDataConfigurationCsvOptionsPtrOutput

func (o TableExternalDataConfigurationCsvOptionsPtrOutput) ToTableExternalDataConfigurationCsvOptionsPtrOutput() TableExternalDataConfigurationCsvOptionsPtrOutput

func (TableExternalDataConfigurationCsvOptionsPtrOutput) ToTableExternalDataConfigurationCsvOptionsPtrOutputWithContext

func (o TableExternalDataConfigurationCsvOptionsPtrOutput) ToTableExternalDataConfigurationCsvOptionsPtrOutputWithContext(ctx context.Context) TableExternalDataConfigurationCsvOptionsPtrOutput

type TableExternalDataConfigurationGoogleSheetsOptions

type TableExternalDataConfigurationGoogleSheetsOptions struct {
	// Information required to partition based on ranges.
	// Structure is documented below.
	Range *string `pulumi:"range"`
	// The number of rows at the top of the sheet
	// that BigQuery will skip when reading the data. At least one of `range` or
	// `skipLeadingRows` must be set.
	SkipLeadingRows *int `pulumi:"skipLeadingRows"`
}

type TableExternalDataConfigurationGoogleSheetsOptionsArgs

type TableExternalDataConfigurationGoogleSheetsOptionsArgs struct {
	// Information required to partition based on ranges.
	// Structure is documented below.
	Range pulumi.StringPtrInput `pulumi:"range"`
	// The number of rows at the top of the sheet
	// that BigQuery will skip when reading the data. At least one of `range` or
	// `skipLeadingRows` must be set.
	SkipLeadingRows pulumi.IntPtrInput `pulumi:"skipLeadingRows"`
}

func (TableExternalDataConfigurationGoogleSheetsOptionsArgs) ElementType

func (TableExternalDataConfigurationGoogleSheetsOptionsArgs) ToTableExternalDataConfigurationGoogleSheetsOptionsOutput

func (i TableExternalDataConfigurationGoogleSheetsOptionsArgs) ToTableExternalDataConfigurationGoogleSheetsOptionsOutput() TableExternalDataConfigurationGoogleSheetsOptionsOutput

func (TableExternalDataConfigurationGoogleSheetsOptionsArgs) ToTableExternalDataConfigurationGoogleSheetsOptionsOutputWithContext

func (i TableExternalDataConfigurationGoogleSheetsOptionsArgs) ToTableExternalDataConfigurationGoogleSheetsOptionsOutputWithContext(ctx context.Context) TableExternalDataConfigurationGoogleSheetsOptionsOutput

func (TableExternalDataConfigurationGoogleSheetsOptionsArgs) ToTableExternalDataConfigurationGoogleSheetsOptionsPtrOutput

func (i TableExternalDataConfigurationGoogleSheetsOptionsArgs) ToTableExternalDataConfigurationGoogleSheetsOptionsPtrOutput() TableExternalDataConfigurationGoogleSheetsOptionsPtrOutput

func (TableExternalDataConfigurationGoogleSheetsOptionsArgs) ToTableExternalDataConfigurationGoogleSheetsOptionsPtrOutputWithContext

func (i TableExternalDataConfigurationGoogleSheetsOptionsArgs) ToTableExternalDataConfigurationGoogleSheetsOptionsPtrOutputWithContext(ctx context.Context) TableExternalDataConfigurationGoogleSheetsOptionsPtrOutput

type TableExternalDataConfigurationGoogleSheetsOptionsInput

type TableExternalDataConfigurationGoogleSheetsOptionsInput interface {
	pulumi.Input

	ToTableExternalDataConfigurationGoogleSheetsOptionsOutput() TableExternalDataConfigurationGoogleSheetsOptionsOutput
	ToTableExternalDataConfigurationGoogleSheetsOptionsOutputWithContext(context.Context) TableExternalDataConfigurationGoogleSheetsOptionsOutput
}

TableExternalDataConfigurationGoogleSheetsOptionsInput is an input type that accepts TableExternalDataConfigurationGoogleSheetsOptionsArgs and TableExternalDataConfigurationGoogleSheetsOptionsOutput values. You can construct a concrete instance of `TableExternalDataConfigurationGoogleSheetsOptionsInput` via:

TableExternalDataConfigurationGoogleSheetsOptionsArgs{...}

type TableExternalDataConfigurationGoogleSheetsOptionsOutput

type TableExternalDataConfigurationGoogleSheetsOptionsOutput struct{ *pulumi.OutputState }

func (TableExternalDataConfigurationGoogleSheetsOptionsOutput) ElementType

func (TableExternalDataConfigurationGoogleSheetsOptionsOutput) Range

Information required to partition based on ranges. Structure is documented below.

func (TableExternalDataConfigurationGoogleSheetsOptionsOutput) SkipLeadingRows

The number of rows at the top of the sheet that BigQuery will skip when reading the data. At least one of `range` or `skipLeadingRows` must be set.

func (TableExternalDataConfigurationGoogleSheetsOptionsOutput) ToTableExternalDataConfigurationGoogleSheetsOptionsOutput

func (TableExternalDataConfigurationGoogleSheetsOptionsOutput) ToTableExternalDataConfigurationGoogleSheetsOptionsOutputWithContext

func (o TableExternalDataConfigurationGoogleSheetsOptionsOutput) ToTableExternalDataConfigurationGoogleSheetsOptionsOutputWithContext(ctx context.Context) TableExternalDataConfigurationGoogleSheetsOptionsOutput

func (TableExternalDataConfigurationGoogleSheetsOptionsOutput) ToTableExternalDataConfigurationGoogleSheetsOptionsPtrOutput

func (TableExternalDataConfigurationGoogleSheetsOptionsOutput) ToTableExternalDataConfigurationGoogleSheetsOptionsPtrOutputWithContext

func (o TableExternalDataConfigurationGoogleSheetsOptionsOutput) ToTableExternalDataConfigurationGoogleSheetsOptionsPtrOutputWithContext(ctx context.Context) TableExternalDataConfigurationGoogleSheetsOptionsPtrOutput

type TableExternalDataConfigurationGoogleSheetsOptionsPtrInput

type TableExternalDataConfigurationGoogleSheetsOptionsPtrInput interface {
	pulumi.Input

	ToTableExternalDataConfigurationGoogleSheetsOptionsPtrOutput() TableExternalDataConfigurationGoogleSheetsOptionsPtrOutput
	ToTableExternalDataConfigurationGoogleSheetsOptionsPtrOutputWithContext(context.Context) TableExternalDataConfigurationGoogleSheetsOptionsPtrOutput
}

TableExternalDataConfigurationGoogleSheetsOptionsPtrInput is an input type that accepts TableExternalDataConfigurationGoogleSheetsOptionsArgs, TableExternalDataConfigurationGoogleSheetsOptionsPtr and TableExternalDataConfigurationGoogleSheetsOptionsPtrOutput values. You can construct a concrete instance of `TableExternalDataConfigurationGoogleSheetsOptionsPtrInput` via:

		 TableExternalDataConfigurationGoogleSheetsOptionsArgs{...}

 or:

		 nil

type TableExternalDataConfigurationGoogleSheetsOptionsPtrOutput

type TableExternalDataConfigurationGoogleSheetsOptionsPtrOutput struct{ *pulumi.OutputState }

func (TableExternalDataConfigurationGoogleSheetsOptionsPtrOutput) Elem

func (TableExternalDataConfigurationGoogleSheetsOptionsPtrOutput) ElementType

func (TableExternalDataConfigurationGoogleSheetsOptionsPtrOutput) Range

Information required to partition based on ranges. Structure is documented below.

func (TableExternalDataConfigurationGoogleSheetsOptionsPtrOutput) SkipLeadingRows

The number of rows at the top of the sheet that BigQuery will skip when reading the data. At least one of `range` or `skipLeadingRows` must be set.

func (TableExternalDataConfigurationGoogleSheetsOptionsPtrOutput) ToTableExternalDataConfigurationGoogleSheetsOptionsPtrOutput

func (TableExternalDataConfigurationGoogleSheetsOptionsPtrOutput) ToTableExternalDataConfigurationGoogleSheetsOptionsPtrOutputWithContext

func (o TableExternalDataConfigurationGoogleSheetsOptionsPtrOutput) ToTableExternalDataConfigurationGoogleSheetsOptionsPtrOutputWithContext(ctx context.Context) TableExternalDataConfigurationGoogleSheetsOptionsPtrOutput

type TableExternalDataConfigurationInput

type TableExternalDataConfigurationInput interface {
	pulumi.Input

	ToTableExternalDataConfigurationOutput() TableExternalDataConfigurationOutput
	ToTableExternalDataConfigurationOutputWithContext(context.Context) TableExternalDataConfigurationOutput
}

TableExternalDataConfigurationInput is an input type that accepts TableExternalDataConfigurationArgs and TableExternalDataConfigurationOutput values. You can construct a concrete instance of `TableExternalDataConfigurationInput` via:

TableExternalDataConfigurationArgs{...}

type TableExternalDataConfigurationOutput

type TableExternalDataConfigurationOutput struct{ *pulumi.OutputState }

func (TableExternalDataConfigurationOutput) Autodetect

- Let BigQuery try to autodetect the schema and format of the table.

func (TableExternalDataConfigurationOutput) Compression

The compression type of the data source. Valid values are "NONE" or "GZIP".

func (TableExternalDataConfigurationOutput) CsvOptions

Additional properties to set if `sourceFormat` is set to "CSV". Structure is documented below.

func (TableExternalDataConfigurationOutput) ElementType

func (TableExternalDataConfigurationOutput) GoogleSheetsOptions

Additional options if `sourceFormat` is set to "GOOGLE_SHEETS". Structure is documented below.

func (TableExternalDataConfigurationOutput) IgnoreUnknownValues

Indicates if BigQuery should allow extra values that are not represented in the table schema. If true, the extra values are ignored. If false, records with extra columns are treated as bad records, and if there are too many bad records, an invalid error is returned in the job result. The default value is false.

func (TableExternalDataConfigurationOutput) MaxBadRecords

The maximum number of bad records that BigQuery can ignore when reading data.

func (TableExternalDataConfigurationOutput) SourceFormat

The data format. Supported values are: "CSV", "GOOGLE_SHEETS", "NEWLINE_DELIMITED_JSON", "AVRO", "PARQUET", and "DATSTORE_BACKUP". To use "GOOGLE_SHEETS" the `scopes` must include "https://www.googleapis.com/auth/drive.readonly".

func (TableExternalDataConfigurationOutput) SourceUris

A list of the fully-qualified URIs that point to your data in Google Cloud.

func (TableExternalDataConfigurationOutput) ToTableExternalDataConfigurationOutput

func (o TableExternalDataConfigurationOutput) ToTableExternalDataConfigurationOutput() TableExternalDataConfigurationOutput

func (TableExternalDataConfigurationOutput) ToTableExternalDataConfigurationOutputWithContext

func (o TableExternalDataConfigurationOutput) ToTableExternalDataConfigurationOutputWithContext(ctx context.Context) TableExternalDataConfigurationOutput

func (TableExternalDataConfigurationOutput) ToTableExternalDataConfigurationPtrOutput

func (o TableExternalDataConfigurationOutput) ToTableExternalDataConfigurationPtrOutput() TableExternalDataConfigurationPtrOutput

func (TableExternalDataConfigurationOutput) ToTableExternalDataConfigurationPtrOutputWithContext

func (o TableExternalDataConfigurationOutput) ToTableExternalDataConfigurationPtrOutputWithContext(ctx context.Context) TableExternalDataConfigurationPtrOutput

type TableExternalDataConfigurationPtrInput

type TableExternalDataConfigurationPtrInput interface {
	pulumi.Input

	ToTableExternalDataConfigurationPtrOutput() TableExternalDataConfigurationPtrOutput
	ToTableExternalDataConfigurationPtrOutputWithContext(context.Context) TableExternalDataConfigurationPtrOutput
}

TableExternalDataConfigurationPtrInput is an input type that accepts TableExternalDataConfigurationArgs, TableExternalDataConfigurationPtr and TableExternalDataConfigurationPtrOutput values. You can construct a concrete instance of `TableExternalDataConfigurationPtrInput` via:

		 TableExternalDataConfigurationArgs{...}

 or:

		 nil

type TableExternalDataConfigurationPtrOutput

type TableExternalDataConfigurationPtrOutput struct{ *pulumi.OutputState }

func (TableExternalDataConfigurationPtrOutput) Autodetect

- Let BigQuery try to autodetect the schema and format of the table.

func (TableExternalDataConfigurationPtrOutput) Compression

The compression type of the data source. Valid values are "NONE" or "GZIP".

func (TableExternalDataConfigurationPtrOutput) CsvOptions

Additional properties to set if `sourceFormat` is set to "CSV". Structure is documented below.

func (TableExternalDataConfigurationPtrOutput) Elem

func (TableExternalDataConfigurationPtrOutput) ElementType

func (TableExternalDataConfigurationPtrOutput) GoogleSheetsOptions

Additional options if `sourceFormat` is set to "GOOGLE_SHEETS". Structure is documented below.

func (TableExternalDataConfigurationPtrOutput) IgnoreUnknownValues

Indicates if BigQuery should allow extra values that are not represented in the table schema. If true, the extra values are ignored. If false, records with extra columns are treated as bad records, and if there are too many bad records, an invalid error is returned in the job result. The default value is false.

func (TableExternalDataConfigurationPtrOutput) MaxBadRecords

The maximum number of bad records that BigQuery can ignore when reading data.

func (TableExternalDataConfigurationPtrOutput) SourceFormat

The data format. Supported values are: "CSV", "GOOGLE_SHEETS", "NEWLINE_DELIMITED_JSON", "AVRO", "PARQUET", and "DATSTORE_BACKUP". To use "GOOGLE_SHEETS" the `scopes` must include "https://www.googleapis.com/auth/drive.readonly".

func (TableExternalDataConfigurationPtrOutput) SourceUris

A list of the fully-qualified URIs that point to your data in Google Cloud.

func (TableExternalDataConfigurationPtrOutput) ToTableExternalDataConfigurationPtrOutput

func (o TableExternalDataConfigurationPtrOutput) ToTableExternalDataConfigurationPtrOutput() TableExternalDataConfigurationPtrOutput

func (TableExternalDataConfigurationPtrOutput) ToTableExternalDataConfigurationPtrOutputWithContext

func (o TableExternalDataConfigurationPtrOutput) ToTableExternalDataConfigurationPtrOutputWithContext(ctx context.Context) TableExternalDataConfigurationPtrOutput

type TableRangePartitioning

type TableRangePartitioning struct {
	// The field used to determine how to create a range-based
	// partition.
	Field string `pulumi:"field"`
	// Information required to partition based on ranges.
	// Structure is documented below.
	Range TableRangePartitioningRange `pulumi:"range"`
}

type TableRangePartitioningArgs

type TableRangePartitioningArgs struct {
	// The field used to determine how to create a range-based
	// partition.
	Field pulumi.StringInput `pulumi:"field"`
	// Information required to partition based on ranges.
	// Structure is documented below.
	Range TableRangePartitioningRangeInput `pulumi:"range"`
}

func (TableRangePartitioningArgs) ElementType

func (TableRangePartitioningArgs) ElementType() reflect.Type

func (TableRangePartitioningArgs) ToTableRangePartitioningOutput

func (i TableRangePartitioningArgs) ToTableRangePartitioningOutput() TableRangePartitioningOutput

func (TableRangePartitioningArgs) ToTableRangePartitioningOutputWithContext

func (i TableRangePartitioningArgs) ToTableRangePartitioningOutputWithContext(ctx context.Context) TableRangePartitioningOutput

func (TableRangePartitioningArgs) ToTableRangePartitioningPtrOutput

func (i TableRangePartitioningArgs) ToTableRangePartitioningPtrOutput() TableRangePartitioningPtrOutput

func (TableRangePartitioningArgs) ToTableRangePartitioningPtrOutputWithContext

func (i TableRangePartitioningArgs) ToTableRangePartitioningPtrOutputWithContext(ctx context.Context) TableRangePartitioningPtrOutput

type TableRangePartitioningInput

type TableRangePartitioningInput interface {
	pulumi.Input

	ToTableRangePartitioningOutput() TableRangePartitioningOutput
	ToTableRangePartitioningOutputWithContext(context.Context) TableRangePartitioningOutput
}

TableRangePartitioningInput is an input type that accepts TableRangePartitioningArgs and TableRangePartitioningOutput values. You can construct a concrete instance of `TableRangePartitioningInput` via:

TableRangePartitioningArgs{...}

type TableRangePartitioningOutput

type TableRangePartitioningOutput struct{ *pulumi.OutputState }

func (TableRangePartitioningOutput) ElementType

func (TableRangePartitioningOutput) Field

The field used to determine how to create a range-based partition.

func (TableRangePartitioningOutput) Range

Information required to partition based on ranges. Structure is documented below.

func (TableRangePartitioningOutput) ToTableRangePartitioningOutput

func (o TableRangePartitioningOutput) ToTableRangePartitioningOutput() TableRangePartitioningOutput

func (TableRangePartitioningOutput) ToTableRangePartitioningOutputWithContext

func (o TableRangePartitioningOutput) ToTableRangePartitioningOutputWithContext(ctx context.Context) TableRangePartitioningOutput

func (TableRangePartitioningOutput) ToTableRangePartitioningPtrOutput

func (o TableRangePartitioningOutput) ToTableRangePartitioningPtrOutput() TableRangePartitioningPtrOutput

func (TableRangePartitioningOutput) ToTableRangePartitioningPtrOutputWithContext

func (o TableRangePartitioningOutput) ToTableRangePartitioningPtrOutputWithContext(ctx context.Context) TableRangePartitioningPtrOutput

type TableRangePartitioningPtrInput

type TableRangePartitioningPtrInput interface {
	pulumi.Input

	ToTableRangePartitioningPtrOutput() TableRangePartitioningPtrOutput
	ToTableRangePartitioningPtrOutputWithContext(context.Context) TableRangePartitioningPtrOutput
}

TableRangePartitioningPtrInput is an input type that accepts TableRangePartitioningArgs, TableRangePartitioningPtr and TableRangePartitioningPtrOutput values. You can construct a concrete instance of `TableRangePartitioningPtrInput` via:

		 TableRangePartitioningArgs{...}

 or:

		 nil

type TableRangePartitioningPtrOutput

type TableRangePartitioningPtrOutput struct{ *pulumi.OutputState }

func (TableRangePartitioningPtrOutput) Elem

func (TableRangePartitioningPtrOutput) ElementType

func (TableRangePartitioningPtrOutput) Field

The field used to determine how to create a range-based partition.

func (TableRangePartitioningPtrOutput) Range

Information required to partition based on ranges. Structure is documented below.

func (TableRangePartitioningPtrOutput) ToTableRangePartitioningPtrOutput

func (o TableRangePartitioningPtrOutput) ToTableRangePartitioningPtrOutput() TableRangePartitioningPtrOutput

func (TableRangePartitioningPtrOutput) ToTableRangePartitioningPtrOutputWithContext

func (o TableRangePartitioningPtrOutput) ToTableRangePartitioningPtrOutputWithContext(ctx context.Context) TableRangePartitioningPtrOutput

type TableRangePartitioningRange

type TableRangePartitioningRange struct {
	// End of the range partitioning, exclusive.
	End int `pulumi:"end"`
	// The width of each range within the partition.
	Interval int `pulumi:"interval"`
	// Start of the range partitioning, inclusive.
	Start int `pulumi:"start"`
}

type TableRangePartitioningRangeArgs

type TableRangePartitioningRangeArgs struct {
	// End of the range partitioning, exclusive.
	End pulumi.IntInput `pulumi:"end"`
	// The width of each range within the partition.
	Interval pulumi.IntInput `pulumi:"interval"`
	// Start of the range partitioning, inclusive.
	Start pulumi.IntInput `pulumi:"start"`
}

func (TableRangePartitioningRangeArgs) ElementType

func (TableRangePartitioningRangeArgs) ToTableRangePartitioningRangeOutput

func (i TableRangePartitioningRangeArgs) ToTableRangePartitioningRangeOutput() TableRangePartitioningRangeOutput

func (TableRangePartitioningRangeArgs) ToTableRangePartitioningRangeOutputWithContext

func (i TableRangePartitioningRangeArgs) ToTableRangePartitioningRangeOutputWithContext(ctx context.Context) TableRangePartitioningRangeOutput

func (TableRangePartitioningRangeArgs) ToTableRangePartitioningRangePtrOutput added in v3.2.0

func (i TableRangePartitioningRangeArgs) ToTableRangePartitioningRangePtrOutput() TableRangePartitioningRangePtrOutput

func (TableRangePartitioningRangeArgs) ToTableRangePartitioningRangePtrOutputWithContext added in v3.2.0

func (i TableRangePartitioningRangeArgs) ToTableRangePartitioningRangePtrOutputWithContext(ctx context.Context) TableRangePartitioningRangePtrOutput

type TableRangePartitioningRangeInput

type TableRangePartitioningRangeInput interface {
	pulumi.Input

	ToTableRangePartitioningRangeOutput() TableRangePartitioningRangeOutput
	ToTableRangePartitioningRangeOutputWithContext(context.Context) TableRangePartitioningRangeOutput
}

TableRangePartitioningRangeInput is an input type that accepts TableRangePartitioningRangeArgs and TableRangePartitioningRangeOutput values. You can construct a concrete instance of `TableRangePartitioningRangeInput` via:

TableRangePartitioningRangeArgs{...}

type TableRangePartitioningRangeOutput

type TableRangePartitioningRangeOutput struct{ *pulumi.OutputState }

func (TableRangePartitioningRangeOutput) ElementType

func (TableRangePartitioningRangeOutput) End

End of the range partitioning, exclusive.

func (TableRangePartitioningRangeOutput) Interval

The width of each range within the partition.

func (TableRangePartitioningRangeOutput) Start

Start of the range partitioning, inclusive.

func (TableRangePartitioningRangeOutput) ToTableRangePartitioningRangeOutput

func (o TableRangePartitioningRangeOutput) ToTableRangePartitioningRangeOutput() TableRangePartitioningRangeOutput

func (TableRangePartitioningRangeOutput) ToTableRangePartitioningRangeOutputWithContext

func (o TableRangePartitioningRangeOutput) ToTableRangePartitioningRangeOutputWithContext(ctx context.Context) TableRangePartitioningRangeOutput

func (TableRangePartitioningRangeOutput) ToTableRangePartitioningRangePtrOutput added in v3.2.0

func (o TableRangePartitioningRangeOutput) ToTableRangePartitioningRangePtrOutput() TableRangePartitioningRangePtrOutput

func (TableRangePartitioningRangeOutput) ToTableRangePartitioningRangePtrOutputWithContext added in v3.2.0

func (o TableRangePartitioningRangeOutput) ToTableRangePartitioningRangePtrOutputWithContext(ctx context.Context) TableRangePartitioningRangePtrOutput

type TableRangePartitioningRangePtrInput added in v3.2.0

type TableRangePartitioningRangePtrInput interface {
	pulumi.Input

	ToTableRangePartitioningRangePtrOutput() TableRangePartitioningRangePtrOutput
	ToTableRangePartitioningRangePtrOutputWithContext(context.Context) TableRangePartitioningRangePtrOutput
}

TableRangePartitioningRangePtrInput is an input type that accepts TableRangePartitioningRangeArgs, TableRangePartitioningRangePtr and TableRangePartitioningRangePtrOutput values. You can construct a concrete instance of `TableRangePartitioningRangePtrInput` via:

		 TableRangePartitioningRangeArgs{...}

 or:

		 nil

func TableRangePartitioningRangePtr added in v3.2.0

type TableRangePartitioningRangePtrOutput added in v3.2.0

type TableRangePartitioningRangePtrOutput struct{ *pulumi.OutputState }

func (TableRangePartitioningRangePtrOutput) Elem added in v3.2.0

func (TableRangePartitioningRangePtrOutput) ElementType added in v3.2.0

func (TableRangePartitioningRangePtrOutput) End added in v3.2.0

End of the range partitioning, exclusive.

func (TableRangePartitioningRangePtrOutput) Interval added in v3.2.0

The width of each range within the partition.

func (TableRangePartitioningRangePtrOutput) Start added in v3.2.0

Start of the range partitioning, inclusive.

func (TableRangePartitioningRangePtrOutput) ToTableRangePartitioningRangePtrOutput added in v3.2.0

func (o TableRangePartitioningRangePtrOutput) ToTableRangePartitioningRangePtrOutput() TableRangePartitioningRangePtrOutput

func (TableRangePartitioningRangePtrOutput) ToTableRangePartitioningRangePtrOutputWithContext added in v3.2.0

func (o TableRangePartitioningRangePtrOutput) ToTableRangePartitioningRangePtrOutputWithContext(ctx context.Context) TableRangePartitioningRangePtrOutput

type TableState

type TableState struct {
	// Specifies column names to use for data clustering.
	// Up to four top-level columns are allowed, and should be specified in
	// descending priority order.
	Clusterings pulumi.StringArrayInput
	// The time when this table was created, in milliseconds since the epoch.
	CreationTime pulumi.IntPtrInput
	// The dataset ID to create the table in.
	// Changing this forces a new resource to be created.
	DatasetId pulumi.StringPtrInput
	// The field description.
	Description pulumi.StringPtrInput
	// Specifies how the table should be encrypted.
	// If left blank, the table will be encrypted with a Google-managed key; that process
	// is transparent to the user.  Structure is documented below.
	EncryptionConfiguration TableEncryptionConfigurationPtrInput
	// A hash of the resource.
	Etag pulumi.StringPtrInput
	// The time when this table expires, in
	// milliseconds since the epoch. If not present, the table will persist
	// indefinitely. Expired tables will be deleted and their storage
	// reclaimed.
	ExpirationTime pulumi.IntPtrInput
	// Describes the data format,
	// location, and other properties of a table stored outside of BigQuery.
	// By defining these properties, the data source can then be queried as
	// if it were a standard BigQuery table. Structure is documented below.
	ExternalDataConfiguration TableExternalDataConfigurationPtrInput
	// A descriptive name for the table.
	FriendlyName pulumi.StringPtrInput
	// A mapping of labels to assign to the resource.
	Labels pulumi.StringMapInput
	// The time when this table was last modified, in milliseconds since the epoch.
	LastModifiedTime pulumi.IntPtrInput
	// The geographic location where the table resides. This value is inherited from the dataset.
	Location pulumi.StringPtrInput
	// The size of this table in bytes, excluding any data in the streaming buffer.
	NumBytes pulumi.IntPtrInput
	// The number of bytes in the table that are considered "long-term storage".
	NumLongTermBytes pulumi.IntPtrInput
	// The number of rows of data in this table, excluding any data in the streaming buffer.
	NumRows pulumi.IntPtrInput
	// The ID of the project in which the resource belongs. If it
	// is not provided, the provider project is used.
	Project pulumi.StringPtrInput
	// If specified, configures range-based
	// partitioning for this table. Structure is documented below.
	RangePartitioning TableRangePartitioningPtrInput
	// A JSON schema for the table. Schema is required
	// for CSV and JSON formats and is disallowed for Google Cloud
	// Bigtable, Cloud Datastore backups, and Avro formats when using
	// external tables. For more information see the
	// [BigQuery API documentation](https://cloud.google.com/bigquery/docs/reference/rest/v2/tables#resource).
	// ~>**NOTE**: Because this field expects a JSON string, any changes to the
	// string will create a diff, even if the JSON itself hasn't changed.
	// If the API returns a different value for the same schema, e.g. it
	// switched the order of values or replaced `STRUCT` field type with `RECORD`
	// field type, we currently cannot suppress the recurring diff this causes.
	// As a workaround, we recommend using the schema as returned by the API.
	Schema pulumi.StringPtrInput
	// The URI of the created resource.
	SelfLink pulumi.StringPtrInput
	// A unique ID for the resource.
	// Changing this forces a new resource to be created.
	TableId pulumi.StringPtrInput
	// If specified, configures time-based
	// partitioning for this table. Structure is documented below.
	TimePartitioning TableTimePartitioningPtrInput
	// The only type supported is DAY, which will generate
	// one partition per day based on data loading time.
	Type pulumi.StringPtrInput
	// If specified, configures this table as a view.
	// Structure is documented below.
	View TableViewPtrInput
}

func (TableState) ElementType

func (TableState) ElementType() reflect.Type

type TableTimePartitioning

type TableTimePartitioning struct {
	// Number of milliseconds for which to keep the
	// storage for a partition.
	ExpirationMs *int `pulumi:"expirationMs"`
	// The field used to determine how to create a range-based
	// partition.
	Field *string `pulumi:"field"`
	// If set to true, queries over this table
	// require a partition filter that can be used for partition elimination to be
	// specified.
	RequirePartitionFilter *bool `pulumi:"requirePartitionFilter"`
	// The only type supported is DAY, which will generate
	// one partition per day based on data loading time.
	Type string `pulumi:"type"`
}

type TableTimePartitioningArgs

type TableTimePartitioningArgs struct {
	// Number of milliseconds for which to keep the
	// storage for a partition.
	ExpirationMs pulumi.IntPtrInput `pulumi:"expirationMs"`
	// The field used to determine how to create a range-based
	// partition.
	Field pulumi.StringPtrInput `pulumi:"field"`
	// If set to true, queries over this table
	// require a partition filter that can be used for partition elimination to be
	// specified.
	RequirePartitionFilter pulumi.BoolPtrInput `pulumi:"requirePartitionFilter"`
	// The only type supported is DAY, which will generate
	// one partition per day based on data loading time.
	Type pulumi.StringInput `pulumi:"type"`
}

func (TableTimePartitioningArgs) ElementType

func (TableTimePartitioningArgs) ElementType() reflect.Type

func (TableTimePartitioningArgs) ToTableTimePartitioningOutput

func (i TableTimePartitioningArgs) ToTableTimePartitioningOutput() TableTimePartitioningOutput

func (TableTimePartitioningArgs) ToTableTimePartitioningOutputWithContext

func (i TableTimePartitioningArgs) ToTableTimePartitioningOutputWithContext(ctx context.Context) TableTimePartitioningOutput

func (TableTimePartitioningArgs) ToTableTimePartitioningPtrOutput

func (i TableTimePartitioningArgs) ToTableTimePartitioningPtrOutput() TableTimePartitioningPtrOutput

func (TableTimePartitioningArgs) ToTableTimePartitioningPtrOutputWithContext

func (i TableTimePartitioningArgs) ToTableTimePartitioningPtrOutputWithContext(ctx context.Context) TableTimePartitioningPtrOutput

type TableTimePartitioningInput

type TableTimePartitioningInput interface {
	pulumi.Input

	ToTableTimePartitioningOutput() TableTimePartitioningOutput
	ToTableTimePartitioningOutputWithContext(context.Context) TableTimePartitioningOutput
}

TableTimePartitioningInput is an input type that accepts TableTimePartitioningArgs and TableTimePartitioningOutput values. You can construct a concrete instance of `TableTimePartitioningInput` via:

TableTimePartitioningArgs{...}

type TableTimePartitioningOutput

type TableTimePartitioningOutput struct{ *pulumi.OutputState }

func (TableTimePartitioningOutput) ElementType

func (TableTimePartitioningOutput) ExpirationMs

Number of milliseconds for which to keep the storage for a partition.

func (TableTimePartitioningOutput) Field

The field used to determine how to create a range-based partition.

func (TableTimePartitioningOutput) RequirePartitionFilter

func (o TableTimePartitioningOutput) RequirePartitionFilter() pulumi.BoolPtrOutput

If set to true, queries over this table require a partition filter that can be used for partition elimination to be specified.

func (TableTimePartitioningOutput) ToTableTimePartitioningOutput

func (o TableTimePartitioningOutput) ToTableTimePartitioningOutput() TableTimePartitioningOutput

func (TableTimePartitioningOutput) ToTableTimePartitioningOutputWithContext

func (o TableTimePartitioningOutput) ToTableTimePartitioningOutputWithContext(ctx context.Context) TableTimePartitioningOutput

func (TableTimePartitioningOutput) ToTableTimePartitioningPtrOutput

func (o TableTimePartitioningOutput) ToTableTimePartitioningPtrOutput() TableTimePartitioningPtrOutput

func (TableTimePartitioningOutput) ToTableTimePartitioningPtrOutputWithContext

func (o TableTimePartitioningOutput) ToTableTimePartitioningPtrOutputWithContext(ctx context.Context) TableTimePartitioningPtrOutput

func (TableTimePartitioningOutput) Type

The only type supported is DAY, which will generate one partition per day based on data loading time.

type TableTimePartitioningPtrInput

type TableTimePartitioningPtrInput interface {
	pulumi.Input

	ToTableTimePartitioningPtrOutput() TableTimePartitioningPtrOutput
	ToTableTimePartitioningPtrOutputWithContext(context.Context) TableTimePartitioningPtrOutput
}

TableTimePartitioningPtrInput is an input type that accepts TableTimePartitioningArgs, TableTimePartitioningPtr and TableTimePartitioningPtrOutput values. You can construct a concrete instance of `TableTimePartitioningPtrInput` via:

		 TableTimePartitioningArgs{...}

 or:

		 nil

type TableTimePartitioningPtrOutput

type TableTimePartitioningPtrOutput struct{ *pulumi.OutputState }

func (TableTimePartitioningPtrOutput) Elem

func (TableTimePartitioningPtrOutput) ElementType

func (TableTimePartitioningPtrOutput) ExpirationMs

Number of milliseconds for which to keep the storage for a partition.

func (TableTimePartitioningPtrOutput) Field

The field used to determine how to create a range-based partition.

func (TableTimePartitioningPtrOutput) RequirePartitionFilter

func (o TableTimePartitioningPtrOutput) RequirePartitionFilter() pulumi.BoolPtrOutput

If set to true, queries over this table require a partition filter that can be used for partition elimination to be specified.

func (TableTimePartitioningPtrOutput) ToTableTimePartitioningPtrOutput

func (o TableTimePartitioningPtrOutput) ToTableTimePartitioningPtrOutput() TableTimePartitioningPtrOutput

func (TableTimePartitioningPtrOutput) ToTableTimePartitioningPtrOutputWithContext

func (o TableTimePartitioningPtrOutput) ToTableTimePartitioningPtrOutputWithContext(ctx context.Context) TableTimePartitioningPtrOutput

func (TableTimePartitioningPtrOutput) Type

The only type supported is DAY, which will generate one partition per day based on data loading time.

type TableView

type TableView struct {
	// A query that BigQuery executes when the view is referenced.
	Query string `pulumi:"query"`
	// Specifies whether to use BigQuery's legacy SQL for this view.
	// The default value is true. If set to false, the view will use BigQuery's standard SQL.
	UseLegacySql *bool `pulumi:"useLegacySql"`
}

type TableViewArgs

type TableViewArgs struct {
	// A query that BigQuery executes when the view is referenced.
	Query pulumi.StringInput `pulumi:"query"`
	// Specifies whether to use BigQuery's legacy SQL for this view.
	// The default value is true. If set to false, the view will use BigQuery's standard SQL.
	UseLegacySql pulumi.BoolPtrInput `pulumi:"useLegacySql"`
}

func (TableViewArgs) ElementType

func (TableViewArgs) ElementType() reflect.Type

func (TableViewArgs) ToTableViewOutput

func (i TableViewArgs) ToTableViewOutput() TableViewOutput

func (TableViewArgs) ToTableViewOutputWithContext

func (i TableViewArgs) ToTableViewOutputWithContext(ctx context.Context) TableViewOutput

func (TableViewArgs) ToTableViewPtrOutput

func (i TableViewArgs) ToTableViewPtrOutput() TableViewPtrOutput

func (TableViewArgs) ToTableViewPtrOutputWithContext

func (i TableViewArgs) ToTableViewPtrOutputWithContext(ctx context.Context) TableViewPtrOutput

type TableViewInput

type TableViewInput interface {
	pulumi.Input

	ToTableViewOutput() TableViewOutput
	ToTableViewOutputWithContext(context.Context) TableViewOutput
}

TableViewInput is an input type that accepts TableViewArgs and TableViewOutput values. You can construct a concrete instance of `TableViewInput` via:

TableViewArgs{...}

type TableViewOutput

type TableViewOutput struct{ *pulumi.OutputState }

func (TableViewOutput) ElementType

func (TableViewOutput) ElementType() reflect.Type

func (TableViewOutput) Query

A query that BigQuery executes when the view is referenced.

func (TableViewOutput) ToTableViewOutput

func (o TableViewOutput) ToTableViewOutput() TableViewOutput

func (TableViewOutput) ToTableViewOutputWithContext

func (o TableViewOutput) ToTableViewOutputWithContext(ctx context.Context) TableViewOutput

func (TableViewOutput) ToTableViewPtrOutput

func (o TableViewOutput) ToTableViewPtrOutput() TableViewPtrOutput

func (TableViewOutput) ToTableViewPtrOutputWithContext

func (o TableViewOutput) ToTableViewPtrOutputWithContext(ctx context.Context) TableViewPtrOutput

func (TableViewOutput) UseLegacySql

func (o TableViewOutput) UseLegacySql() pulumi.BoolPtrOutput

Specifies whether to use BigQuery's legacy SQL for this view. The default value is true. If set to false, the view will use BigQuery's standard SQL.

type TableViewPtrInput

type TableViewPtrInput interface {
	pulumi.Input

	ToTableViewPtrOutput() TableViewPtrOutput
	ToTableViewPtrOutputWithContext(context.Context) TableViewPtrOutput
}

TableViewPtrInput is an input type that accepts TableViewArgs, TableViewPtr and TableViewPtrOutput values. You can construct a concrete instance of `TableViewPtrInput` via:

		 TableViewArgs{...}

 or:

		 nil

func TableViewPtr

func TableViewPtr(v *TableViewArgs) TableViewPtrInput

type TableViewPtrOutput

type TableViewPtrOutput struct{ *pulumi.OutputState }

func (TableViewPtrOutput) Elem

func (TableViewPtrOutput) ElementType

func (TableViewPtrOutput) ElementType() reflect.Type

func (TableViewPtrOutput) Query

A query that BigQuery executes when the view is referenced.

func (TableViewPtrOutput) ToTableViewPtrOutput

func (o TableViewPtrOutput) ToTableViewPtrOutput() TableViewPtrOutput

func (TableViewPtrOutput) ToTableViewPtrOutputWithContext

func (o TableViewPtrOutput) ToTableViewPtrOutputWithContext(ctx context.Context) TableViewPtrOutput

func (TableViewPtrOutput) UseLegacySql

func (o TableViewPtrOutput) UseLegacySql() pulumi.BoolPtrOutput

Specifies whether to use BigQuery's legacy SQL for this view. The default value is true. If set to false, the view will use BigQuery's standard SQL.

Jump to

Keyboard shortcuts

? : This menu
/ : Search site
f or F : Jump to
y or Y : Canonical URL