bigquery

package
v4.0.0 Latest Latest
Warning

This package is not in the latest version of its module.

Go to latest
Published: Oct 26, 2020 License: Apache-2.0 Imports: 4 Imported by: 0

Documentation

Index

Constants

This section is empty.

Variables

This section is empty.

Functions

This section is empty.

Types

type AppProfile

type AppProfile struct {
	pulumi.CustomResourceState

	// The unique name of the app profile in the form `[_a-zA-Z0-9][-_.a-zA-Z0-9]*`.
	AppProfileId pulumi.StringOutput `pulumi:"appProfileId"`
	// Long form description of the use case for this app profile.
	Description pulumi.StringPtrOutput `pulumi:"description"`
	// If true, ignore safety checks when deleting/updating the app profile.
	IgnoreWarnings pulumi.BoolPtrOutput `pulumi:"ignoreWarnings"`
	// The name of the instance to create the app profile within.
	Instance pulumi.StringPtrOutput `pulumi:"instance"`
	// If true, read/write requests are routed to the nearest cluster in the instance, and will fail over to the nearest cluster that is available
	// in the event of transient errors or delays. Clusters in a region are considered equidistant. Choosing this option sacrifices read-your-writes
	// consistency to improve availability.
	MultiClusterRoutingUseAny pulumi.BoolPtrOutput `pulumi:"multiClusterRoutingUseAny"`
	// The unique name of the requested app profile. Values are of the form
	// 'projects/<project>/instances/<instance>/appProfiles/<appProfileId>'.
	Name pulumi.StringOutput `pulumi:"name"`
	// The ID of the project in which the resource belongs.
	// If it is not provided, the provider project is used.
	Project pulumi.StringOutput `pulumi:"project"`
	// Use a single-cluster routing policy.
	// Structure is documented below.
	SingleClusterRouting AppProfileSingleClusterRoutingPtrOutput `pulumi:"singleClusterRouting"`
}

App profile is a configuration object describing how Cloud Bigtable should treat traffic from a particular end user application.

## Example Usage

func GetAppProfile

func GetAppProfile(ctx *pulumi.Context,
	name string, id pulumi.IDInput, state *AppProfileState, opts ...pulumi.ResourceOption) (*AppProfile, error)

GetAppProfile gets an existing AppProfile resource's state with the given name, ID, and optional state properties that are used to uniquely qualify the lookup (nil if not required).

func NewAppProfile

func NewAppProfile(ctx *pulumi.Context,
	name string, args *AppProfileArgs, opts ...pulumi.ResourceOption) (*AppProfile, error)

NewAppProfile registers a new resource with the given unique name, arguments, and options.

type AppProfileArgs

type AppProfileArgs struct {
	// The unique name of the app profile in the form `[_a-zA-Z0-9][-_.a-zA-Z0-9]*`.
	AppProfileId pulumi.StringInput
	// Long form description of the use case for this app profile.
	Description pulumi.StringPtrInput
	// If true, ignore safety checks when deleting/updating the app profile.
	IgnoreWarnings pulumi.BoolPtrInput
	// The name of the instance to create the app profile within.
	Instance pulumi.StringPtrInput
	// If true, read/write requests are routed to the nearest cluster in the instance, and will fail over to the nearest cluster that is available
	// in the event of transient errors or delays. Clusters in a region are considered equidistant. Choosing this option sacrifices read-your-writes
	// consistency to improve availability.
	MultiClusterRoutingUseAny pulumi.BoolPtrInput
	// The ID of the project in which the resource belongs.
	// If it is not provided, the provider project is used.
	Project pulumi.StringPtrInput
	// Use a single-cluster routing policy.
	// Structure is documented below.
	SingleClusterRouting AppProfileSingleClusterRoutingPtrInput
}

The set of arguments for constructing a AppProfile resource.

func (AppProfileArgs) ElementType

func (AppProfileArgs) ElementType() reflect.Type

type AppProfileSingleClusterRouting

type AppProfileSingleClusterRouting struct {
	// If true, CheckAndMutateRow and ReadModifyWriteRow requests are allowed by this app profile.
	// It is unsafe to send these requests to the same table/row/column in multiple clusters.
	AllowTransactionalWrites *bool `pulumi:"allowTransactionalWrites"`
	// The cluster to which read/write requests should be routed.
	ClusterId string `pulumi:"clusterId"`
}

type AppProfileSingleClusterRoutingArgs

type AppProfileSingleClusterRoutingArgs struct {
	// If true, CheckAndMutateRow and ReadModifyWriteRow requests are allowed by this app profile.
	// It is unsafe to send these requests to the same table/row/column in multiple clusters.
	AllowTransactionalWrites pulumi.BoolPtrInput `pulumi:"allowTransactionalWrites"`
	// The cluster to which read/write requests should be routed.
	ClusterId pulumi.StringInput `pulumi:"clusterId"`
}

func (AppProfileSingleClusterRoutingArgs) ElementType

func (AppProfileSingleClusterRoutingArgs) ToAppProfileSingleClusterRoutingOutput

func (i AppProfileSingleClusterRoutingArgs) ToAppProfileSingleClusterRoutingOutput() AppProfileSingleClusterRoutingOutput

func (AppProfileSingleClusterRoutingArgs) ToAppProfileSingleClusterRoutingOutputWithContext

func (i AppProfileSingleClusterRoutingArgs) ToAppProfileSingleClusterRoutingOutputWithContext(ctx context.Context) AppProfileSingleClusterRoutingOutput

func (AppProfileSingleClusterRoutingArgs) ToAppProfileSingleClusterRoutingPtrOutput

func (i AppProfileSingleClusterRoutingArgs) ToAppProfileSingleClusterRoutingPtrOutput() AppProfileSingleClusterRoutingPtrOutput

func (AppProfileSingleClusterRoutingArgs) ToAppProfileSingleClusterRoutingPtrOutputWithContext

func (i AppProfileSingleClusterRoutingArgs) ToAppProfileSingleClusterRoutingPtrOutputWithContext(ctx context.Context) AppProfileSingleClusterRoutingPtrOutput

type AppProfileSingleClusterRoutingInput

type AppProfileSingleClusterRoutingInput interface {
	pulumi.Input

	ToAppProfileSingleClusterRoutingOutput() AppProfileSingleClusterRoutingOutput
	ToAppProfileSingleClusterRoutingOutputWithContext(context.Context) AppProfileSingleClusterRoutingOutput
}

AppProfileSingleClusterRoutingInput is an input type that accepts AppProfileSingleClusterRoutingArgs and AppProfileSingleClusterRoutingOutput values. You can construct a concrete instance of `AppProfileSingleClusterRoutingInput` via:

AppProfileSingleClusterRoutingArgs{...}

type AppProfileSingleClusterRoutingOutput

type AppProfileSingleClusterRoutingOutput struct{ *pulumi.OutputState }

func (AppProfileSingleClusterRoutingOutput) AllowTransactionalWrites

func (o AppProfileSingleClusterRoutingOutput) AllowTransactionalWrites() pulumi.BoolPtrOutput

If true, CheckAndMutateRow and ReadModifyWriteRow requests are allowed by this app profile. It is unsafe to send these requests to the same table/row/column in multiple clusters.

func (AppProfileSingleClusterRoutingOutput) ClusterId

The cluster to which read/write requests should be routed.

func (AppProfileSingleClusterRoutingOutput) ElementType

func (AppProfileSingleClusterRoutingOutput) ToAppProfileSingleClusterRoutingOutput

func (o AppProfileSingleClusterRoutingOutput) ToAppProfileSingleClusterRoutingOutput() AppProfileSingleClusterRoutingOutput

func (AppProfileSingleClusterRoutingOutput) ToAppProfileSingleClusterRoutingOutputWithContext

func (o AppProfileSingleClusterRoutingOutput) ToAppProfileSingleClusterRoutingOutputWithContext(ctx context.Context) AppProfileSingleClusterRoutingOutput

func (AppProfileSingleClusterRoutingOutput) ToAppProfileSingleClusterRoutingPtrOutput

func (o AppProfileSingleClusterRoutingOutput) ToAppProfileSingleClusterRoutingPtrOutput() AppProfileSingleClusterRoutingPtrOutput

func (AppProfileSingleClusterRoutingOutput) ToAppProfileSingleClusterRoutingPtrOutputWithContext

func (o AppProfileSingleClusterRoutingOutput) ToAppProfileSingleClusterRoutingPtrOutputWithContext(ctx context.Context) AppProfileSingleClusterRoutingPtrOutput

type AppProfileSingleClusterRoutingPtrInput

type AppProfileSingleClusterRoutingPtrInput interface {
	pulumi.Input

	ToAppProfileSingleClusterRoutingPtrOutput() AppProfileSingleClusterRoutingPtrOutput
	ToAppProfileSingleClusterRoutingPtrOutputWithContext(context.Context) AppProfileSingleClusterRoutingPtrOutput
}

AppProfileSingleClusterRoutingPtrInput is an input type that accepts AppProfileSingleClusterRoutingArgs, AppProfileSingleClusterRoutingPtr and AppProfileSingleClusterRoutingPtrOutput values. You can construct a concrete instance of `AppProfileSingleClusterRoutingPtrInput` via:

        AppProfileSingleClusterRoutingArgs{...}

or:

        nil

type AppProfileSingleClusterRoutingPtrOutput

type AppProfileSingleClusterRoutingPtrOutput struct{ *pulumi.OutputState }

func (AppProfileSingleClusterRoutingPtrOutput) AllowTransactionalWrites

func (o AppProfileSingleClusterRoutingPtrOutput) AllowTransactionalWrites() pulumi.BoolPtrOutput

If true, CheckAndMutateRow and ReadModifyWriteRow requests are allowed by this app profile. It is unsafe to send these requests to the same table/row/column in multiple clusters.

func (AppProfileSingleClusterRoutingPtrOutput) ClusterId

The cluster to which read/write requests should be routed.

func (AppProfileSingleClusterRoutingPtrOutput) Elem

func (AppProfileSingleClusterRoutingPtrOutput) ElementType

func (AppProfileSingleClusterRoutingPtrOutput) ToAppProfileSingleClusterRoutingPtrOutput

func (o AppProfileSingleClusterRoutingPtrOutput) ToAppProfileSingleClusterRoutingPtrOutput() AppProfileSingleClusterRoutingPtrOutput

func (AppProfileSingleClusterRoutingPtrOutput) ToAppProfileSingleClusterRoutingPtrOutputWithContext

func (o AppProfileSingleClusterRoutingPtrOutput) ToAppProfileSingleClusterRoutingPtrOutputWithContext(ctx context.Context) AppProfileSingleClusterRoutingPtrOutput

type AppProfileState

type AppProfileState struct {
	// The unique name of the app profile in the form `[_a-zA-Z0-9][-_.a-zA-Z0-9]*`.
	AppProfileId pulumi.StringPtrInput
	// Long form description of the use case for this app profile.
	Description pulumi.StringPtrInput
	// If true, ignore safety checks when deleting/updating the app profile.
	IgnoreWarnings pulumi.BoolPtrInput
	// The name of the instance to create the app profile within.
	Instance pulumi.StringPtrInput
	// If true, read/write requests are routed to the nearest cluster in the instance, and will fail over to the nearest cluster that is available
	// in the event of transient errors or delays. Clusters in a region are considered equidistant. Choosing this option sacrifices read-your-writes
	// consistency to improve availability.
	MultiClusterRoutingUseAny pulumi.BoolPtrInput
	// The unique name of the requested app profile. Values are of the form
	// 'projects/<project>/instances/<instance>/appProfiles/<appProfileId>'.
	Name pulumi.StringPtrInput
	// The ID of the project in which the resource belongs.
	// If it is not provided, the provider project is used.
	Project pulumi.StringPtrInput
	// Use a single-cluster routing policy.
	// Structure is documented below.
	SingleClusterRouting AppProfileSingleClusterRoutingPtrInput
}

func (AppProfileState) ElementType

func (AppProfileState) ElementType() reflect.Type

type Connection

type Connection struct {
	pulumi.CustomResourceState

	// Cloud SQL properties.
	// Structure is documented below.
	CloudSql ConnectionCloudSqlOutput `pulumi:"cloudSql"`
	// Optional connection id that should be assigned to the created connection.
	ConnectionId pulumi.StringPtrOutput `pulumi:"connectionId"`
	// A descriptive description for the connection
	Description pulumi.StringPtrOutput `pulumi:"description"`
	// A descriptive name for the connection
	FriendlyName pulumi.StringPtrOutput `pulumi:"friendlyName"`
	// True if the connection has credential assigned.
	HasCredential pulumi.BoolOutput `pulumi:"hasCredential"`
	// The geographic location where the connection should reside.
	// Cloud SQL instance must be in the same location as the connection
	// with following exceptions: Cloud SQL us-central1 maps to BigQuery US, Cloud SQL europe-west1 maps to BigQuery EU.
	// Examples: US, EU, asia-northeast1, us-central1, europe-west1. The default value is US.
	Location pulumi.StringPtrOutput `pulumi:"location"`
	// The resource name of the connection in the form of:
	// "projects/{project_id}/locations/{location_id}/connections/{connectionId}"
	Name pulumi.StringOutput `pulumi:"name"`
	// The ID of the project in which the resource belongs.
	// If it is not provided, the provider project is used.
	Project pulumi.StringOutput `pulumi:"project"`
}

A connection allows BigQuery connections to external data sources..

To get more information about Connection, see:

* [API documentation](https://cloud.google.com/bigquery/docs/reference/bigqueryconnection/rest/v1beta1/projects.locations.connections/create) * How-to Guides

> **Warning:** All arguments including `cloud_sql.credential.password` will be stored in the raw state as plain-text. [Read more about sensitive data in state](https://www.terraform.io/docs/state/sensitive-data.html).

## Example Usage

func GetConnection

func GetConnection(ctx *pulumi.Context,
	name string, id pulumi.IDInput, state *ConnectionState, opts ...pulumi.ResourceOption) (*Connection, error)

GetConnection gets an existing Connection resource's state with the given name, ID, and optional state properties that are used to uniquely qualify the lookup (nil if not required).

func NewConnection

func NewConnection(ctx *pulumi.Context,
	name string, args *ConnectionArgs, opts ...pulumi.ResourceOption) (*Connection, error)

NewConnection registers a new resource with the given unique name, arguments, and options.

type ConnectionArgs

type ConnectionArgs struct {
	// Cloud SQL properties.
	// Structure is documented below.
	CloudSql ConnectionCloudSqlInput
	// Optional connection id that should be assigned to the created connection.
	ConnectionId pulumi.StringPtrInput
	// A descriptive description for the connection
	Description pulumi.StringPtrInput
	// A descriptive name for the connection
	FriendlyName pulumi.StringPtrInput
	// The geographic location where the connection should reside.
	// Cloud SQL instance must be in the same location as the connection
	// with following exceptions: Cloud SQL us-central1 maps to BigQuery US, Cloud SQL europe-west1 maps to BigQuery EU.
	// Examples: US, EU, asia-northeast1, us-central1, europe-west1. The default value is US.
	Location pulumi.StringPtrInput
	// The ID of the project in which the resource belongs.
	// If it is not provided, the provider project is used.
	Project pulumi.StringPtrInput
}

The set of arguments for constructing a Connection resource.

func (ConnectionArgs) ElementType

func (ConnectionArgs) ElementType() reflect.Type

type ConnectionCloudSql

type ConnectionCloudSql struct {
	// Cloud SQL properties.
	// Structure is documented below.
	Credential ConnectionCloudSqlCredential `pulumi:"credential"`
	// Database name.
	Database string `pulumi:"database"`
	// Cloud SQL instance ID in the form project:location:instance.
	InstanceId string `pulumi:"instanceId"`
	// Type of the Cloud SQL database.
	// Possible values are `DATABASE_TYPE_UNSPECIFIED`, `POSTGRES`, and `MYSQL`.
	Type string `pulumi:"type"`
}

type ConnectionCloudSqlArgs

type ConnectionCloudSqlArgs struct {
	// Cloud SQL properties.
	// Structure is documented below.
	Credential ConnectionCloudSqlCredentialInput `pulumi:"credential"`
	// Database name.
	Database pulumi.StringInput `pulumi:"database"`
	// Cloud SQL instance ID in the form project:location:instance.
	InstanceId pulumi.StringInput `pulumi:"instanceId"`
	// Type of the Cloud SQL database.
	// Possible values are `DATABASE_TYPE_UNSPECIFIED`, `POSTGRES`, and `MYSQL`.
	Type pulumi.StringInput `pulumi:"type"`
}

func (ConnectionCloudSqlArgs) ElementType

func (ConnectionCloudSqlArgs) ElementType() reflect.Type

func (ConnectionCloudSqlArgs) ToConnectionCloudSqlOutput

func (i ConnectionCloudSqlArgs) ToConnectionCloudSqlOutput() ConnectionCloudSqlOutput

func (ConnectionCloudSqlArgs) ToConnectionCloudSqlOutputWithContext

func (i ConnectionCloudSqlArgs) ToConnectionCloudSqlOutputWithContext(ctx context.Context) ConnectionCloudSqlOutput

func (ConnectionCloudSqlArgs) ToConnectionCloudSqlPtrOutput

func (i ConnectionCloudSqlArgs) ToConnectionCloudSqlPtrOutput() ConnectionCloudSqlPtrOutput

func (ConnectionCloudSqlArgs) ToConnectionCloudSqlPtrOutputWithContext

func (i ConnectionCloudSqlArgs) ToConnectionCloudSqlPtrOutputWithContext(ctx context.Context) ConnectionCloudSqlPtrOutput

type ConnectionCloudSqlCredential

type ConnectionCloudSqlCredential struct {
	// Password for database.
	// **Note**: This property is sensitive and will not be displayed in the plan.
	Password string `pulumi:"password"`
	// Username for database.
	Username string `pulumi:"username"`
}

type ConnectionCloudSqlCredentialArgs

type ConnectionCloudSqlCredentialArgs struct {
	// Password for database.
	// **Note**: This property is sensitive and will not be displayed in the plan.
	Password pulumi.StringInput `pulumi:"password"`
	// Username for database.
	Username pulumi.StringInput `pulumi:"username"`
}

func (ConnectionCloudSqlCredentialArgs) ElementType

func (ConnectionCloudSqlCredentialArgs) ToConnectionCloudSqlCredentialOutput

func (i ConnectionCloudSqlCredentialArgs) ToConnectionCloudSqlCredentialOutput() ConnectionCloudSqlCredentialOutput

func (ConnectionCloudSqlCredentialArgs) ToConnectionCloudSqlCredentialOutputWithContext

func (i ConnectionCloudSqlCredentialArgs) ToConnectionCloudSqlCredentialOutputWithContext(ctx context.Context) ConnectionCloudSqlCredentialOutput

func (ConnectionCloudSqlCredentialArgs) ToConnectionCloudSqlCredentialPtrOutput

func (i ConnectionCloudSqlCredentialArgs) ToConnectionCloudSqlCredentialPtrOutput() ConnectionCloudSqlCredentialPtrOutput

func (ConnectionCloudSqlCredentialArgs) ToConnectionCloudSqlCredentialPtrOutputWithContext

func (i ConnectionCloudSqlCredentialArgs) ToConnectionCloudSqlCredentialPtrOutputWithContext(ctx context.Context) ConnectionCloudSqlCredentialPtrOutput

type ConnectionCloudSqlCredentialInput

type ConnectionCloudSqlCredentialInput interface {
	pulumi.Input

	ToConnectionCloudSqlCredentialOutput() ConnectionCloudSqlCredentialOutput
	ToConnectionCloudSqlCredentialOutputWithContext(context.Context) ConnectionCloudSqlCredentialOutput
}

ConnectionCloudSqlCredentialInput is an input type that accepts ConnectionCloudSqlCredentialArgs and ConnectionCloudSqlCredentialOutput values. You can construct a concrete instance of `ConnectionCloudSqlCredentialInput` via:

ConnectionCloudSqlCredentialArgs{...}

type ConnectionCloudSqlCredentialOutput

type ConnectionCloudSqlCredentialOutput struct{ *pulumi.OutputState }

func (ConnectionCloudSqlCredentialOutput) ElementType

func (ConnectionCloudSqlCredentialOutput) Password

Password for database. **Note**: This property is sensitive and will not be displayed in the plan.

func (ConnectionCloudSqlCredentialOutput) ToConnectionCloudSqlCredentialOutput

func (o ConnectionCloudSqlCredentialOutput) ToConnectionCloudSqlCredentialOutput() ConnectionCloudSqlCredentialOutput

func (ConnectionCloudSqlCredentialOutput) ToConnectionCloudSqlCredentialOutputWithContext

func (o ConnectionCloudSqlCredentialOutput) ToConnectionCloudSqlCredentialOutputWithContext(ctx context.Context) ConnectionCloudSqlCredentialOutput

func (ConnectionCloudSqlCredentialOutput) ToConnectionCloudSqlCredentialPtrOutput

func (o ConnectionCloudSqlCredentialOutput) ToConnectionCloudSqlCredentialPtrOutput() ConnectionCloudSqlCredentialPtrOutput

func (ConnectionCloudSqlCredentialOutput) ToConnectionCloudSqlCredentialPtrOutputWithContext

func (o ConnectionCloudSqlCredentialOutput) ToConnectionCloudSqlCredentialPtrOutputWithContext(ctx context.Context) ConnectionCloudSqlCredentialPtrOutput

func (ConnectionCloudSqlCredentialOutput) Username

Username for database.

type ConnectionCloudSqlCredentialPtrInput

type ConnectionCloudSqlCredentialPtrInput interface {
	pulumi.Input

	ToConnectionCloudSqlCredentialPtrOutput() ConnectionCloudSqlCredentialPtrOutput
	ToConnectionCloudSqlCredentialPtrOutputWithContext(context.Context) ConnectionCloudSqlCredentialPtrOutput
}

ConnectionCloudSqlCredentialPtrInput is an input type that accepts ConnectionCloudSqlCredentialArgs, ConnectionCloudSqlCredentialPtr and ConnectionCloudSqlCredentialPtrOutput values. You can construct a concrete instance of `ConnectionCloudSqlCredentialPtrInput` via:

        ConnectionCloudSqlCredentialArgs{...}

or:

        nil

type ConnectionCloudSqlCredentialPtrOutput

type ConnectionCloudSqlCredentialPtrOutput struct{ *pulumi.OutputState }

func (ConnectionCloudSqlCredentialPtrOutput) Elem

func (ConnectionCloudSqlCredentialPtrOutput) ElementType

func (ConnectionCloudSqlCredentialPtrOutput) Password

Password for database. **Note**: This property is sensitive and will not be displayed in the plan.

func (ConnectionCloudSqlCredentialPtrOutput) ToConnectionCloudSqlCredentialPtrOutput

func (o ConnectionCloudSqlCredentialPtrOutput) ToConnectionCloudSqlCredentialPtrOutput() ConnectionCloudSqlCredentialPtrOutput

func (ConnectionCloudSqlCredentialPtrOutput) ToConnectionCloudSqlCredentialPtrOutputWithContext

func (o ConnectionCloudSqlCredentialPtrOutput) ToConnectionCloudSqlCredentialPtrOutputWithContext(ctx context.Context) ConnectionCloudSqlCredentialPtrOutput

func (ConnectionCloudSqlCredentialPtrOutput) Username

Username for database.

type ConnectionCloudSqlInput

type ConnectionCloudSqlInput interface {
	pulumi.Input

	ToConnectionCloudSqlOutput() ConnectionCloudSqlOutput
	ToConnectionCloudSqlOutputWithContext(context.Context) ConnectionCloudSqlOutput
}

ConnectionCloudSqlInput is an input type that accepts ConnectionCloudSqlArgs and ConnectionCloudSqlOutput values. You can construct a concrete instance of `ConnectionCloudSqlInput` via:

ConnectionCloudSqlArgs{...}

type ConnectionCloudSqlOutput

type ConnectionCloudSqlOutput struct{ *pulumi.OutputState }

func (ConnectionCloudSqlOutput) Credential

Cloud SQL properties. Structure is documented below.

func (ConnectionCloudSqlOutput) Database

Database name.

func (ConnectionCloudSqlOutput) ElementType

func (ConnectionCloudSqlOutput) ElementType() reflect.Type

func (ConnectionCloudSqlOutput) InstanceId

Cloud SQL instance ID in the form project:location:instance.

func (ConnectionCloudSqlOutput) ToConnectionCloudSqlOutput

func (o ConnectionCloudSqlOutput) ToConnectionCloudSqlOutput() ConnectionCloudSqlOutput

func (ConnectionCloudSqlOutput) ToConnectionCloudSqlOutputWithContext

func (o ConnectionCloudSqlOutput) ToConnectionCloudSqlOutputWithContext(ctx context.Context) ConnectionCloudSqlOutput

func (ConnectionCloudSqlOutput) ToConnectionCloudSqlPtrOutput

func (o ConnectionCloudSqlOutput) ToConnectionCloudSqlPtrOutput() ConnectionCloudSqlPtrOutput

func (ConnectionCloudSqlOutput) ToConnectionCloudSqlPtrOutputWithContext

func (o ConnectionCloudSqlOutput) ToConnectionCloudSqlPtrOutputWithContext(ctx context.Context) ConnectionCloudSqlPtrOutput

func (ConnectionCloudSqlOutput) Type

Type of the Cloud SQL database. Possible values are `DATABASE_TYPE_UNSPECIFIED`, `POSTGRES`, and `MYSQL`.

type ConnectionCloudSqlPtrInput

type ConnectionCloudSqlPtrInput interface {
	pulumi.Input

	ToConnectionCloudSqlPtrOutput() ConnectionCloudSqlPtrOutput
	ToConnectionCloudSqlPtrOutputWithContext(context.Context) ConnectionCloudSqlPtrOutput
}

ConnectionCloudSqlPtrInput is an input type that accepts ConnectionCloudSqlArgs, ConnectionCloudSqlPtr and ConnectionCloudSqlPtrOutput values. You can construct a concrete instance of `ConnectionCloudSqlPtrInput` via:

        ConnectionCloudSqlArgs{...}

or:

        nil

type ConnectionCloudSqlPtrOutput

type ConnectionCloudSqlPtrOutput struct{ *pulumi.OutputState }

func (ConnectionCloudSqlPtrOutput) Credential

Cloud SQL properties. Structure is documented below.

func (ConnectionCloudSqlPtrOutput) Database

Database name.

func (ConnectionCloudSqlPtrOutput) Elem

func (ConnectionCloudSqlPtrOutput) ElementType

func (ConnectionCloudSqlPtrOutput) InstanceId

Cloud SQL instance ID in the form project:location:instance.

func (ConnectionCloudSqlPtrOutput) ToConnectionCloudSqlPtrOutput

func (o ConnectionCloudSqlPtrOutput) ToConnectionCloudSqlPtrOutput() ConnectionCloudSqlPtrOutput

func (ConnectionCloudSqlPtrOutput) ToConnectionCloudSqlPtrOutputWithContext

func (o ConnectionCloudSqlPtrOutput) ToConnectionCloudSqlPtrOutputWithContext(ctx context.Context) ConnectionCloudSqlPtrOutput

func (ConnectionCloudSqlPtrOutput) Type

Type of the Cloud SQL database. Possible values are `DATABASE_TYPE_UNSPECIFIED`, `POSTGRES`, and `MYSQL`.

type ConnectionState

type ConnectionState struct {
	// Cloud SQL properties.
	// Structure is documented below.
	CloudSql ConnectionCloudSqlPtrInput
	// Optional connection id that should be assigned to the created connection.
	ConnectionId pulumi.StringPtrInput
	// A descriptive description for the connection
	Description pulumi.StringPtrInput
	// A descriptive name for the connection
	FriendlyName pulumi.StringPtrInput
	// True if the connection has credential assigned.
	HasCredential pulumi.BoolPtrInput
	// The geographic location where the connection should reside.
	// Cloud SQL instance must be in the same location as the connection
	// with following exceptions: Cloud SQL us-central1 maps to BigQuery US, Cloud SQL europe-west1 maps to BigQuery EU.
	// Examples: US, EU, asia-northeast1, us-central1, europe-west1. The default value is US.
	Location pulumi.StringPtrInput
	// The resource name of the connection in the form of:
	// "projects/{project_id}/locations/{location_id}/connections/{connectionId}"
	Name pulumi.StringPtrInput
	// The ID of the project in which the resource belongs.
	// If it is not provided, the provider project is used.
	Project pulumi.StringPtrInput
}

func (ConnectionState) ElementType

func (ConnectionState) ElementType() reflect.Type

type DataTransferConfig

type DataTransferConfig struct {
	pulumi.CustomResourceState

	// The number of days to look back to automatically refresh the data.
	// For example, if dataRefreshWindowDays = 10, then every day BigQuery
	// reingests data for [today-10, today-1], rather than ingesting data for
	// just [today-1]. Only valid if the data source supports the feature.
	// Set the value to 0 to use the default value.
	DataRefreshWindowDays pulumi.IntPtrOutput `pulumi:"dataRefreshWindowDays"`
	// The data source id. Cannot be changed once the transfer config is created.
	DataSourceId pulumi.StringOutput `pulumi:"dataSourceId"`
	// The BigQuery target dataset id.
	DestinationDatasetId pulumi.StringOutput `pulumi:"destinationDatasetId"`
	// When set to true, no runs are scheduled for a given transfer.
	Disabled pulumi.BoolPtrOutput `pulumi:"disabled"`
	// The user specified display name for the transfer config.
	DisplayName pulumi.StringOutput `pulumi:"displayName"`
	// The geographic location where the transfer config should reside.
	// Examples: US, EU, asia-northeast1. The default value is US.
	Location pulumi.StringPtrOutput `pulumi:"location"`
	// The resource name of the transfer config. Transfer config names have the form
	// projects/{projectId}/locations/{location}/transferConfigs/{configId}. Where configId is usually a uuid, but this is not
	// required. The name is ignored when creating a transfer config.
	Name pulumi.StringOutput `pulumi:"name"`
	// Pub/Sub topic where notifications will be sent after transfer runs
	// associated with this transfer config finish.
	NotificationPubsubTopic pulumi.StringPtrOutput `pulumi:"notificationPubsubTopic"`
	// These parameters are specific to each data source.
	Params pulumi.StringMapOutput `pulumi:"params"`
	// The ID of the project in which the resource belongs.
	// If it is not provided, the provider project is used.
	Project pulumi.StringOutput `pulumi:"project"`
	// Data transfer schedule. If the data source does not support a custom
	// schedule, this should be empty. If it is empty, the default value for
	// the data source will be used. The specified times are in UTC. Examples
	// of valid format: 1st,3rd monday of month 15:30, every wed,fri of jan,
	// jun 13:15, and first sunday of quarter 00:00. See more explanation
	// about the format here:
	// https://cloud.google.com/appengine/docs/flexible/python/scheduling-jobs-with-cron-yaml#the_schedule_format
	// NOTE: the granularity should be at least 8 hours, or less frequent.
	Schedule pulumi.StringPtrOutput `pulumi:"schedule"`
	// Different parameters are configured primarily using the the `params` field on this
	// resource. This block contains the parameters which contain secrets or passwords so that they can be marked
	// sensitive and hidden from plan output. The name of the field, eg: secret_access_key, will be the key
	// in the `params` map in the api request.
	// Credentials may not be specified in both locations and will cause an error. Changing from one location
	// to a different credential configuration in the config will require an apply to update state.
	// Structure is documented below.
	SensitiveParams DataTransferConfigSensitiveParamsPtrOutput `pulumi:"sensitiveParams"`
	// Optional service account name. If this field is set, transfer config will
	// be created with this service account credentials. It requires that
	// requesting user calling this API has permissions to act as this service account.
	ServiceAccountName pulumi.StringPtrOutput `pulumi:"serviceAccountName"`
}

Represents a data transfer configuration. A transfer configuration contains all metadata needed to perform a data transfer.

To get more information about Config, see:

* [API documentation](https://cloud.google.com/bigquery/docs/reference/datatransfer/rest/v1/projects.locations.transferConfigs/create) * How-to Guides

> **Warning:** All arguments including `sensitive_params.secret_access_key` will be stored in the raw state as plain-text. [Read more about sensitive data in state](https://www.terraform.io/docs/state/sensitive-data.html).

## Example Usage

func GetDataTransferConfig

func GetDataTransferConfig(ctx *pulumi.Context,
	name string, id pulumi.IDInput, state *DataTransferConfigState, opts ...pulumi.ResourceOption) (*DataTransferConfig, error)

GetDataTransferConfig gets an existing DataTransferConfig resource's state with the given name, ID, and optional state properties that are used to uniquely qualify the lookup (nil if not required).

func NewDataTransferConfig

func NewDataTransferConfig(ctx *pulumi.Context,
	name string, args *DataTransferConfigArgs, opts ...pulumi.ResourceOption) (*DataTransferConfig, error)

NewDataTransferConfig registers a new resource with the given unique name, arguments, and options.

type DataTransferConfigArgs

type DataTransferConfigArgs struct {
	// The number of days to look back to automatically refresh the data.
	// For example, if dataRefreshWindowDays = 10, then every day BigQuery
	// reingests data for [today-10, today-1], rather than ingesting data for
	// just [today-1]. Only valid if the data source supports the feature.
	// Set the value to 0 to use the default value.
	DataRefreshWindowDays pulumi.IntPtrInput
	// The data source id. Cannot be changed once the transfer config is created.
	DataSourceId pulumi.StringInput
	// The BigQuery target dataset id.
	DestinationDatasetId pulumi.StringInput
	// When set to true, no runs are scheduled for a given transfer.
	Disabled pulumi.BoolPtrInput
	// The user specified display name for the transfer config.
	DisplayName pulumi.StringInput
	// The geographic location where the transfer config should reside.
	// Examples: US, EU, asia-northeast1. The default value is US.
	Location pulumi.StringPtrInput
	// Pub/Sub topic where notifications will be sent after transfer runs
	// associated with this transfer config finish.
	NotificationPubsubTopic pulumi.StringPtrInput
	// These parameters are specific to each data source.
	Params pulumi.StringMapInput
	// The ID of the project in which the resource belongs.
	// If it is not provided, the provider project is used.
	Project pulumi.StringPtrInput
	// Data transfer schedule. If the data source does not support a custom
	// schedule, this should be empty. If it is empty, the default value for
	// the data source will be used. The specified times are in UTC. Examples
	// of valid format: 1st,3rd monday of month 15:30, every wed,fri of jan,
	// jun 13:15, and first sunday of quarter 00:00. See more explanation
	// about the format here:
	// https://cloud.google.com/appengine/docs/flexible/python/scheduling-jobs-with-cron-yaml#the_schedule_format
	// NOTE: the granularity should be at least 8 hours, or less frequent.
	Schedule pulumi.StringPtrInput
	// Different parameters are configured primarily using the the `params` field on this
	// resource. This block contains the parameters which contain secrets or passwords so that they can be marked
	// sensitive and hidden from plan output. The name of the field, eg: secret_access_key, will be the key
	// in the `params` map in the api request.
	// Credentials may not be specified in both locations and will cause an error. Changing from one location
	// to a different credential configuration in the config will require an apply to update state.
	// Structure is documented below.
	SensitiveParams DataTransferConfigSensitiveParamsPtrInput
	// Optional service account name. If this field is set, transfer config will
	// be created with this service account credentials. It requires that
	// requesting user calling this API has permissions to act as this service account.
	ServiceAccountName pulumi.StringPtrInput
}

The set of arguments for constructing a DataTransferConfig resource.

func (DataTransferConfigArgs) ElementType

func (DataTransferConfigArgs) ElementType() reflect.Type

type DataTransferConfigSensitiveParams

type DataTransferConfigSensitiveParams struct {
	// The Secret Access Key of the AWS account transferring data from.
	// **Note**: This property is sensitive and will not be displayed in the plan.
	SecretAccessKey string `pulumi:"secretAccessKey"`
}

type DataTransferConfigSensitiveParamsArgs

type DataTransferConfigSensitiveParamsArgs struct {
	// The Secret Access Key of the AWS account transferring data from.
	// **Note**: This property is sensitive and will not be displayed in the plan.
	SecretAccessKey pulumi.StringInput `pulumi:"secretAccessKey"`
}

func (DataTransferConfigSensitiveParamsArgs) ElementType

func (DataTransferConfigSensitiveParamsArgs) ToDataTransferConfigSensitiveParamsOutput

func (i DataTransferConfigSensitiveParamsArgs) ToDataTransferConfigSensitiveParamsOutput() DataTransferConfigSensitiveParamsOutput

func (DataTransferConfigSensitiveParamsArgs) ToDataTransferConfigSensitiveParamsOutputWithContext

func (i DataTransferConfigSensitiveParamsArgs) ToDataTransferConfigSensitiveParamsOutputWithContext(ctx context.Context) DataTransferConfigSensitiveParamsOutput

func (DataTransferConfigSensitiveParamsArgs) ToDataTransferConfigSensitiveParamsPtrOutput

func (i DataTransferConfigSensitiveParamsArgs) ToDataTransferConfigSensitiveParamsPtrOutput() DataTransferConfigSensitiveParamsPtrOutput

func (DataTransferConfigSensitiveParamsArgs) ToDataTransferConfigSensitiveParamsPtrOutputWithContext

func (i DataTransferConfigSensitiveParamsArgs) ToDataTransferConfigSensitiveParamsPtrOutputWithContext(ctx context.Context) DataTransferConfigSensitiveParamsPtrOutput

type DataTransferConfigSensitiveParamsInput

type DataTransferConfigSensitiveParamsInput interface {
	pulumi.Input

	ToDataTransferConfigSensitiveParamsOutput() DataTransferConfigSensitiveParamsOutput
	ToDataTransferConfigSensitiveParamsOutputWithContext(context.Context) DataTransferConfigSensitiveParamsOutput
}

DataTransferConfigSensitiveParamsInput is an input type that accepts DataTransferConfigSensitiveParamsArgs and DataTransferConfigSensitiveParamsOutput values. You can construct a concrete instance of `DataTransferConfigSensitiveParamsInput` via:

DataTransferConfigSensitiveParamsArgs{...}

type DataTransferConfigSensitiveParamsOutput

type DataTransferConfigSensitiveParamsOutput struct{ *pulumi.OutputState }

func (DataTransferConfigSensitiveParamsOutput) ElementType

func (DataTransferConfigSensitiveParamsOutput) SecretAccessKey

The Secret Access Key of the AWS account transferring data from. **Note**: This property is sensitive and will not be displayed in the plan.

func (DataTransferConfigSensitiveParamsOutput) ToDataTransferConfigSensitiveParamsOutput

func (o DataTransferConfigSensitiveParamsOutput) ToDataTransferConfigSensitiveParamsOutput() DataTransferConfigSensitiveParamsOutput

func (DataTransferConfigSensitiveParamsOutput) ToDataTransferConfigSensitiveParamsOutputWithContext

func (o DataTransferConfigSensitiveParamsOutput) ToDataTransferConfigSensitiveParamsOutputWithContext(ctx context.Context) DataTransferConfigSensitiveParamsOutput

func (DataTransferConfigSensitiveParamsOutput) ToDataTransferConfigSensitiveParamsPtrOutput

func (o DataTransferConfigSensitiveParamsOutput) ToDataTransferConfigSensitiveParamsPtrOutput() DataTransferConfigSensitiveParamsPtrOutput

func (DataTransferConfigSensitiveParamsOutput) ToDataTransferConfigSensitiveParamsPtrOutputWithContext

func (o DataTransferConfigSensitiveParamsOutput) ToDataTransferConfigSensitiveParamsPtrOutputWithContext(ctx context.Context) DataTransferConfigSensitiveParamsPtrOutput

type DataTransferConfigSensitiveParamsPtrInput

type DataTransferConfigSensitiveParamsPtrInput interface {
	pulumi.Input

	ToDataTransferConfigSensitiveParamsPtrOutput() DataTransferConfigSensitiveParamsPtrOutput
	ToDataTransferConfigSensitiveParamsPtrOutputWithContext(context.Context) DataTransferConfigSensitiveParamsPtrOutput
}

DataTransferConfigSensitiveParamsPtrInput is an input type that accepts DataTransferConfigSensitiveParamsArgs, DataTransferConfigSensitiveParamsPtr and DataTransferConfigSensitiveParamsPtrOutput values. You can construct a concrete instance of `DataTransferConfigSensitiveParamsPtrInput` via:

        DataTransferConfigSensitiveParamsArgs{...}

or:

        nil

type DataTransferConfigSensitiveParamsPtrOutput

type DataTransferConfigSensitiveParamsPtrOutput struct{ *pulumi.OutputState }

func (DataTransferConfigSensitiveParamsPtrOutput) Elem

func (DataTransferConfigSensitiveParamsPtrOutput) ElementType

func (DataTransferConfigSensitiveParamsPtrOutput) SecretAccessKey

The Secret Access Key of the AWS account transferring data from. **Note**: This property is sensitive and will not be displayed in the plan.

func (DataTransferConfigSensitiveParamsPtrOutput) ToDataTransferConfigSensitiveParamsPtrOutput

func (o DataTransferConfigSensitiveParamsPtrOutput) ToDataTransferConfigSensitiveParamsPtrOutput() DataTransferConfigSensitiveParamsPtrOutput

func (DataTransferConfigSensitiveParamsPtrOutput) ToDataTransferConfigSensitiveParamsPtrOutputWithContext

func (o DataTransferConfigSensitiveParamsPtrOutput) ToDataTransferConfigSensitiveParamsPtrOutputWithContext(ctx context.Context) DataTransferConfigSensitiveParamsPtrOutput

type DataTransferConfigState

type DataTransferConfigState struct {
	// The number of days to look back to automatically refresh the data.
	// For example, if dataRefreshWindowDays = 10, then every day BigQuery
	// reingests data for [today-10, today-1], rather than ingesting data for
	// just [today-1]. Only valid if the data source supports the feature.
	// Set the value to 0 to use the default value.
	DataRefreshWindowDays pulumi.IntPtrInput
	// The data source id. Cannot be changed once the transfer config is created.
	DataSourceId pulumi.StringPtrInput
	// The BigQuery target dataset id.
	DestinationDatasetId pulumi.StringPtrInput
	// When set to true, no runs are scheduled for a given transfer.
	Disabled pulumi.BoolPtrInput
	// The user specified display name for the transfer config.
	DisplayName pulumi.StringPtrInput
	// The geographic location where the transfer config should reside.
	// Examples: US, EU, asia-northeast1. The default value is US.
	Location pulumi.StringPtrInput
	// The resource name of the transfer config. Transfer config names have the form
	// projects/{projectId}/locations/{location}/transferConfigs/{configId}. Where configId is usually a uuid, but this is not
	// required. The name is ignored when creating a transfer config.
	Name pulumi.StringPtrInput
	// Pub/Sub topic where notifications will be sent after transfer runs
	// associated with this transfer config finish.
	NotificationPubsubTopic pulumi.StringPtrInput
	// These parameters are specific to each data source.
	Params pulumi.StringMapInput
	// The ID of the project in which the resource belongs.
	// If it is not provided, the provider project is used.
	Project pulumi.StringPtrInput
	// Data transfer schedule. If the data source does not support a custom
	// schedule, this should be empty. If it is empty, the default value for
	// the data source will be used. The specified times are in UTC. Examples
	// of valid format: 1st,3rd monday of month 15:30, every wed,fri of jan,
	// jun 13:15, and first sunday of quarter 00:00. See more explanation
	// about the format here:
	// https://cloud.google.com/appengine/docs/flexible/python/scheduling-jobs-with-cron-yaml#the_schedule_format
	// NOTE: the granularity should be at least 8 hours, or less frequent.
	Schedule pulumi.StringPtrInput
	// Different parameters are configured primarily using the the `params` field on this
	// resource. This block contains the parameters which contain secrets or passwords so that they can be marked
	// sensitive and hidden from plan output. The name of the field, eg: secret_access_key, will be the key
	// in the `params` map in the api request.
	// Credentials may not be specified in both locations and will cause an error. Changing from one location
	// to a different credential configuration in the config will require an apply to update state.
	// Structure is documented below.
	SensitiveParams DataTransferConfigSensitiveParamsPtrInput
	// Optional service account name. If this field is set, transfer config will
	// be created with this service account credentials. It requires that
	// requesting user calling this API has permissions to act as this service account.
	ServiceAccountName pulumi.StringPtrInput
}

func (DataTransferConfigState) ElementType

func (DataTransferConfigState) ElementType() reflect.Type

type Dataset

type Dataset struct {
	pulumi.CustomResourceState

	// An array of objects that define dataset access for one or more entities.
	// Structure is documented below.
	Accesses DatasetAccessTypeArrayOutput `pulumi:"accesses"`
	// The time when this dataset was created, in milliseconds since the epoch.
	CreationTime pulumi.IntOutput `pulumi:"creationTime"`
	// The ID of the dataset containing this table.
	DatasetId pulumi.StringOutput `pulumi:"datasetId"`
	// The default encryption key for all tables in the dataset. Once this property is set,
	// all newly-created partitioned tables in the dataset will have encryption key set to
	// this value, unless table creation request (or query) overrides the key.
	// Structure is documented below.
	DefaultEncryptionConfiguration DatasetDefaultEncryptionConfigurationPtrOutput `pulumi:"defaultEncryptionConfiguration"`
	// The default partition expiration for all partitioned tables in
	// the dataset, in milliseconds.
	DefaultPartitionExpirationMs pulumi.IntPtrOutput `pulumi:"defaultPartitionExpirationMs"`
	// The default lifetime of all tables in the dataset, in milliseconds.
	// The minimum value is 3600000 milliseconds (one hour).
	DefaultTableExpirationMs pulumi.IntPtrOutput `pulumi:"defaultTableExpirationMs"`
	// If set to `true`, delete all the tables in the
	// dataset when destroying the resource; otherwise,
	// destroying the resource will fail if tables are present.
	DeleteContentsOnDestroy pulumi.BoolPtrOutput `pulumi:"deleteContentsOnDestroy"`
	// A user-friendly description of the dataset
	Description pulumi.StringPtrOutput `pulumi:"description"`
	// A hash of the resource.
	Etag pulumi.StringOutput `pulumi:"etag"`
	// A descriptive name for the dataset
	FriendlyName pulumi.StringPtrOutput `pulumi:"friendlyName"`
	// The labels associated with this dataset. You can use these to
	// organize and group your datasets
	Labels pulumi.StringMapOutput `pulumi:"labels"`
	// The date when this dataset or any of its tables was last modified, in milliseconds since the epoch.
	LastModifiedTime pulumi.IntOutput `pulumi:"lastModifiedTime"`
	// The geographic location where the dataset should reside.
	// See [official docs](https://cloud.google.com/bigquery/docs/dataset-locations).
	Location pulumi.StringPtrOutput `pulumi:"location"`
	// The ID of the project in which the resource belongs.
	// If it is not provided, the provider project is used.
	Project pulumi.StringOutput `pulumi:"project"`
	// The URI of the created resource.
	SelfLink pulumi.StringOutput `pulumi:"selfLink"`
}

Datasets allow you to organize and control access to your tables.

To get more information about Dataset, see:

* [API documentation](https://cloud.google.com/bigquery/docs/reference/rest/v2/datasets) * How-to Guides

## Example Usage

func GetDataset

func GetDataset(ctx *pulumi.Context,
	name string, id pulumi.IDInput, state *DatasetState, opts ...pulumi.ResourceOption) (*Dataset, error)

GetDataset gets an existing Dataset resource's state with the given name, ID, and optional state properties that are used to uniquely qualify the lookup (nil if not required).

func NewDataset

func NewDataset(ctx *pulumi.Context,
	name string, args *DatasetArgs, opts ...pulumi.ResourceOption) (*Dataset, error)

NewDataset registers a new resource with the given unique name, arguments, and options.

type DatasetAccess

type DatasetAccess struct {
	pulumi.CustomResourceState

	// If true, represents that that the iam_member in the config was translated to a different member type by the API, and is
	// stored in state as a different member type
	ApiUpdatedMember pulumi.BoolOutput `pulumi:"apiUpdatedMember"`
	// The ID of the dataset containing this table.
	DatasetId pulumi.StringOutput `pulumi:"datasetId"`
	// A domain to grant access to. Any users signed in with the
	// domain specified will be granted the specified access
	Domain pulumi.StringPtrOutput `pulumi:"domain"`
	// An email address of a Google Group to grant access to.
	GroupByEmail pulumi.StringPtrOutput `pulumi:"groupByEmail"`
	// Some other type of member that appears in the IAM Policy but isn't a user,
	// group, domain, or special group. For example: `allUsers`
	IamMember pulumi.StringPtrOutput `pulumi:"iamMember"`
	// The ID of the project in which the resource belongs.
	// If it is not provided, the provider project is used.
	Project pulumi.StringOutput `pulumi:"project"`
	// Describes the rights granted to the user specified by the other
	// member of the access object. Basic, predefined, and custom roles are
	// supported. Predefined roles that have equivalent basic roles are
	// swapped by the API to their basic counterparts, and will show a diff
	// post-create. See
	// [official docs](https://cloud.google.com/bigquery/docs/access-control).
	Role pulumi.StringPtrOutput `pulumi:"role"`
	// A special group to grant access to. Possible values include:
	SpecialGroup pulumi.StringPtrOutput `pulumi:"specialGroup"`
	// An email address of a user to grant access to. For example:
	// fred@example.com
	UserByEmail pulumi.StringPtrOutput `pulumi:"userByEmail"`
	// A view from a different dataset to grant access to. Queries
	// executed against that view will have read access to tables in
	// this dataset. The role field is not required when this field is
	// set. If that view is updated by any user, access to the view
	// needs to be granted again via an update operation.
	// Structure is documented below.
	View DatasetAccessViewPtrOutput `pulumi:"view"`
}

Gives dataset access for a single entity. This resource is intended to be used in cases where it is not possible to compile a full list of access blocks to include in a `bigquery.Dataset` resource, to enable them to be added separately.

> **Note:** If this resource is used alongside a `bigquery.Dataset` resource, the dataset resource must either have no defined `access` blocks or a `lifecycle` block with `ignoreChanges = [access]` so they don't fight over which accesses should be on the dataset.

To get more information about DatasetAccess, see:

* [API documentation](https://cloud.google.com/bigquery/docs/reference/rest/v2/datasets) * How-to Guides

## Example Usage

func GetDatasetAccess

func GetDatasetAccess(ctx *pulumi.Context,
	name string, id pulumi.IDInput, state *DatasetAccessState, opts ...pulumi.ResourceOption) (*DatasetAccess, error)

GetDatasetAccess gets an existing DatasetAccess resource's state with the given name, ID, and optional state properties that are used to uniquely qualify the lookup (nil if not required).

func NewDatasetAccess

func NewDatasetAccess(ctx *pulumi.Context,
	name string, args *DatasetAccessArgs, opts ...pulumi.ResourceOption) (*DatasetAccess, error)

NewDatasetAccess registers a new resource with the given unique name, arguments, and options.

type DatasetAccessArgs

type DatasetAccessArgs struct {
	// The ID of the dataset containing this table.
	DatasetId pulumi.StringInput
	// A domain to grant access to. Any users signed in with the
	// domain specified will be granted the specified access
	Domain pulumi.StringPtrInput
	// An email address of a Google Group to grant access to.
	GroupByEmail pulumi.StringPtrInput
	// Some other type of member that appears in the IAM Policy but isn't a user,
	// group, domain, or special group. For example: `allUsers`
	IamMember pulumi.StringPtrInput
	// The ID of the project in which the resource belongs.
	// If it is not provided, the provider project is used.
	Project pulumi.StringPtrInput
	// Describes the rights granted to the user specified by the other
	// member of the access object. Basic, predefined, and custom roles are
	// supported. Predefined roles that have equivalent basic roles are
	// swapped by the API to their basic counterparts, and will show a diff
	// post-create. See
	// [official docs](https://cloud.google.com/bigquery/docs/access-control).
	Role pulumi.StringPtrInput
	// A special group to grant access to. Possible values include:
	SpecialGroup pulumi.StringPtrInput
	// An email address of a user to grant access to. For example:
	// fred@example.com
	UserByEmail pulumi.StringPtrInput
	// A view from a different dataset to grant access to. Queries
	// executed against that view will have read access to tables in
	// this dataset. The role field is not required when this field is
	// set. If that view is updated by any user, access to the view
	// needs to be granted again via an update operation.
	// Structure is documented below.
	View DatasetAccessViewPtrInput
}

The set of arguments for constructing a DatasetAccess resource.

func (DatasetAccessArgs) ElementType

func (DatasetAccessArgs) ElementType() reflect.Type

type DatasetAccessState

type DatasetAccessState struct {
	// If true, represents that that the iam_member in the config was translated to a different member type by the API, and is
	// stored in state as a different member type
	ApiUpdatedMember pulumi.BoolPtrInput
	// The ID of the dataset containing this table.
	DatasetId pulumi.StringPtrInput
	// A domain to grant access to. Any users signed in with the
	// domain specified will be granted the specified access
	Domain pulumi.StringPtrInput
	// An email address of a Google Group to grant access to.
	GroupByEmail pulumi.StringPtrInput
	// Some other type of member that appears in the IAM Policy but isn't a user,
	// group, domain, or special group. For example: `allUsers`
	IamMember pulumi.StringPtrInput
	// The ID of the project in which the resource belongs.
	// If it is not provided, the provider project is used.
	Project pulumi.StringPtrInput
	// Describes the rights granted to the user specified by the other
	// member of the access object. Basic, predefined, and custom roles are
	// supported. Predefined roles that have equivalent basic roles are
	// swapped by the API to their basic counterparts, and will show a diff
	// post-create. See
	// [official docs](https://cloud.google.com/bigquery/docs/access-control).
	Role pulumi.StringPtrInput
	// A special group to grant access to. Possible values include:
	SpecialGroup pulumi.StringPtrInput
	// An email address of a user to grant access to. For example:
	// fred@example.com
	UserByEmail pulumi.StringPtrInput
	// A view from a different dataset to grant access to. Queries
	// executed against that view will have read access to tables in
	// this dataset. The role field is not required when this field is
	// set. If that view is updated by any user, access to the view
	// needs to be granted again via an update operation.
	// Structure is documented below.
	View DatasetAccessViewPtrInput
}

func (DatasetAccessState) ElementType

func (DatasetAccessState) ElementType() reflect.Type

type DatasetAccessType

type DatasetAccessType struct {
	// A domain to grant access to. Any users signed in with the
	// domain specified will be granted the specified access
	Domain *string `pulumi:"domain"`
	// An email address of a Google Group to grant access to.
	GroupByEmail *string `pulumi:"groupByEmail"`
	// Describes the rights granted to the user specified by the other
	// member of the access object. Basic, predefined, and custom roles
	// are supported. Predefined roles that have equivalent basic roles
	// are swapped by the API to their basic counterparts. See
	// [official docs](https://cloud.google.com/bigquery/docs/access-control).
	Role *string `pulumi:"role"`
	// A special group to grant access to. Possible values include:
	SpecialGroup *string `pulumi:"specialGroup"`
	// An email address of a user to grant access to. For example:
	// fred@example.com
	UserByEmail *string `pulumi:"userByEmail"`
	// A view from a different dataset to grant access to. Queries
	// executed against that view will have read access to tables in
	// this dataset. The role field is not required when this field is
	// set. If that view is updated by any user, access to the view
	// needs to be granted again via an update operation.
	// Structure is documented below.
	View *DatasetAccessView `pulumi:"view"`
}

type DatasetAccessTypeArgs

type DatasetAccessTypeArgs struct {
	// A domain to grant access to. Any users signed in with the
	// domain specified will be granted the specified access
	Domain pulumi.StringPtrInput `pulumi:"domain"`
	// An email address of a Google Group to grant access to.
	GroupByEmail pulumi.StringPtrInput `pulumi:"groupByEmail"`
	// Describes the rights granted to the user specified by the other
	// member of the access object. Basic, predefined, and custom roles
	// are supported. Predefined roles that have equivalent basic roles
	// are swapped by the API to their basic counterparts. See
	// [official docs](https://cloud.google.com/bigquery/docs/access-control).
	Role pulumi.StringPtrInput `pulumi:"role"`
	// A special group to grant access to. Possible values include:
	SpecialGroup pulumi.StringPtrInput `pulumi:"specialGroup"`
	// An email address of a user to grant access to. For example:
	// fred@example.com
	UserByEmail pulumi.StringPtrInput `pulumi:"userByEmail"`
	// A view from a different dataset to grant access to. Queries
	// executed against that view will have read access to tables in
	// this dataset. The role field is not required when this field is
	// set. If that view is updated by any user, access to the view
	// needs to be granted again via an update operation.
	// Structure is documented below.
	View DatasetAccessViewPtrInput `pulumi:"view"`
}

func (DatasetAccessTypeArgs) ElementType

func (DatasetAccessTypeArgs) ElementType() reflect.Type

func (DatasetAccessTypeArgs) ToDatasetAccessTypeOutput

func (i DatasetAccessTypeArgs) ToDatasetAccessTypeOutput() DatasetAccessTypeOutput

func (DatasetAccessTypeArgs) ToDatasetAccessTypeOutputWithContext

func (i DatasetAccessTypeArgs) ToDatasetAccessTypeOutputWithContext(ctx context.Context) DatasetAccessTypeOutput

type DatasetAccessTypeArray

type DatasetAccessTypeArray []DatasetAccessTypeInput

func (DatasetAccessTypeArray) ElementType

func (DatasetAccessTypeArray) ElementType() reflect.Type

func (DatasetAccessTypeArray) ToDatasetAccessTypeArrayOutput

func (i DatasetAccessTypeArray) ToDatasetAccessTypeArrayOutput() DatasetAccessTypeArrayOutput

func (DatasetAccessTypeArray) ToDatasetAccessTypeArrayOutputWithContext

func (i DatasetAccessTypeArray) ToDatasetAccessTypeArrayOutputWithContext(ctx context.Context) DatasetAccessTypeArrayOutput

type DatasetAccessTypeArrayInput

type DatasetAccessTypeArrayInput interface {
	pulumi.Input

	ToDatasetAccessTypeArrayOutput() DatasetAccessTypeArrayOutput
	ToDatasetAccessTypeArrayOutputWithContext(context.Context) DatasetAccessTypeArrayOutput
}

DatasetAccessTypeArrayInput is an input type that accepts DatasetAccessTypeArray and DatasetAccessTypeArrayOutput values. You can construct a concrete instance of `DatasetAccessTypeArrayInput` via:

DatasetAccessTypeArray{ DatasetAccessTypeArgs{...} }

type DatasetAccessTypeArrayOutput

type DatasetAccessTypeArrayOutput struct{ *pulumi.OutputState }

func (DatasetAccessTypeArrayOutput) ElementType

func (DatasetAccessTypeArrayOutput) Index

func (DatasetAccessTypeArrayOutput) ToDatasetAccessTypeArrayOutput

func (o DatasetAccessTypeArrayOutput) ToDatasetAccessTypeArrayOutput() DatasetAccessTypeArrayOutput

func (DatasetAccessTypeArrayOutput) ToDatasetAccessTypeArrayOutputWithContext

func (o DatasetAccessTypeArrayOutput) ToDatasetAccessTypeArrayOutputWithContext(ctx context.Context) DatasetAccessTypeArrayOutput

type DatasetAccessTypeInput

type DatasetAccessTypeInput interface {
	pulumi.Input

	ToDatasetAccessTypeOutput() DatasetAccessTypeOutput
	ToDatasetAccessTypeOutputWithContext(context.Context) DatasetAccessTypeOutput
}

DatasetAccessTypeInput is an input type that accepts DatasetAccessTypeArgs and DatasetAccessTypeOutput values. You can construct a concrete instance of `DatasetAccessTypeInput` via:

DatasetAccessTypeArgs{...}

type DatasetAccessTypeOutput

type DatasetAccessTypeOutput struct{ *pulumi.OutputState }

func (DatasetAccessTypeOutput) Domain

A domain to grant access to. Any users signed in with the domain specified will be granted the specified access

func (DatasetAccessTypeOutput) ElementType

func (DatasetAccessTypeOutput) ElementType() reflect.Type

func (DatasetAccessTypeOutput) GroupByEmail

An email address of a Google Group to grant access to.

func (DatasetAccessTypeOutput) Role

Describes the rights granted to the user specified by the other member of the access object. Basic, predefined, and custom roles are supported. Predefined roles that have equivalent basic roles are swapped by the API to their basic counterparts. See [official docs](https://cloud.google.com/bigquery/docs/access-control).

func (DatasetAccessTypeOutput) SpecialGroup

A special group to grant access to. Possible values include:

func (DatasetAccessTypeOutput) ToDatasetAccessTypeOutput

func (o DatasetAccessTypeOutput) ToDatasetAccessTypeOutput() DatasetAccessTypeOutput

func (DatasetAccessTypeOutput) ToDatasetAccessTypeOutputWithContext

func (o DatasetAccessTypeOutput) ToDatasetAccessTypeOutputWithContext(ctx context.Context) DatasetAccessTypeOutput

func (DatasetAccessTypeOutput) UserByEmail

An email address of a user to grant access to. For example: fred@example.com

func (DatasetAccessTypeOutput) View

A view from a different dataset to grant access to. Queries executed against that view will have read access to tables in this dataset. The role field is not required when this field is set. If that view is updated by any user, access to the view needs to be granted again via an update operation. Structure is documented below.

type DatasetAccessView

type DatasetAccessView struct {
	// The ID of the dataset containing this table.
	DatasetId string `pulumi:"datasetId"`
	// The ID of the project containing this table.
	ProjectId string `pulumi:"projectId"`
	// The ID of the table. The ID must contain only letters (a-z,
	// A-Z), numbers (0-9), or underscores (_). The maximum length
	// is 1,024 characters.
	TableId string `pulumi:"tableId"`
}

type DatasetAccessViewArgs

type DatasetAccessViewArgs struct {
	// The ID of the dataset containing this table.
	DatasetId pulumi.StringInput `pulumi:"datasetId"`
	// The ID of the project containing this table.
	ProjectId pulumi.StringInput `pulumi:"projectId"`
	// The ID of the table. The ID must contain only letters (a-z,
	// A-Z), numbers (0-9), or underscores (_). The maximum length
	// is 1,024 characters.
	TableId pulumi.StringInput `pulumi:"tableId"`
}

func (DatasetAccessViewArgs) ElementType

func (DatasetAccessViewArgs) ElementType() reflect.Type

func (DatasetAccessViewArgs) ToDatasetAccessViewOutput

func (i DatasetAccessViewArgs) ToDatasetAccessViewOutput() DatasetAccessViewOutput

func (DatasetAccessViewArgs) ToDatasetAccessViewOutputWithContext

func (i DatasetAccessViewArgs) ToDatasetAccessViewOutputWithContext(ctx context.Context) DatasetAccessViewOutput

func (DatasetAccessViewArgs) ToDatasetAccessViewPtrOutput

func (i DatasetAccessViewArgs) ToDatasetAccessViewPtrOutput() DatasetAccessViewPtrOutput

func (DatasetAccessViewArgs) ToDatasetAccessViewPtrOutputWithContext

func (i DatasetAccessViewArgs) ToDatasetAccessViewPtrOutputWithContext(ctx context.Context) DatasetAccessViewPtrOutput

type DatasetAccessViewInput

type DatasetAccessViewInput interface {
	pulumi.Input

	ToDatasetAccessViewOutput() DatasetAccessViewOutput
	ToDatasetAccessViewOutputWithContext(context.Context) DatasetAccessViewOutput
}

DatasetAccessViewInput is an input type that accepts DatasetAccessViewArgs and DatasetAccessViewOutput values. You can construct a concrete instance of `DatasetAccessViewInput` via:

DatasetAccessViewArgs{...}

type DatasetAccessViewOutput

type DatasetAccessViewOutput struct{ *pulumi.OutputState }

func (DatasetAccessViewOutput) DatasetId

The ID of the dataset containing this table.

func (DatasetAccessViewOutput) ElementType

func (DatasetAccessViewOutput) ElementType() reflect.Type

func (DatasetAccessViewOutput) ProjectId

The ID of the project containing this table.

func (DatasetAccessViewOutput) TableId

The ID of the table. The ID must contain only letters (a-z, A-Z), numbers (0-9), or underscores (_). The maximum length is 1,024 characters.

func (DatasetAccessViewOutput) ToDatasetAccessViewOutput

func (o DatasetAccessViewOutput) ToDatasetAccessViewOutput() DatasetAccessViewOutput

func (DatasetAccessViewOutput) ToDatasetAccessViewOutputWithContext

func (o DatasetAccessViewOutput) ToDatasetAccessViewOutputWithContext(ctx context.Context) DatasetAccessViewOutput

func (DatasetAccessViewOutput) ToDatasetAccessViewPtrOutput

func (o DatasetAccessViewOutput) ToDatasetAccessViewPtrOutput() DatasetAccessViewPtrOutput

func (DatasetAccessViewOutput) ToDatasetAccessViewPtrOutputWithContext

func (o DatasetAccessViewOutput) ToDatasetAccessViewPtrOutputWithContext(ctx context.Context) DatasetAccessViewPtrOutput

type DatasetAccessViewPtrInput

type DatasetAccessViewPtrInput interface {
	pulumi.Input

	ToDatasetAccessViewPtrOutput() DatasetAccessViewPtrOutput
	ToDatasetAccessViewPtrOutputWithContext(context.Context) DatasetAccessViewPtrOutput
}

DatasetAccessViewPtrInput is an input type that accepts DatasetAccessViewArgs, DatasetAccessViewPtr and DatasetAccessViewPtrOutput values. You can construct a concrete instance of `DatasetAccessViewPtrInput` via:

        DatasetAccessViewArgs{...}

or:

        nil

type DatasetAccessViewPtrOutput

type DatasetAccessViewPtrOutput struct{ *pulumi.OutputState }

func (DatasetAccessViewPtrOutput) DatasetId

The ID of the dataset containing this table.

func (DatasetAccessViewPtrOutput) Elem

func (DatasetAccessViewPtrOutput) ElementType

func (DatasetAccessViewPtrOutput) ElementType() reflect.Type

func (DatasetAccessViewPtrOutput) ProjectId

The ID of the project containing this table.

func (DatasetAccessViewPtrOutput) TableId

The ID of the table. The ID must contain only letters (a-z, A-Z), numbers (0-9), or underscores (_). The maximum length is 1,024 characters.

func (DatasetAccessViewPtrOutput) ToDatasetAccessViewPtrOutput

func (o DatasetAccessViewPtrOutput) ToDatasetAccessViewPtrOutput() DatasetAccessViewPtrOutput

func (DatasetAccessViewPtrOutput) ToDatasetAccessViewPtrOutputWithContext

func (o DatasetAccessViewPtrOutput) ToDatasetAccessViewPtrOutputWithContext(ctx context.Context) DatasetAccessViewPtrOutput

type DatasetArgs

type DatasetArgs struct {
	// An array of objects that define dataset access for one or more entities.
	// Structure is documented below.
	Accesses DatasetAccessTypeArrayInput
	// The ID of the dataset containing this table.
	DatasetId pulumi.StringInput
	// The default encryption key for all tables in the dataset. Once this property is set,
	// all newly-created partitioned tables in the dataset will have encryption key set to
	// this value, unless table creation request (or query) overrides the key.
	// Structure is documented below.
	DefaultEncryptionConfiguration DatasetDefaultEncryptionConfigurationPtrInput
	// The default partition expiration for all partitioned tables in
	// the dataset, in milliseconds.
	DefaultPartitionExpirationMs pulumi.IntPtrInput
	// The default lifetime of all tables in the dataset, in milliseconds.
	// The minimum value is 3600000 milliseconds (one hour).
	DefaultTableExpirationMs pulumi.IntPtrInput
	// If set to `true`, delete all the tables in the
	// dataset when destroying the resource; otherwise,
	// destroying the resource will fail if tables are present.
	DeleteContentsOnDestroy pulumi.BoolPtrInput
	// A user-friendly description of the dataset
	Description pulumi.StringPtrInput
	// A descriptive name for the dataset
	FriendlyName pulumi.StringPtrInput
	// The labels associated with this dataset. You can use these to
	// organize and group your datasets
	Labels pulumi.StringMapInput
	// The geographic location where the dataset should reside.
	// See [official docs](https://cloud.google.com/bigquery/docs/dataset-locations).
	Location pulumi.StringPtrInput
	// The ID of the project in which the resource belongs.
	// If it is not provided, the provider project is used.
	Project pulumi.StringPtrInput
}

The set of arguments for constructing a Dataset resource.

func (DatasetArgs) ElementType

func (DatasetArgs) ElementType() reflect.Type

type DatasetDefaultEncryptionConfiguration

type DatasetDefaultEncryptionConfiguration struct {
	// Describes the Cloud KMS encryption key that will be used to protect destination
	// BigQuery table. The BigQuery Service Account associated with your project requires
	// access to this encryption key.
	KmsKeyName string `pulumi:"kmsKeyName"`
}

type DatasetDefaultEncryptionConfigurationArgs

type DatasetDefaultEncryptionConfigurationArgs struct {
	// Describes the Cloud KMS encryption key that will be used to protect destination
	// BigQuery table. The BigQuery Service Account associated with your project requires
	// access to this encryption key.
	KmsKeyName pulumi.StringInput `pulumi:"kmsKeyName"`
}

func (DatasetDefaultEncryptionConfigurationArgs) ElementType

func (DatasetDefaultEncryptionConfigurationArgs) ToDatasetDefaultEncryptionConfigurationOutput

func (i DatasetDefaultEncryptionConfigurationArgs) ToDatasetDefaultEncryptionConfigurationOutput() DatasetDefaultEncryptionConfigurationOutput

func (DatasetDefaultEncryptionConfigurationArgs) ToDatasetDefaultEncryptionConfigurationOutputWithContext

func (i DatasetDefaultEncryptionConfigurationArgs) ToDatasetDefaultEncryptionConfigurationOutputWithContext(ctx context.Context) DatasetDefaultEncryptionConfigurationOutput

func (DatasetDefaultEncryptionConfigurationArgs) ToDatasetDefaultEncryptionConfigurationPtrOutput

func (i DatasetDefaultEncryptionConfigurationArgs) ToDatasetDefaultEncryptionConfigurationPtrOutput() DatasetDefaultEncryptionConfigurationPtrOutput

func (DatasetDefaultEncryptionConfigurationArgs) ToDatasetDefaultEncryptionConfigurationPtrOutputWithContext

func (i DatasetDefaultEncryptionConfigurationArgs) ToDatasetDefaultEncryptionConfigurationPtrOutputWithContext(ctx context.Context) DatasetDefaultEncryptionConfigurationPtrOutput

type DatasetDefaultEncryptionConfigurationInput

type DatasetDefaultEncryptionConfigurationInput interface {
	pulumi.Input

	ToDatasetDefaultEncryptionConfigurationOutput() DatasetDefaultEncryptionConfigurationOutput
	ToDatasetDefaultEncryptionConfigurationOutputWithContext(context.Context) DatasetDefaultEncryptionConfigurationOutput
}

DatasetDefaultEncryptionConfigurationInput is an input type that accepts DatasetDefaultEncryptionConfigurationArgs and DatasetDefaultEncryptionConfigurationOutput values. You can construct a concrete instance of `DatasetDefaultEncryptionConfigurationInput` via:

DatasetDefaultEncryptionConfigurationArgs{...}

type DatasetDefaultEncryptionConfigurationOutput

type DatasetDefaultEncryptionConfigurationOutput struct{ *pulumi.OutputState }

func (DatasetDefaultEncryptionConfigurationOutput) ElementType

func (DatasetDefaultEncryptionConfigurationOutput) KmsKeyName

Describes the Cloud KMS encryption key that will be used to protect destination BigQuery table. The BigQuery Service Account associated with your project requires access to this encryption key.

func (DatasetDefaultEncryptionConfigurationOutput) ToDatasetDefaultEncryptionConfigurationOutput

func (o DatasetDefaultEncryptionConfigurationOutput) ToDatasetDefaultEncryptionConfigurationOutput() DatasetDefaultEncryptionConfigurationOutput

func (DatasetDefaultEncryptionConfigurationOutput) ToDatasetDefaultEncryptionConfigurationOutputWithContext

func (o DatasetDefaultEncryptionConfigurationOutput) ToDatasetDefaultEncryptionConfigurationOutputWithContext(ctx context.Context) DatasetDefaultEncryptionConfigurationOutput

func (DatasetDefaultEncryptionConfigurationOutput) ToDatasetDefaultEncryptionConfigurationPtrOutput

func (o DatasetDefaultEncryptionConfigurationOutput) ToDatasetDefaultEncryptionConfigurationPtrOutput() DatasetDefaultEncryptionConfigurationPtrOutput

func (DatasetDefaultEncryptionConfigurationOutput) ToDatasetDefaultEncryptionConfigurationPtrOutputWithContext

func (o DatasetDefaultEncryptionConfigurationOutput) ToDatasetDefaultEncryptionConfigurationPtrOutputWithContext(ctx context.Context) DatasetDefaultEncryptionConfigurationPtrOutput

type DatasetDefaultEncryptionConfigurationPtrInput

type DatasetDefaultEncryptionConfigurationPtrInput interface {
	pulumi.Input

	ToDatasetDefaultEncryptionConfigurationPtrOutput() DatasetDefaultEncryptionConfigurationPtrOutput
	ToDatasetDefaultEncryptionConfigurationPtrOutputWithContext(context.Context) DatasetDefaultEncryptionConfigurationPtrOutput
}

DatasetDefaultEncryptionConfigurationPtrInput is an input type that accepts DatasetDefaultEncryptionConfigurationArgs, DatasetDefaultEncryptionConfigurationPtr and DatasetDefaultEncryptionConfigurationPtrOutput values. You can construct a concrete instance of `DatasetDefaultEncryptionConfigurationPtrInput` via:

        DatasetDefaultEncryptionConfigurationArgs{...}

or:

        nil

type DatasetDefaultEncryptionConfigurationPtrOutput

type DatasetDefaultEncryptionConfigurationPtrOutput struct{ *pulumi.OutputState }

func (DatasetDefaultEncryptionConfigurationPtrOutput) Elem

func (DatasetDefaultEncryptionConfigurationPtrOutput) ElementType

func (DatasetDefaultEncryptionConfigurationPtrOutput) KmsKeyName

Describes the Cloud KMS encryption key that will be used to protect destination BigQuery table. The BigQuery Service Account associated with your project requires access to this encryption key.

func (DatasetDefaultEncryptionConfigurationPtrOutput) ToDatasetDefaultEncryptionConfigurationPtrOutput

func (o DatasetDefaultEncryptionConfigurationPtrOutput) ToDatasetDefaultEncryptionConfigurationPtrOutput() DatasetDefaultEncryptionConfigurationPtrOutput

func (DatasetDefaultEncryptionConfigurationPtrOutput) ToDatasetDefaultEncryptionConfigurationPtrOutputWithContext

func (o DatasetDefaultEncryptionConfigurationPtrOutput) ToDatasetDefaultEncryptionConfigurationPtrOutputWithContext(ctx context.Context) DatasetDefaultEncryptionConfigurationPtrOutput

type DatasetIamBinding

type DatasetIamBinding struct {
	pulumi.CustomResourceState

	Condition DatasetIamBindingConditionPtrOutput `pulumi:"condition"`
	// The dataset ID.
	DatasetId pulumi.StringOutput `pulumi:"datasetId"`
	// (Computed) The etag of the dataset's IAM policy.
	Etag    pulumi.StringOutput      `pulumi:"etag"`
	Members pulumi.StringArrayOutput `pulumi:"members"`
	Project pulumi.StringOutput      `pulumi:"project"`
	// The role that should be applied. Only one
	// `bigquery.DatasetIamBinding` can be used per role. Note that custom roles must be of the format
	// `[projects|organizations]/{parent-name}/roles/{role-name}`.
	Role pulumi.StringOutput `pulumi:"role"`
}

Three different resources help you manage your IAM policy for BigQuery dataset. Each of these resources serves a different use case:

* `bigquery.DatasetIamPolicy`: Authoritative. Sets the IAM policy for the dataset and replaces any existing policy already attached. * `bigquery.DatasetIamBinding`: Authoritative for a given role. Updates the IAM policy to grant a role to a list of members. Other roles within the IAM policy for the dataset are preserved. * `bigquery.DatasetIamMember`: Non-authoritative. Updates the IAM policy to grant a role to a new member. Other members for the role for the dataset are preserved.

These resources are intended to convert the permissions system for BigQuery datasets to the standard IAM interface. For advanced usages, including [creating authorized views](https://cloud.google.com/bigquery/docs/share-access-views), please use either `bigquery.DatasetAccess` or the `access` field on `bigquery.Dataset`.

> **Note:** These resources **cannot** be used with `bigquery.DatasetAccess` resources or the `access` field on `bigquery.Dataset` or they will fight over what the policy should be.

> **Note:** Using any of these resources will remove any authorized view permissions from the dataset. To assign and preserve authorized view permissions use the `bigquery.DatasetAccess` instead.

> **Note:** Legacy BigQuery roles `OWNER` `WRITER` and `READER` **cannot** be used with any of these IAM resources. Instead use the full role form of: `roles/bigquery.dataOwner` `roles/bigquery.dataEditor` and `roles/bigquery.dataViewer`.

> **Note:** `bigquery.DatasetIamPolicy` **cannot** be used in conjunction with `bigquery.DatasetIamBinding` and `bigquery.DatasetIamMember` or they will fight over what your policy should be.

> **Note:** `bigquery.DatasetIamBinding` resources **can be** used in conjunction with `bigquery.DatasetIamMember` resources **only if** they do not grant privilege to the same role.

func GetDatasetIamBinding

func GetDatasetIamBinding(ctx *pulumi.Context,
	name string, id pulumi.IDInput, state *DatasetIamBindingState, opts ...pulumi.ResourceOption) (*DatasetIamBinding, error)

GetDatasetIamBinding gets an existing DatasetIamBinding resource's state with the given name, ID, and optional state properties that are used to uniquely qualify the lookup (nil if not required).

func NewDatasetIamBinding

func NewDatasetIamBinding(ctx *pulumi.Context,
	name string, args *DatasetIamBindingArgs, opts ...pulumi.ResourceOption) (*DatasetIamBinding, error)

NewDatasetIamBinding registers a new resource with the given unique name, arguments, and options.

type DatasetIamBindingArgs

type DatasetIamBindingArgs struct {
	Condition DatasetIamBindingConditionPtrInput
	// The dataset ID.
	DatasetId pulumi.StringInput
	Members   pulumi.StringArrayInput
	Project   pulumi.StringPtrInput
	// The role that should be applied. Only one
	// `bigquery.DatasetIamBinding` can be used per role. Note that custom roles must be of the format
	// `[projects|organizations]/{parent-name}/roles/{role-name}`.
	Role pulumi.StringInput
}

The set of arguments for constructing a DatasetIamBinding resource.

func (DatasetIamBindingArgs) ElementType

func (DatasetIamBindingArgs) ElementType() reflect.Type

type DatasetIamBindingCondition

type DatasetIamBindingCondition struct {
	Description *string `pulumi:"description"`
	Expression  string  `pulumi:"expression"`
	Title       string  `pulumi:"title"`
}

type DatasetIamBindingConditionArgs

type DatasetIamBindingConditionArgs struct {
	Description pulumi.StringPtrInput `pulumi:"description"`
	Expression  pulumi.StringInput    `pulumi:"expression"`
	Title       pulumi.StringInput    `pulumi:"title"`
}

func (DatasetIamBindingConditionArgs) ElementType

func (DatasetIamBindingConditionArgs) ToDatasetIamBindingConditionOutput

func (i DatasetIamBindingConditionArgs) ToDatasetIamBindingConditionOutput() DatasetIamBindingConditionOutput

func (DatasetIamBindingConditionArgs) ToDatasetIamBindingConditionOutputWithContext

func (i DatasetIamBindingConditionArgs) ToDatasetIamBindingConditionOutputWithContext(ctx context.Context) DatasetIamBindingConditionOutput

func (DatasetIamBindingConditionArgs) ToDatasetIamBindingConditionPtrOutput

func (i DatasetIamBindingConditionArgs) ToDatasetIamBindingConditionPtrOutput() DatasetIamBindingConditionPtrOutput

func (DatasetIamBindingConditionArgs) ToDatasetIamBindingConditionPtrOutputWithContext

func (i DatasetIamBindingConditionArgs) ToDatasetIamBindingConditionPtrOutputWithContext(ctx context.Context) DatasetIamBindingConditionPtrOutput

type DatasetIamBindingConditionInput

type DatasetIamBindingConditionInput interface {
	pulumi.Input

	ToDatasetIamBindingConditionOutput() DatasetIamBindingConditionOutput
	ToDatasetIamBindingConditionOutputWithContext(context.Context) DatasetIamBindingConditionOutput
}

DatasetIamBindingConditionInput is an input type that accepts DatasetIamBindingConditionArgs and DatasetIamBindingConditionOutput values. You can construct a concrete instance of `DatasetIamBindingConditionInput` via:

DatasetIamBindingConditionArgs{...}

type DatasetIamBindingConditionOutput

type DatasetIamBindingConditionOutput struct{ *pulumi.OutputState }

func (DatasetIamBindingConditionOutput) Description

func (DatasetIamBindingConditionOutput) ElementType

func (DatasetIamBindingConditionOutput) Expression

func (DatasetIamBindingConditionOutput) Title

func (DatasetIamBindingConditionOutput) ToDatasetIamBindingConditionOutput

func (o DatasetIamBindingConditionOutput) ToDatasetIamBindingConditionOutput() DatasetIamBindingConditionOutput

func (DatasetIamBindingConditionOutput) ToDatasetIamBindingConditionOutputWithContext

func (o DatasetIamBindingConditionOutput) ToDatasetIamBindingConditionOutputWithContext(ctx context.Context) DatasetIamBindingConditionOutput

func (DatasetIamBindingConditionOutput) ToDatasetIamBindingConditionPtrOutput

func (o DatasetIamBindingConditionOutput) ToDatasetIamBindingConditionPtrOutput() DatasetIamBindingConditionPtrOutput

func (DatasetIamBindingConditionOutput) ToDatasetIamBindingConditionPtrOutputWithContext

func (o DatasetIamBindingConditionOutput) ToDatasetIamBindingConditionPtrOutputWithContext(ctx context.Context) DatasetIamBindingConditionPtrOutput

type DatasetIamBindingConditionPtrInput

type DatasetIamBindingConditionPtrInput interface {
	pulumi.Input

	ToDatasetIamBindingConditionPtrOutput() DatasetIamBindingConditionPtrOutput
	ToDatasetIamBindingConditionPtrOutputWithContext(context.Context) DatasetIamBindingConditionPtrOutput
}

DatasetIamBindingConditionPtrInput is an input type that accepts DatasetIamBindingConditionArgs, DatasetIamBindingConditionPtr and DatasetIamBindingConditionPtrOutput values. You can construct a concrete instance of `DatasetIamBindingConditionPtrInput` via:

        DatasetIamBindingConditionArgs{...}

or:

        nil

type DatasetIamBindingConditionPtrOutput

type DatasetIamBindingConditionPtrOutput struct{ *pulumi.OutputState }

func (DatasetIamBindingConditionPtrOutput) Description

func (DatasetIamBindingConditionPtrOutput) Elem

func (DatasetIamBindingConditionPtrOutput) ElementType

func (DatasetIamBindingConditionPtrOutput) Expression

func (DatasetIamBindingConditionPtrOutput) Title

func (DatasetIamBindingConditionPtrOutput) ToDatasetIamBindingConditionPtrOutput

func (o DatasetIamBindingConditionPtrOutput) ToDatasetIamBindingConditionPtrOutput() DatasetIamBindingConditionPtrOutput

func (DatasetIamBindingConditionPtrOutput) ToDatasetIamBindingConditionPtrOutputWithContext

func (o DatasetIamBindingConditionPtrOutput) ToDatasetIamBindingConditionPtrOutputWithContext(ctx context.Context) DatasetIamBindingConditionPtrOutput

type DatasetIamBindingState

type DatasetIamBindingState struct {
	Condition DatasetIamBindingConditionPtrInput
	// The dataset ID.
	DatasetId pulumi.StringPtrInput
	// (Computed) The etag of the dataset's IAM policy.
	Etag    pulumi.StringPtrInput
	Members pulumi.StringArrayInput
	Project pulumi.StringPtrInput
	// The role that should be applied. Only one
	// `bigquery.DatasetIamBinding` can be used per role. Note that custom roles must be of the format
	// `[projects|organizations]/{parent-name}/roles/{role-name}`.
	Role pulumi.StringPtrInput
}

func (DatasetIamBindingState) ElementType

func (DatasetIamBindingState) ElementType() reflect.Type

type DatasetIamMember

type DatasetIamMember struct {
	pulumi.CustomResourceState

	Condition DatasetIamMemberConditionPtrOutput `pulumi:"condition"`
	// The dataset ID.
	DatasetId pulumi.StringOutput `pulumi:"datasetId"`
	// (Computed) The etag of the dataset's IAM policy.
	Etag    pulumi.StringOutput `pulumi:"etag"`
	Member  pulumi.StringOutput `pulumi:"member"`
	Project pulumi.StringOutput `pulumi:"project"`
	// The role that should be applied. Only one
	// `bigquery.DatasetIamBinding` can be used per role. Note that custom roles must be of the format
	// `[projects|organizations]/{parent-name}/roles/{role-name}`.
	Role pulumi.StringOutput `pulumi:"role"`
}

Three different resources help you manage your IAM policy for BigQuery dataset. Each of these resources serves a different use case:

* `bigquery.DatasetIamPolicy`: Authoritative. Sets the IAM policy for the dataset and replaces any existing policy already attached. * `bigquery.DatasetIamBinding`: Authoritative for a given role. Updates the IAM policy to grant a role to a list of members. Other roles within the IAM policy for the dataset are preserved. * `bigquery.DatasetIamMember`: Non-authoritative. Updates the IAM policy to grant a role to a new member. Other members for the role for the dataset are preserved.

These resources are intended to convert the permissions system for BigQuery datasets to the standard IAM interface. For advanced usages, including [creating authorized views](https://cloud.google.com/bigquery/docs/share-access-views), please use either `bigquery.DatasetAccess` or the `access` field on `bigquery.Dataset`.

> **Note:** These resources **cannot** be used with `bigquery.DatasetAccess` resources or the `access` field on `bigquery.Dataset` or they will fight over what the policy should be.

> **Note:** Using any of these resources will remove any authorized view permissions from the dataset. To assign and preserve authorized view permissions use the `bigquery.DatasetAccess` instead.

> **Note:** Legacy BigQuery roles `OWNER` `WRITER` and `READER` **cannot** be used with any of these IAM resources. Instead use the full role form of: `roles/bigquery.dataOwner` `roles/bigquery.dataEditor` and `roles/bigquery.dataViewer`.

> **Note:** `bigquery.DatasetIamPolicy` **cannot** be used in conjunction with `bigquery.DatasetIamBinding` and `bigquery.DatasetIamMember` or they will fight over what your policy should be.

> **Note:** `bigquery.DatasetIamBinding` resources **can be** used in conjunction with `bigquery.DatasetIamMember` resources **only if** they do not grant privilege to the same role.

func GetDatasetIamMember

func GetDatasetIamMember(ctx *pulumi.Context,
	name string, id pulumi.IDInput, state *DatasetIamMemberState, opts ...pulumi.ResourceOption) (*DatasetIamMember, error)

GetDatasetIamMember gets an existing DatasetIamMember resource's state with the given name, ID, and optional state properties that are used to uniquely qualify the lookup (nil if not required).

func NewDatasetIamMember

func NewDatasetIamMember(ctx *pulumi.Context,
	name string, args *DatasetIamMemberArgs, opts ...pulumi.ResourceOption) (*DatasetIamMember, error)

NewDatasetIamMember registers a new resource with the given unique name, arguments, and options.

type DatasetIamMemberArgs

type DatasetIamMemberArgs struct {
	Condition DatasetIamMemberConditionPtrInput
	// The dataset ID.
	DatasetId pulumi.StringInput
	Member    pulumi.StringInput
	Project   pulumi.StringPtrInput
	// The role that should be applied. Only one
	// `bigquery.DatasetIamBinding` can be used per role. Note that custom roles must be of the format
	// `[projects|organizations]/{parent-name}/roles/{role-name}`.
	Role pulumi.StringInput
}

The set of arguments for constructing a DatasetIamMember resource.

func (DatasetIamMemberArgs) ElementType

func (DatasetIamMemberArgs) ElementType() reflect.Type

type DatasetIamMemberCondition

type DatasetIamMemberCondition struct {
	Description *string `pulumi:"description"`
	Expression  string  `pulumi:"expression"`
	Title       string  `pulumi:"title"`
}

type DatasetIamMemberConditionArgs

type DatasetIamMemberConditionArgs struct {
	Description pulumi.StringPtrInput `pulumi:"description"`
	Expression  pulumi.StringInput    `pulumi:"expression"`
	Title       pulumi.StringInput    `pulumi:"title"`
}

func (DatasetIamMemberConditionArgs) ElementType

func (DatasetIamMemberConditionArgs) ToDatasetIamMemberConditionOutput

func (i DatasetIamMemberConditionArgs) ToDatasetIamMemberConditionOutput() DatasetIamMemberConditionOutput

func (DatasetIamMemberConditionArgs) ToDatasetIamMemberConditionOutputWithContext

func (i DatasetIamMemberConditionArgs) ToDatasetIamMemberConditionOutputWithContext(ctx context.Context) DatasetIamMemberConditionOutput

func (DatasetIamMemberConditionArgs) ToDatasetIamMemberConditionPtrOutput

func (i DatasetIamMemberConditionArgs) ToDatasetIamMemberConditionPtrOutput() DatasetIamMemberConditionPtrOutput

func (DatasetIamMemberConditionArgs) ToDatasetIamMemberConditionPtrOutputWithContext

func (i DatasetIamMemberConditionArgs) ToDatasetIamMemberConditionPtrOutputWithContext(ctx context.Context) DatasetIamMemberConditionPtrOutput

type DatasetIamMemberConditionInput

type DatasetIamMemberConditionInput interface {
	pulumi.Input

	ToDatasetIamMemberConditionOutput() DatasetIamMemberConditionOutput
	ToDatasetIamMemberConditionOutputWithContext(context.Context) DatasetIamMemberConditionOutput
}

DatasetIamMemberConditionInput is an input type that accepts DatasetIamMemberConditionArgs and DatasetIamMemberConditionOutput values. You can construct a concrete instance of `DatasetIamMemberConditionInput` via:

DatasetIamMemberConditionArgs{...}

type DatasetIamMemberConditionOutput

type DatasetIamMemberConditionOutput struct{ *pulumi.OutputState }

func (DatasetIamMemberConditionOutput) Description

func (DatasetIamMemberConditionOutput) ElementType

func (DatasetIamMemberConditionOutput) Expression

func (DatasetIamMemberConditionOutput) Title

func (DatasetIamMemberConditionOutput) ToDatasetIamMemberConditionOutput

func (o DatasetIamMemberConditionOutput) ToDatasetIamMemberConditionOutput() DatasetIamMemberConditionOutput

func (DatasetIamMemberConditionOutput) ToDatasetIamMemberConditionOutputWithContext

func (o DatasetIamMemberConditionOutput) ToDatasetIamMemberConditionOutputWithContext(ctx context.Context) DatasetIamMemberConditionOutput

func (DatasetIamMemberConditionOutput) ToDatasetIamMemberConditionPtrOutput

func (o DatasetIamMemberConditionOutput) ToDatasetIamMemberConditionPtrOutput() DatasetIamMemberConditionPtrOutput

func (DatasetIamMemberConditionOutput) ToDatasetIamMemberConditionPtrOutputWithContext

func (o DatasetIamMemberConditionOutput) ToDatasetIamMemberConditionPtrOutputWithContext(ctx context.Context) DatasetIamMemberConditionPtrOutput

type DatasetIamMemberConditionPtrInput

type DatasetIamMemberConditionPtrInput interface {
	pulumi.Input

	ToDatasetIamMemberConditionPtrOutput() DatasetIamMemberConditionPtrOutput
	ToDatasetIamMemberConditionPtrOutputWithContext(context.Context) DatasetIamMemberConditionPtrOutput
}

DatasetIamMemberConditionPtrInput is an input type that accepts DatasetIamMemberConditionArgs, DatasetIamMemberConditionPtr and DatasetIamMemberConditionPtrOutput values. You can construct a concrete instance of `DatasetIamMemberConditionPtrInput` via:

        DatasetIamMemberConditionArgs{...}

or:

        nil

type DatasetIamMemberConditionPtrOutput

type DatasetIamMemberConditionPtrOutput struct{ *pulumi.OutputState }

func (DatasetIamMemberConditionPtrOutput) Description

func (DatasetIamMemberConditionPtrOutput) Elem

func (DatasetIamMemberConditionPtrOutput) ElementType

func (DatasetIamMemberConditionPtrOutput) Expression

func (DatasetIamMemberConditionPtrOutput) Title

func (DatasetIamMemberConditionPtrOutput) ToDatasetIamMemberConditionPtrOutput

func (o DatasetIamMemberConditionPtrOutput) ToDatasetIamMemberConditionPtrOutput() DatasetIamMemberConditionPtrOutput

func (DatasetIamMemberConditionPtrOutput) ToDatasetIamMemberConditionPtrOutputWithContext

func (o DatasetIamMemberConditionPtrOutput) ToDatasetIamMemberConditionPtrOutputWithContext(ctx context.Context) DatasetIamMemberConditionPtrOutput

type DatasetIamMemberState

type DatasetIamMemberState struct {
	Condition DatasetIamMemberConditionPtrInput
	// The dataset ID.
	DatasetId pulumi.StringPtrInput
	// (Computed) The etag of the dataset's IAM policy.
	Etag    pulumi.StringPtrInput
	Member  pulumi.StringPtrInput
	Project pulumi.StringPtrInput
	// The role that should be applied. Only one
	// `bigquery.DatasetIamBinding` can be used per role. Note that custom roles must be of the format
	// `[projects|organizations]/{parent-name}/roles/{role-name}`.
	Role pulumi.StringPtrInput
}

func (DatasetIamMemberState) ElementType

func (DatasetIamMemberState) ElementType() reflect.Type

type DatasetIamPolicy

type DatasetIamPolicy struct {
	pulumi.CustomResourceState

	// The dataset ID.
	DatasetId pulumi.StringOutput `pulumi:"datasetId"`
	// (Computed) The etag of the dataset's IAM policy.
	Etag pulumi.StringOutput `pulumi:"etag"`
	// The policy data generated by
	// a `organizations.getIAMPolicy` data source.
	PolicyData pulumi.StringOutput `pulumi:"policyData"`
	Project    pulumi.StringOutput `pulumi:"project"`
}

Three different resources help you manage your IAM policy for BigQuery dataset. Each of these resources serves a different use case:

* `bigquery.DatasetIamPolicy`: Authoritative. Sets the IAM policy for the dataset and replaces any existing policy already attached. * `bigquery.DatasetIamBinding`: Authoritative for a given role. Updates the IAM policy to grant a role to a list of members. Other roles within the IAM policy for the dataset are preserved. * `bigquery.DatasetIamMember`: Non-authoritative. Updates the IAM policy to grant a role to a new member. Other members for the role for the dataset are preserved.

These resources are intended to convert the permissions system for BigQuery datasets to the standard IAM interface. For advanced usages, including [creating authorized views](https://cloud.google.com/bigquery/docs/share-access-views), please use either `bigquery.DatasetAccess` or the `access` field on `bigquery.Dataset`.

> **Note:** These resources **cannot** be used with `bigquery.DatasetAccess` resources or the `access` field on `bigquery.Dataset` or they will fight over what the policy should be.

> **Note:** Using any of these resources will remove any authorized view permissions from the dataset. To assign and preserve authorized view permissions use the `bigquery.DatasetAccess` instead.

> **Note:** Legacy BigQuery roles `OWNER` `WRITER` and `READER` **cannot** be used with any of these IAM resources. Instead use the full role form of: `roles/bigquery.dataOwner` `roles/bigquery.dataEditor` and `roles/bigquery.dataViewer`.

> **Note:** `bigquery.DatasetIamPolicy` **cannot** be used in conjunction with `bigquery.DatasetIamBinding` and `bigquery.DatasetIamMember` or they will fight over what your policy should be.

> **Note:** `bigquery.DatasetIamBinding` resources **can be** used in conjunction with `bigquery.DatasetIamMember` resources **only if** they do not grant privilege to the same role.

func GetDatasetIamPolicy

func GetDatasetIamPolicy(ctx *pulumi.Context,
	name string, id pulumi.IDInput, state *DatasetIamPolicyState, opts ...pulumi.ResourceOption) (*DatasetIamPolicy, error)

GetDatasetIamPolicy gets an existing DatasetIamPolicy resource's state with the given name, ID, and optional state properties that are used to uniquely qualify the lookup (nil if not required).

func NewDatasetIamPolicy

func NewDatasetIamPolicy(ctx *pulumi.Context,
	name string, args *DatasetIamPolicyArgs, opts ...pulumi.ResourceOption) (*DatasetIamPolicy, error)

NewDatasetIamPolicy registers a new resource with the given unique name, arguments, and options.

type DatasetIamPolicyArgs

type DatasetIamPolicyArgs struct {
	// The dataset ID.
	DatasetId pulumi.StringInput
	// The policy data generated by
	// a `organizations.getIAMPolicy` data source.
	PolicyData pulumi.StringInput
	Project    pulumi.StringPtrInput
}

The set of arguments for constructing a DatasetIamPolicy resource.

func (DatasetIamPolicyArgs) ElementType

func (DatasetIamPolicyArgs) ElementType() reflect.Type

type DatasetIamPolicyState

type DatasetIamPolicyState struct {
	// The dataset ID.
	DatasetId pulumi.StringPtrInput
	// (Computed) The etag of the dataset's IAM policy.
	Etag pulumi.StringPtrInput
	// The policy data generated by
	// a `organizations.getIAMPolicy` data source.
	PolicyData pulumi.StringPtrInput
	Project    pulumi.StringPtrInput
}

func (DatasetIamPolicyState) ElementType

func (DatasetIamPolicyState) ElementType() reflect.Type

type DatasetState

type DatasetState struct {
	// An array of objects that define dataset access for one or more entities.
	// Structure is documented below.
	Accesses DatasetAccessTypeArrayInput
	// The time when this dataset was created, in milliseconds since the epoch.
	CreationTime pulumi.IntPtrInput
	// The ID of the dataset containing this table.
	DatasetId pulumi.StringPtrInput
	// The default encryption key for all tables in the dataset. Once this property is set,
	// all newly-created partitioned tables in the dataset will have encryption key set to
	// this value, unless table creation request (or query) overrides the key.
	// Structure is documented below.
	DefaultEncryptionConfiguration DatasetDefaultEncryptionConfigurationPtrInput
	// The default partition expiration for all partitioned tables in
	// the dataset, in milliseconds.
	DefaultPartitionExpirationMs pulumi.IntPtrInput
	// The default lifetime of all tables in the dataset, in milliseconds.
	// The minimum value is 3600000 milliseconds (one hour).
	DefaultTableExpirationMs pulumi.IntPtrInput
	// If set to `true`, delete all the tables in the
	// dataset when destroying the resource; otherwise,
	// destroying the resource will fail if tables are present.
	DeleteContentsOnDestroy pulumi.BoolPtrInput
	// A user-friendly description of the dataset
	Description pulumi.StringPtrInput
	// A hash of the resource.
	Etag pulumi.StringPtrInput
	// A descriptive name for the dataset
	FriendlyName pulumi.StringPtrInput
	// The labels associated with this dataset. You can use these to
	// organize and group your datasets
	Labels pulumi.StringMapInput
	// The date when this dataset or any of its tables was last modified, in milliseconds since the epoch.
	LastModifiedTime pulumi.IntPtrInput
	// The geographic location where the dataset should reside.
	// See [official docs](https://cloud.google.com/bigquery/docs/dataset-locations).
	Location pulumi.StringPtrInput
	// The ID of the project in which the resource belongs.
	// If it is not provided, the provider project is used.
	Project pulumi.StringPtrInput
	// The URI of the created resource.
	SelfLink pulumi.StringPtrInput
}

func (DatasetState) ElementType

func (DatasetState) ElementType() reflect.Type

type GetDefaultServiceAccountArgs

type GetDefaultServiceAccountArgs struct {
	// The project the unique service account was created for. If it is not provided, the provider project is used.
	Project *string `pulumi:"project"`
}

A collection of arguments for invoking getDefaultServiceAccount.

type GetDefaultServiceAccountResult

type GetDefaultServiceAccountResult struct {
	// The email address of the service account. This value is often used to refer to the service account
	// in order to grant IAM permissions.
	Email string `pulumi:"email"`
	// The provider-assigned unique ID for this managed resource.
	Id      string `pulumi:"id"`
	Project string `pulumi:"project"`
}

A collection of values returned by getDefaultServiceAccount.

func GetDefaultServiceAccount

func GetDefaultServiceAccount(ctx *pulumi.Context, args *GetDefaultServiceAccountArgs, opts ...pulumi.InvokeOption) (*GetDefaultServiceAccountResult, error)

Get the email address of a project's unique BigQuery service account.

Each Google Cloud project has a unique service account used by BigQuery. When using BigQuery with [customer-managed encryption keys](https://cloud.google.com/bigquery/docs/customer-managed-encryption), this account needs to be granted the `cloudkms.cryptoKeyEncrypterDecrypter` IAM role on the customer-managed Cloud KMS key used to protect the data.

For more information see [the API reference](https://cloud.google.com/bigquery/docs/reference/rest/v2/projects/getServiceAccount).

type IamBinding

type IamBinding struct {
	pulumi.CustomResourceState

	// An [IAM Condition](https://cloud.google.com/iam/docs/conditions-overview) for a given binding.
	// Structure is documented below.
	Condition IamBindingConditionPtrOutput `pulumi:"condition"`
	DatasetId pulumi.StringOutput          `pulumi:"datasetId"`
	// (Computed) The etag of the IAM policy.
	Etag    pulumi.StringOutput      `pulumi:"etag"`
	Members pulumi.StringArrayOutput `pulumi:"members"`
	// The ID of the project in which the resource belongs.
	// If it is not provided, the project will be parsed from the identifier of the parent resource. If no project is provided in the parent identifier and no project is specified, the provider project is used.
	Project pulumi.StringOutput `pulumi:"project"`
	// The role that should be applied. Only one
	// `bigquery.IamBinding` can be used per role. Note that custom roles must be of the format
	// `[projects|organizations]/{parent-name}/roles/{role-name}`.
	Role    pulumi.StringOutput `pulumi:"role"`
	TableId pulumi.StringOutput `pulumi:"tableId"`
}

Three different resources help you manage your IAM policy for BigQuery Table. Each of these resources serves a different use case:

* `bigquery.IamPolicy`: Authoritative. Sets the IAM policy for the table and replaces any existing policy already attached. * `bigquery.IamBinding`: Authoritative for a given role. Updates the IAM policy to grant a role to a list of members. Other roles within the IAM policy for the table are preserved. * `bigquery.IamMember`: Non-authoritative. Updates the IAM policy to grant a role to a new member. Other members for the role for the table are preserved.

> **Note:** `bigquery.IamPolicy` **cannot** be used in conjunction with `bigquery.IamBinding` and `bigquery.IamMember` or they will fight over what your policy should be.

> **Note:** `bigquery.IamBinding` resources **can be** used in conjunction with `bigquery.IamMember` resources **only if** they do not grant privilege to the same role.

func GetIamBinding

func GetIamBinding(ctx *pulumi.Context,
	name string, id pulumi.IDInput, state *IamBindingState, opts ...pulumi.ResourceOption) (*IamBinding, error)

GetIamBinding gets an existing IamBinding resource's state with the given name, ID, and optional state properties that are used to uniquely qualify the lookup (nil if not required).

func NewIamBinding

func NewIamBinding(ctx *pulumi.Context,
	name string, args *IamBindingArgs, opts ...pulumi.ResourceOption) (*IamBinding, error)

NewIamBinding registers a new resource with the given unique name, arguments, and options.

type IamBindingArgs

type IamBindingArgs struct {
	// An [IAM Condition](https://cloud.google.com/iam/docs/conditions-overview) for a given binding.
	// Structure is documented below.
	Condition IamBindingConditionPtrInput
	DatasetId pulumi.StringInput
	Members   pulumi.StringArrayInput
	// The ID of the project in which the resource belongs.
	// If it is not provided, the project will be parsed from the identifier of the parent resource. If no project is provided in the parent identifier and no project is specified, the provider project is used.
	Project pulumi.StringPtrInput
	// The role that should be applied. Only one
	// `bigquery.IamBinding` can be used per role. Note that custom roles must be of the format
	// `[projects|organizations]/{parent-name}/roles/{role-name}`.
	Role    pulumi.StringInput
	TableId pulumi.StringInput
}

The set of arguments for constructing a IamBinding resource.

func (IamBindingArgs) ElementType

func (IamBindingArgs) ElementType() reflect.Type

type IamBindingCondition

type IamBindingCondition struct {
	// An optional description of the expression. This is a longer text which describes the expression, e.g. when hovered over it in a UI.
	Description *string `pulumi:"description"`
	// Textual representation of an expression in Common Expression Language syntax.
	Expression string `pulumi:"expression"`
	// A title for the expression, i.e. a short string describing its purpose.
	Title string `pulumi:"title"`
}

type IamBindingConditionArgs

type IamBindingConditionArgs struct {
	// An optional description of the expression. This is a longer text which describes the expression, e.g. when hovered over it in a UI.
	Description pulumi.StringPtrInput `pulumi:"description"`
	// Textual representation of an expression in Common Expression Language syntax.
	Expression pulumi.StringInput `pulumi:"expression"`
	// A title for the expression, i.e. a short string describing its purpose.
	Title pulumi.StringInput `pulumi:"title"`
}

func (IamBindingConditionArgs) ElementType

func (IamBindingConditionArgs) ElementType() reflect.Type

func (IamBindingConditionArgs) ToIamBindingConditionOutput

func (i IamBindingConditionArgs) ToIamBindingConditionOutput() IamBindingConditionOutput

func (IamBindingConditionArgs) ToIamBindingConditionOutputWithContext

func (i IamBindingConditionArgs) ToIamBindingConditionOutputWithContext(ctx context.Context) IamBindingConditionOutput

func (IamBindingConditionArgs) ToIamBindingConditionPtrOutput

func (i IamBindingConditionArgs) ToIamBindingConditionPtrOutput() IamBindingConditionPtrOutput

func (IamBindingConditionArgs) ToIamBindingConditionPtrOutputWithContext

func (i IamBindingConditionArgs) ToIamBindingConditionPtrOutputWithContext(ctx context.Context) IamBindingConditionPtrOutput

type IamBindingConditionInput

type IamBindingConditionInput interface {
	pulumi.Input

	ToIamBindingConditionOutput() IamBindingConditionOutput
	ToIamBindingConditionOutputWithContext(context.Context) IamBindingConditionOutput
}

IamBindingConditionInput is an input type that accepts IamBindingConditionArgs and IamBindingConditionOutput values. You can construct a concrete instance of `IamBindingConditionInput` via:

IamBindingConditionArgs{...}

type IamBindingConditionOutput

type IamBindingConditionOutput struct{ *pulumi.OutputState }

func (IamBindingConditionOutput) Description

An optional description of the expression. This is a longer text which describes the expression, e.g. when hovered over it in a UI.

func (IamBindingConditionOutput) ElementType

func (IamBindingConditionOutput) ElementType() reflect.Type

func (IamBindingConditionOutput) Expression

Textual representation of an expression in Common Expression Language syntax.

func (IamBindingConditionOutput) Title

A title for the expression, i.e. a short string describing its purpose.

func (IamBindingConditionOutput) ToIamBindingConditionOutput

func (o IamBindingConditionOutput) ToIamBindingConditionOutput() IamBindingConditionOutput

func (IamBindingConditionOutput) ToIamBindingConditionOutputWithContext

func (o IamBindingConditionOutput) ToIamBindingConditionOutputWithContext(ctx context.Context) IamBindingConditionOutput

func (IamBindingConditionOutput) ToIamBindingConditionPtrOutput

func (o IamBindingConditionOutput) ToIamBindingConditionPtrOutput() IamBindingConditionPtrOutput

func (IamBindingConditionOutput) ToIamBindingConditionPtrOutputWithContext

func (o IamBindingConditionOutput) ToIamBindingConditionPtrOutputWithContext(ctx context.Context) IamBindingConditionPtrOutput

type IamBindingConditionPtrInput

type IamBindingConditionPtrInput interface {
	pulumi.Input

	ToIamBindingConditionPtrOutput() IamBindingConditionPtrOutput
	ToIamBindingConditionPtrOutputWithContext(context.Context) IamBindingConditionPtrOutput
}

IamBindingConditionPtrInput is an input type that accepts IamBindingConditionArgs, IamBindingConditionPtr and IamBindingConditionPtrOutput values. You can construct a concrete instance of `IamBindingConditionPtrInput` via:

        IamBindingConditionArgs{...}

or:

        nil

type IamBindingConditionPtrOutput

type IamBindingConditionPtrOutput struct{ *pulumi.OutputState }

func (IamBindingConditionPtrOutput) Description

An optional description of the expression. This is a longer text which describes the expression, e.g. when hovered over it in a UI.

func (IamBindingConditionPtrOutput) Elem

func (IamBindingConditionPtrOutput) ElementType

func (IamBindingConditionPtrOutput) Expression

Textual representation of an expression in Common Expression Language syntax.

func (IamBindingConditionPtrOutput) Title

A title for the expression, i.e. a short string describing its purpose.

func (IamBindingConditionPtrOutput) ToIamBindingConditionPtrOutput

func (o IamBindingConditionPtrOutput) ToIamBindingConditionPtrOutput() IamBindingConditionPtrOutput

func (IamBindingConditionPtrOutput) ToIamBindingConditionPtrOutputWithContext

func (o IamBindingConditionPtrOutput) ToIamBindingConditionPtrOutputWithContext(ctx context.Context) IamBindingConditionPtrOutput

type IamBindingState

type IamBindingState struct {
	// An [IAM Condition](https://cloud.google.com/iam/docs/conditions-overview) for a given binding.
	// Structure is documented below.
	Condition IamBindingConditionPtrInput
	DatasetId pulumi.StringPtrInput
	// (Computed) The etag of the IAM policy.
	Etag    pulumi.StringPtrInput
	Members pulumi.StringArrayInput
	// The ID of the project in which the resource belongs.
	// If it is not provided, the project will be parsed from the identifier of the parent resource. If no project is provided in the parent identifier and no project is specified, the provider project is used.
	Project pulumi.StringPtrInput
	// The role that should be applied. Only one
	// `bigquery.IamBinding` can be used per role. Note that custom roles must be of the format
	// `[projects|organizations]/{parent-name}/roles/{role-name}`.
	Role    pulumi.StringPtrInput
	TableId pulumi.StringPtrInput
}

func (IamBindingState) ElementType

func (IamBindingState) ElementType() reflect.Type

type IamMember

type IamMember struct {
	pulumi.CustomResourceState

	// An [IAM Condition](https://cloud.google.com/iam/docs/conditions-overview) for a given binding.
	// Structure is documented below.
	Condition IamMemberConditionPtrOutput `pulumi:"condition"`
	DatasetId pulumi.StringOutput         `pulumi:"datasetId"`
	// (Computed) The etag of the IAM policy.
	Etag   pulumi.StringOutput `pulumi:"etag"`
	Member pulumi.StringOutput `pulumi:"member"`
	// The ID of the project in which the resource belongs.
	// If it is not provided, the project will be parsed from the identifier of the parent resource. If no project is provided in the parent identifier and no project is specified, the provider project is used.
	Project pulumi.StringOutput `pulumi:"project"`
	// The role that should be applied. Only one
	// `bigquery.IamBinding` can be used per role. Note that custom roles must be of the format
	// `[projects|organizations]/{parent-name}/roles/{role-name}`.
	Role    pulumi.StringOutput `pulumi:"role"`
	TableId pulumi.StringOutput `pulumi:"tableId"`
}

Three different resources help you manage your IAM policy for BigQuery Table. Each of these resources serves a different use case:

* `bigquery.IamPolicy`: Authoritative. Sets the IAM policy for the table and replaces any existing policy already attached. * `bigquery.IamBinding`: Authoritative for a given role. Updates the IAM policy to grant a role to a list of members. Other roles within the IAM policy for the table are preserved. * `bigquery.IamMember`: Non-authoritative. Updates the IAM policy to grant a role to a new member. Other members for the role for the table are preserved.

> **Note:** `bigquery.IamPolicy` **cannot** be used in conjunction with `bigquery.IamBinding` and `bigquery.IamMember` or they will fight over what your policy should be.

> **Note:** `bigquery.IamBinding` resources **can be** used in conjunction with `bigquery.IamMember` resources **only if** they do not grant privilege to the same role.

func GetIamMember

func GetIamMember(ctx *pulumi.Context,
	name string, id pulumi.IDInput, state *IamMemberState, opts ...pulumi.ResourceOption) (*IamMember, error)

GetIamMember gets an existing IamMember resource's state with the given name, ID, and optional state properties that are used to uniquely qualify the lookup (nil if not required).

func NewIamMember

func NewIamMember(ctx *pulumi.Context,
	name string, args *IamMemberArgs, opts ...pulumi.ResourceOption) (*IamMember, error)

NewIamMember registers a new resource with the given unique name, arguments, and options.

type IamMemberArgs

type IamMemberArgs struct {
	// An [IAM Condition](https://cloud.google.com/iam/docs/conditions-overview) for a given binding.
	// Structure is documented below.
	Condition IamMemberConditionPtrInput
	DatasetId pulumi.StringInput
	Member    pulumi.StringInput
	// The ID of the project in which the resource belongs.
	// If it is not provided, the project will be parsed from the identifier of the parent resource. If no project is provided in the parent identifier and no project is specified, the provider project is used.
	Project pulumi.StringPtrInput
	// The role that should be applied. Only one
	// `bigquery.IamBinding` can be used per role. Note that custom roles must be of the format
	// `[projects|organizations]/{parent-name}/roles/{role-name}`.
	Role    pulumi.StringInput
	TableId pulumi.StringInput
}

The set of arguments for constructing a IamMember resource.

func (IamMemberArgs) ElementType

func (IamMemberArgs) ElementType() reflect.Type

type IamMemberCondition

type IamMemberCondition struct {
	// An optional description of the expression. This is a longer text which describes the expression, e.g. when hovered over it in a UI.
	Description *string `pulumi:"description"`
	// Textual representation of an expression in Common Expression Language syntax.
	Expression string `pulumi:"expression"`
	// A title for the expression, i.e. a short string describing its purpose.
	Title string `pulumi:"title"`
}

type IamMemberConditionArgs

type IamMemberConditionArgs struct {
	// An optional description of the expression. This is a longer text which describes the expression, e.g. when hovered over it in a UI.
	Description pulumi.StringPtrInput `pulumi:"description"`
	// Textual representation of an expression in Common Expression Language syntax.
	Expression pulumi.StringInput `pulumi:"expression"`
	// A title for the expression, i.e. a short string describing its purpose.
	Title pulumi.StringInput `pulumi:"title"`
}

func (IamMemberConditionArgs) ElementType

func (IamMemberConditionArgs) ElementType() reflect.Type

func (IamMemberConditionArgs) ToIamMemberConditionOutput

func (i IamMemberConditionArgs) ToIamMemberConditionOutput() IamMemberConditionOutput

func (IamMemberConditionArgs) ToIamMemberConditionOutputWithContext

func (i IamMemberConditionArgs) ToIamMemberConditionOutputWithContext(ctx context.Context) IamMemberConditionOutput

func (IamMemberConditionArgs) ToIamMemberConditionPtrOutput

func (i IamMemberConditionArgs) ToIamMemberConditionPtrOutput() IamMemberConditionPtrOutput

func (IamMemberConditionArgs) ToIamMemberConditionPtrOutputWithContext

func (i IamMemberConditionArgs) ToIamMemberConditionPtrOutputWithContext(ctx context.Context) IamMemberConditionPtrOutput

type IamMemberConditionInput

type IamMemberConditionInput interface {
	pulumi.Input

	ToIamMemberConditionOutput() IamMemberConditionOutput
	ToIamMemberConditionOutputWithContext(context.Context) IamMemberConditionOutput
}

IamMemberConditionInput is an input type that accepts IamMemberConditionArgs and IamMemberConditionOutput values. You can construct a concrete instance of `IamMemberConditionInput` via:

IamMemberConditionArgs{...}

type IamMemberConditionOutput

type IamMemberConditionOutput struct{ *pulumi.OutputState }

func (IamMemberConditionOutput) Description

An optional description of the expression. This is a longer text which describes the expression, e.g. when hovered over it in a UI.

func (IamMemberConditionOutput) ElementType

func (IamMemberConditionOutput) ElementType() reflect.Type

func (IamMemberConditionOutput) Expression

Textual representation of an expression in Common Expression Language syntax.

func (IamMemberConditionOutput) Title

A title for the expression, i.e. a short string describing its purpose.

func (IamMemberConditionOutput) ToIamMemberConditionOutput

func (o IamMemberConditionOutput) ToIamMemberConditionOutput() IamMemberConditionOutput

func (IamMemberConditionOutput) ToIamMemberConditionOutputWithContext

func (o IamMemberConditionOutput) ToIamMemberConditionOutputWithContext(ctx context.Context) IamMemberConditionOutput

func (IamMemberConditionOutput) ToIamMemberConditionPtrOutput

func (o IamMemberConditionOutput) ToIamMemberConditionPtrOutput() IamMemberConditionPtrOutput

func (IamMemberConditionOutput) ToIamMemberConditionPtrOutputWithContext

func (o IamMemberConditionOutput) ToIamMemberConditionPtrOutputWithContext(ctx context.Context) IamMemberConditionPtrOutput

type IamMemberConditionPtrInput

type IamMemberConditionPtrInput interface {
	pulumi.Input

	ToIamMemberConditionPtrOutput() IamMemberConditionPtrOutput
	ToIamMemberConditionPtrOutputWithContext(context.Context) IamMemberConditionPtrOutput
}

IamMemberConditionPtrInput is an input type that accepts IamMemberConditionArgs, IamMemberConditionPtr and IamMemberConditionPtrOutput values. You can construct a concrete instance of `IamMemberConditionPtrInput` via:

        IamMemberConditionArgs{...}

or:

        nil

type IamMemberConditionPtrOutput

type IamMemberConditionPtrOutput struct{ *pulumi.OutputState }

func (IamMemberConditionPtrOutput) Description

An optional description of the expression. This is a longer text which describes the expression, e.g. when hovered over it in a UI.

func (IamMemberConditionPtrOutput) Elem

func (IamMemberConditionPtrOutput) ElementType

func (IamMemberConditionPtrOutput) Expression

Textual representation of an expression in Common Expression Language syntax.

func (IamMemberConditionPtrOutput) Title

A title for the expression, i.e. a short string describing its purpose.

func (IamMemberConditionPtrOutput) ToIamMemberConditionPtrOutput

func (o IamMemberConditionPtrOutput) ToIamMemberConditionPtrOutput() IamMemberConditionPtrOutput

func (IamMemberConditionPtrOutput) ToIamMemberConditionPtrOutputWithContext

func (o IamMemberConditionPtrOutput) ToIamMemberConditionPtrOutputWithContext(ctx context.Context) IamMemberConditionPtrOutput

type IamMemberState

type IamMemberState struct {
	// An [IAM Condition](https://cloud.google.com/iam/docs/conditions-overview) for a given binding.
	// Structure is documented below.
	Condition IamMemberConditionPtrInput
	DatasetId pulumi.StringPtrInput
	// (Computed) The etag of the IAM policy.
	Etag   pulumi.StringPtrInput
	Member pulumi.StringPtrInput
	// The ID of the project in which the resource belongs.
	// If it is not provided, the project will be parsed from the identifier of the parent resource. If no project is provided in the parent identifier and no project is specified, the provider project is used.
	Project pulumi.StringPtrInput
	// The role that should be applied. Only one
	// `bigquery.IamBinding` can be used per role. Note that custom roles must be of the format
	// `[projects|organizations]/{parent-name}/roles/{role-name}`.
	Role    pulumi.StringPtrInput
	TableId pulumi.StringPtrInput
}

func (IamMemberState) ElementType

func (IamMemberState) ElementType() reflect.Type

type IamPolicy

type IamPolicy struct {
	pulumi.CustomResourceState

	DatasetId pulumi.StringOutput `pulumi:"datasetId"`
	// (Computed) The etag of the IAM policy.
	Etag pulumi.StringOutput `pulumi:"etag"`
	// The policy data generated by
	// a `organizations.getIAMPolicy` data source.
	PolicyData pulumi.StringOutput `pulumi:"policyData"`
	// The ID of the project in which the resource belongs.
	// If it is not provided, the project will be parsed from the identifier of the parent resource. If no project is provided in the parent identifier and no project is specified, the provider project is used.
	Project pulumi.StringOutput `pulumi:"project"`
	TableId pulumi.StringOutput `pulumi:"tableId"`
}

Three different resources help you manage your IAM policy for BigQuery Table. Each of these resources serves a different use case:

* `bigquery.IamPolicy`: Authoritative. Sets the IAM policy for the table and replaces any existing policy already attached. * `bigquery.IamBinding`: Authoritative for a given role. Updates the IAM policy to grant a role to a list of members. Other roles within the IAM policy for the table are preserved. * `bigquery.IamMember`: Non-authoritative. Updates the IAM policy to grant a role to a new member. Other members for the role for the table are preserved.

> **Note:** `bigquery.IamPolicy` **cannot** be used in conjunction with `bigquery.IamBinding` and `bigquery.IamMember` or they will fight over what your policy should be.

> **Note:** `bigquery.IamBinding` resources **can be** used in conjunction with `bigquery.IamMember` resources **only if** they do not grant privilege to the same role.

func GetIamPolicy

func GetIamPolicy(ctx *pulumi.Context,
	name string, id pulumi.IDInput, state *IamPolicyState, opts ...pulumi.ResourceOption) (*IamPolicy, error)

GetIamPolicy gets an existing IamPolicy resource's state with the given name, ID, and optional state properties that are used to uniquely qualify the lookup (nil if not required).

func NewIamPolicy

func NewIamPolicy(ctx *pulumi.Context,
	name string, args *IamPolicyArgs, opts ...pulumi.ResourceOption) (*IamPolicy, error)

NewIamPolicy registers a new resource with the given unique name, arguments, and options.

type IamPolicyArgs

type IamPolicyArgs struct {
	DatasetId pulumi.StringInput
	// The policy data generated by
	// a `organizations.getIAMPolicy` data source.
	PolicyData pulumi.StringInput
	// The ID of the project in which the resource belongs.
	// If it is not provided, the project will be parsed from the identifier of the parent resource. If no project is provided in the parent identifier and no project is specified, the provider project is used.
	Project pulumi.StringPtrInput
	TableId pulumi.StringInput
}

The set of arguments for constructing a IamPolicy resource.

func (IamPolicyArgs) ElementType

func (IamPolicyArgs) ElementType() reflect.Type

type IamPolicyState

type IamPolicyState struct {
	DatasetId pulumi.StringPtrInput
	// (Computed) The etag of the IAM policy.
	Etag pulumi.StringPtrInput
	// The policy data generated by
	// a `organizations.getIAMPolicy` data source.
	PolicyData pulumi.StringPtrInput
	// The ID of the project in which the resource belongs.
	// If it is not provided, the project will be parsed from the identifier of the parent resource. If no project is provided in the parent identifier and no project is specified, the provider project is used.
	Project pulumi.StringPtrInput
	TableId pulumi.StringPtrInput
}

func (IamPolicyState) ElementType

func (IamPolicyState) ElementType() reflect.Type

type Job

type Job struct {
	pulumi.CustomResourceState

	// Copies a table.
	// Structure is documented below.
	Copy JobCopyPtrOutput `pulumi:"copy"`
	// Configures an extract job.
	// Structure is documented below.
	Extract JobExtractPtrOutput `pulumi:"extract"`
	// The ID of the job. The ID must contain only letters (a-z, A-Z), numbers (0-9), underscores (_), or dashes (-). The maximum length is 1,024 characters.
	JobId pulumi.StringOutput `pulumi:"jobId"`
	// Job timeout in milliseconds. If this time limit is exceeded, BigQuery may attempt to terminate the job.
	JobTimeoutMs pulumi.StringPtrOutput `pulumi:"jobTimeoutMs"`
	// The type of the job.
	JobType pulumi.StringOutput `pulumi:"jobType"`
	// The labels associated with this job. You can use these to organize and group your jobs.
	Labels pulumi.StringMapOutput `pulumi:"labels"`
	// Configures a load job.
	// Structure is documented below.
	Load JobLoadPtrOutput `pulumi:"load"`
	// The geographic location of the job. The default value is US.
	Location pulumi.StringPtrOutput `pulumi:"location"`
	// The ID of the project in which the resource belongs.
	// If it is not provided, the provider project is used.
	Project pulumi.StringOutput `pulumi:"project"`
	// Configures a query job.
	// Structure is documented below.
	Query JobQueryPtrOutput `pulumi:"query"`
	// Email address of the user who ran the job.
	UserEmail pulumi.StringOutput `pulumi:"userEmail"`
}

Jobs are actions that BigQuery runs on your behalf to load data, export data, query data, or copy data. Once a BigQuery job is created, it cannot be changed or deleted.

## Example Usage

func GetJob

func GetJob(ctx *pulumi.Context,
	name string, id pulumi.IDInput, state *JobState, opts ...pulumi.ResourceOption) (*Job, error)

GetJob gets an existing Job resource's state with the given name, ID, and optional state properties that are used to uniquely qualify the lookup (nil if not required).

func NewJob

func NewJob(ctx *pulumi.Context,
	name string, args *JobArgs, opts ...pulumi.ResourceOption) (*Job, error)

NewJob registers a new resource with the given unique name, arguments, and options.

type JobArgs

type JobArgs struct {
	// Copies a table.
	// Structure is documented below.
	Copy JobCopyPtrInput
	// Configures an extract job.
	// Structure is documented below.
	Extract JobExtractPtrInput
	// The ID of the job. The ID must contain only letters (a-z, A-Z), numbers (0-9), underscores (_), or dashes (-). The maximum length is 1,024 characters.
	JobId pulumi.StringInput
	// Job timeout in milliseconds. If this time limit is exceeded, BigQuery may attempt to terminate the job.
	JobTimeoutMs pulumi.StringPtrInput
	// The labels associated with this job. You can use these to organize and group your jobs.
	Labels pulumi.StringMapInput
	// Configures a load job.
	// Structure is documented below.
	Load JobLoadPtrInput
	// The geographic location of the job. The default value is US.
	Location pulumi.StringPtrInput
	// The ID of the project in which the resource belongs.
	// If it is not provided, the provider project is used.
	Project pulumi.StringPtrInput
	// Configures a query job.
	// Structure is documented below.
	Query JobQueryPtrInput
}

The set of arguments for constructing a Job resource.

func (JobArgs) ElementType

func (JobArgs) ElementType() reflect.Type

type JobCopy

type JobCopy struct {
	// Specifies whether the job is allowed to create new tables. The following values are supported:
	// CREATE_IF_NEEDED: If the table does not exist, BigQuery creates the table.
	// CREATE_NEVER: The table must already exist. If it does not, a 'notFound' error is returned in the job result.
	// Creation, truncation and append actions occur as one atomic update upon job completion
	// Default value is `CREATE_IF_NEEDED`.
	// Possible values are `CREATE_IF_NEEDED` and `CREATE_NEVER`.
	CreateDisposition *string `pulumi:"createDisposition"`
	// Custom encryption configuration (e.g., Cloud KMS keys)
	// Structure is documented below.
	DestinationEncryptionConfiguration *JobCopyDestinationEncryptionConfiguration `pulumi:"destinationEncryptionConfiguration"`
	// The destination table.
	// Structure is documented below.
	DestinationTable *JobCopyDestinationTable `pulumi:"destinationTable"`
	// Source tables to copy.
	// Structure is documented below.
	SourceTables []JobCopySourceTable `pulumi:"sourceTables"`
	// Specifies the action that occurs if the destination table already exists. The following values are supported:
	// WRITE_TRUNCATE: If the table already exists, BigQuery overwrites the table data and uses the schema from the query result.
	// WRITE_APPEND: If the table already exists, BigQuery appends the data to the table.
	// WRITE_EMPTY: If the table already exists and contains data, a 'duplicate' error is returned in the job result.
	// Each action is atomic and only occurs if BigQuery is able to complete the job successfully.
	// Creation, truncation and append actions occur as one atomic update upon job completion.
	// Default value is `WRITE_EMPTY`.
	// Possible values are `WRITE_TRUNCATE`, `WRITE_APPEND`, and `WRITE_EMPTY`.
	WriteDisposition *string `pulumi:"writeDisposition"`
}

type JobCopyArgs

type JobCopyArgs struct {
	// Specifies whether the job is allowed to create new tables. The following values are supported:
	// CREATE_IF_NEEDED: If the table does not exist, BigQuery creates the table.
	// CREATE_NEVER: The table must already exist. If it does not, a 'notFound' error is returned in the job result.
	// Creation, truncation and append actions occur as one atomic update upon job completion
	// Default value is `CREATE_IF_NEEDED`.
	// Possible values are `CREATE_IF_NEEDED` and `CREATE_NEVER`.
	CreateDisposition pulumi.StringPtrInput `pulumi:"createDisposition"`
	// Custom encryption configuration (e.g., Cloud KMS keys)
	// Structure is documented below.
	DestinationEncryptionConfiguration JobCopyDestinationEncryptionConfigurationPtrInput `pulumi:"destinationEncryptionConfiguration"`
	// The destination table.
	// Structure is documented below.
	DestinationTable JobCopyDestinationTablePtrInput `pulumi:"destinationTable"`
	// Source tables to copy.
	// Structure is documented below.
	SourceTables JobCopySourceTableArrayInput `pulumi:"sourceTables"`
	// Specifies the action that occurs if the destination table already exists. The following values are supported:
	// WRITE_TRUNCATE: If the table already exists, BigQuery overwrites the table data and uses the schema from the query result.
	// WRITE_APPEND: If the table already exists, BigQuery appends the data to the table.
	// WRITE_EMPTY: If the table already exists and contains data, a 'duplicate' error is returned in the job result.
	// Each action is atomic and only occurs if BigQuery is able to complete the job successfully.
	// Creation, truncation and append actions occur as one atomic update upon job completion.
	// Default value is `WRITE_EMPTY`.
	// Possible values are `WRITE_TRUNCATE`, `WRITE_APPEND`, and `WRITE_EMPTY`.
	WriteDisposition pulumi.StringPtrInput `pulumi:"writeDisposition"`
}

func (JobCopyArgs) ElementType

func (JobCopyArgs) ElementType() reflect.Type

func (JobCopyArgs) ToJobCopyOutput

func (i JobCopyArgs) ToJobCopyOutput() JobCopyOutput

func (JobCopyArgs) ToJobCopyOutputWithContext

func (i JobCopyArgs) ToJobCopyOutputWithContext(ctx context.Context) JobCopyOutput

func (JobCopyArgs) ToJobCopyPtrOutput

func (i JobCopyArgs) ToJobCopyPtrOutput() JobCopyPtrOutput

func (JobCopyArgs) ToJobCopyPtrOutputWithContext

func (i JobCopyArgs) ToJobCopyPtrOutputWithContext(ctx context.Context) JobCopyPtrOutput

type JobCopyDestinationEncryptionConfiguration

type JobCopyDestinationEncryptionConfiguration struct {
	// Describes the Cloud KMS encryption key that will be used to protect destination BigQuery table.
	// The BigQuery Service Account associated with your project requires access to this encryption key.
	KmsKeyName string `pulumi:"kmsKeyName"`
}

type JobCopyDestinationEncryptionConfigurationArgs

type JobCopyDestinationEncryptionConfigurationArgs struct {
	// Describes the Cloud KMS encryption key that will be used to protect destination BigQuery table.
	// The BigQuery Service Account associated with your project requires access to this encryption key.
	KmsKeyName pulumi.StringInput `pulumi:"kmsKeyName"`
}

func (JobCopyDestinationEncryptionConfigurationArgs) ElementType

func (JobCopyDestinationEncryptionConfigurationArgs) ToJobCopyDestinationEncryptionConfigurationOutput

func (i JobCopyDestinationEncryptionConfigurationArgs) ToJobCopyDestinationEncryptionConfigurationOutput() JobCopyDestinationEncryptionConfigurationOutput

func (JobCopyDestinationEncryptionConfigurationArgs) ToJobCopyDestinationEncryptionConfigurationOutputWithContext

func (i JobCopyDestinationEncryptionConfigurationArgs) ToJobCopyDestinationEncryptionConfigurationOutputWithContext(ctx context.Context) JobCopyDestinationEncryptionConfigurationOutput

func (JobCopyDestinationEncryptionConfigurationArgs) ToJobCopyDestinationEncryptionConfigurationPtrOutput

func (i JobCopyDestinationEncryptionConfigurationArgs) ToJobCopyDestinationEncryptionConfigurationPtrOutput() JobCopyDestinationEncryptionConfigurationPtrOutput

func (JobCopyDestinationEncryptionConfigurationArgs) ToJobCopyDestinationEncryptionConfigurationPtrOutputWithContext

func (i JobCopyDestinationEncryptionConfigurationArgs) ToJobCopyDestinationEncryptionConfigurationPtrOutputWithContext(ctx context.Context) JobCopyDestinationEncryptionConfigurationPtrOutput

type JobCopyDestinationEncryptionConfigurationInput

type JobCopyDestinationEncryptionConfigurationInput interface {
	pulumi.Input

	ToJobCopyDestinationEncryptionConfigurationOutput() JobCopyDestinationEncryptionConfigurationOutput
	ToJobCopyDestinationEncryptionConfigurationOutputWithContext(context.Context) JobCopyDestinationEncryptionConfigurationOutput
}

JobCopyDestinationEncryptionConfigurationInput is an input type that accepts JobCopyDestinationEncryptionConfigurationArgs and JobCopyDestinationEncryptionConfigurationOutput values. You can construct a concrete instance of `JobCopyDestinationEncryptionConfigurationInput` via:

JobCopyDestinationEncryptionConfigurationArgs{...}

type JobCopyDestinationEncryptionConfigurationOutput

type JobCopyDestinationEncryptionConfigurationOutput struct{ *pulumi.OutputState }

func (JobCopyDestinationEncryptionConfigurationOutput) ElementType

func (JobCopyDestinationEncryptionConfigurationOutput) KmsKeyName

Describes the Cloud KMS encryption key that will be used to protect destination BigQuery table. The BigQuery Service Account associated with your project requires access to this encryption key.

func (JobCopyDestinationEncryptionConfigurationOutput) ToJobCopyDestinationEncryptionConfigurationOutput

func (o JobCopyDestinationEncryptionConfigurationOutput) ToJobCopyDestinationEncryptionConfigurationOutput() JobCopyDestinationEncryptionConfigurationOutput

func (JobCopyDestinationEncryptionConfigurationOutput) ToJobCopyDestinationEncryptionConfigurationOutputWithContext

func (o JobCopyDestinationEncryptionConfigurationOutput) ToJobCopyDestinationEncryptionConfigurationOutputWithContext(ctx context.Context) JobCopyDestinationEncryptionConfigurationOutput

func (JobCopyDestinationEncryptionConfigurationOutput) ToJobCopyDestinationEncryptionConfigurationPtrOutput

func (o JobCopyDestinationEncryptionConfigurationOutput) ToJobCopyDestinationEncryptionConfigurationPtrOutput() JobCopyDestinationEncryptionConfigurationPtrOutput

func (JobCopyDestinationEncryptionConfigurationOutput) ToJobCopyDestinationEncryptionConfigurationPtrOutputWithContext

func (o JobCopyDestinationEncryptionConfigurationOutput) ToJobCopyDestinationEncryptionConfigurationPtrOutputWithContext(ctx context.Context) JobCopyDestinationEncryptionConfigurationPtrOutput

type JobCopyDestinationEncryptionConfigurationPtrInput

type JobCopyDestinationEncryptionConfigurationPtrInput interface {
	pulumi.Input

	ToJobCopyDestinationEncryptionConfigurationPtrOutput() JobCopyDestinationEncryptionConfigurationPtrOutput
	ToJobCopyDestinationEncryptionConfigurationPtrOutputWithContext(context.Context) JobCopyDestinationEncryptionConfigurationPtrOutput
}

JobCopyDestinationEncryptionConfigurationPtrInput is an input type that accepts JobCopyDestinationEncryptionConfigurationArgs, JobCopyDestinationEncryptionConfigurationPtr and JobCopyDestinationEncryptionConfigurationPtrOutput values. You can construct a concrete instance of `JobCopyDestinationEncryptionConfigurationPtrInput` via:

        JobCopyDestinationEncryptionConfigurationArgs{...}

or:

        nil

type JobCopyDestinationEncryptionConfigurationPtrOutput

type JobCopyDestinationEncryptionConfigurationPtrOutput struct{ *pulumi.OutputState }

func (JobCopyDestinationEncryptionConfigurationPtrOutput) Elem

func (JobCopyDestinationEncryptionConfigurationPtrOutput) ElementType

func (JobCopyDestinationEncryptionConfigurationPtrOutput) KmsKeyName

Describes the Cloud KMS encryption key that will be used to protect destination BigQuery table. The BigQuery Service Account associated with your project requires access to this encryption key.

func (JobCopyDestinationEncryptionConfigurationPtrOutput) ToJobCopyDestinationEncryptionConfigurationPtrOutput

func (o JobCopyDestinationEncryptionConfigurationPtrOutput) ToJobCopyDestinationEncryptionConfigurationPtrOutput() JobCopyDestinationEncryptionConfigurationPtrOutput

func (JobCopyDestinationEncryptionConfigurationPtrOutput) ToJobCopyDestinationEncryptionConfigurationPtrOutputWithContext

func (o JobCopyDestinationEncryptionConfigurationPtrOutput) ToJobCopyDestinationEncryptionConfigurationPtrOutputWithContext(ctx context.Context) JobCopyDestinationEncryptionConfigurationPtrOutput

type JobCopyDestinationTable

type JobCopyDestinationTable struct {
	// The ID of the dataset containing this model.
	DatasetId *string `pulumi:"datasetId"`
	// The ID of the project containing this model.
	ProjectId *string `pulumi:"projectId"`
	// The table. Can be specified `{{table_id}}` if `projectId` and `datasetId` are also set,
	// or of the form `projects/{{project}}/datasets/{{dataset_id}}/tables/{{table_id}}` if not.
	TableId string `pulumi:"tableId"`
}

type JobCopyDestinationTableArgs

type JobCopyDestinationTableArgs struct {
	// The ID of the dataset containing this model.
	DatasetId pulumi.StringPtrInput `pulumi:"datasetId"`
	// The ID of the project containing this model.
	ProjectId pulumi.StringPtrInput `pulumi:"projectId"`
	// The table. Can be specified `{{table_id}}` if `projectId` and `datasetId` are also set,
	// or of the form `projects/{{project}}/datasets/{{dataset_id}}/tables/{{table_id}}` if not.
	TableId pulumi.StringInput `pulumi:"tableId"`
}

func (JobCopyDestinationTableArgs) ElementType

func (JobCopyDestinationTableArgs) ToJobCopyDestinationTableOutput

func (i JobCopyDestinationTableArgs) ToJobCopyDestinationTableOutput() JobCopyDestinationTableOutput

func (JobCopyDestinationTableArgs) ToJobCopyDestinationTableOutputWithContext

func (i JobCopyDestinationTableArgs) ToJobCopyDestinationTableOutputWithContext(ctx context.Context) JobCopyDestinationTableOutput

func (JobCopyDestinationTableArgs) ToJobCopyDestinationTablePtrOutput

func (i JobCopyDestinationTableArgs) ToJobCopyDestinationTablePtrOutput() JobCopyDestinationTablePtrOutput

func (JobCopyDestinationTableArgs) ToJobCopyDestinationTablePtrOutputWithContext

func (i JobCopyDestinationTableArgs) ToJobCopyDestinationTablePtrOutputWithContext(ctx context.Context) JobCopyDestinationTablePtrOutput

type JobCopyDestinationTableInput

type JobCopyDestinationTableInput interface {
	pulumi.Input

	ToJobCopyDestinationTableOutput() JobCopyDestinationTableOutput
	ToJobCopyDestinationTableOutputWithContext(context.Context) JobCopyDestinationTableOutput
}

JobCopyDestinationTableInput is an input type that accepts JobCopyDestinationTableArgs and JobCopyDestinationTableOutput values. You can construct a concrete instance of `JobCopyDestinationTableInput` via:

JobCopyDestinationTableArgs{...}

type JobCopyDestinationTableOutput

type JobCopyDestinationTableOutput struct{ *pulumi.OutputState }

func (JobCopyDestinationTableOutput) DatasetId

The ID of the dataset containing this model.

func (JobCopyDestinationTableOutput) ElementType

func (JobCopyDestinationTableOutput) ProjectId

The ID of the project containing this model.

func (JobCopyDestinationTableOutput) TableId

The table. Can be specified `{{table_id}}` if `projectId` and `datasetId` are also set, or of the form `projects/{{project}}/datasets/{{dataset_id}}/tables/{{table_id}}` if not.

func (JobCopyDestinationTableOutput) ToJobCopyDestinationTableOutput

func (o JobCopyDestinationTableOutput) ToJobCopyDestinationTableOutput() JobCopyDestinationTableOutput

func (JobCopyDestinationTableOutput) ToJobCopyDestinationTableOutputWithContext

func (o JobCopyDestinationTableOutput) ToJobCopyDestinationTableOutputWithContext(ctx context.Context) JobCopyDestinationTableOutput

func (JobCopyDestinationTableOutput) ToJobCopyDestinationTablePtrOutput

func (o JobCopyDestinationTableOutput) ToJobCopyDestinationTablePtrOutput() JobCopyDestinationTablePtrOutput

func (JobCopyDestinationTableOutput) ToJobCopyDestinationTablePtrOutputWithContext

func (o JobCopyDestinationTableOutput) ToJobCopyDestinationTablePtrOutputWithContext(ctx context.Context) JobCopyDestinationTablePtrOutput

type JobCopyDestinationTablePtrInput

type JobCopyDestinationTablePtrInput interface {
	pulumi.Input

	ToJobCopyDestinationTablePtrOutput() JobCopyDestinationTablePtrOutput
	ToJobCopyDestinationTablePtrOutputWithContext(context.Context) JobCopyDestinationTablePtrOutput
}

JobCopyDestinationTablePtrInput is an input type that accepts JobCopyDestinationTableArgs, JobCopyDestinationTablePtr and JobCopyDestinationTablePtrOutput values. You can construct a concrete instance of `JobCopyDestinationTablePtrInput` via:

        JobCopyDestinationTableArgs{...}

or:

        nil

type JobCopyDestinationTablePtrOutput

type JobCopyDestinationTablePtrOutput struct{ *pulumi.OutputState }

func (JobCopyDestinationTablePtrOutput) DatasetId

The ID of the dataset containing this model.

func (JobCopyDestinationTablePtrOutput) Elem

func (JobCopyDestinationTablePtrOutput) ElementType

func (JobCopyDestinationTablePtrOutput) ProjectId

The ID of the project containing this model.

func (JobCopyDestinationTablePtrOutput) TableId

The table. Can be specified `{{table_id}}` if `projectId` and `datasetId` are also set, or of the form `projects/{{project}}/datasets/{{dataset_id}}/tables/{{table_id}}` if not.

func (JobCopyDestinationTablePtrOutput) ToJobCopyDestinationTablePtrOutput

func (o JobCopyDestinationTablePtrOutput) ToJobCopyDestinationTablePtrOutput() JobCopyDestinationTablePtrOutput

func (JobCopyDestinationTablePtrOutput) ToJobCopyDestinationTablePtrOutputWithContext

func (o JobCopyDestinationTablePtrOutput) ToJobCopyDestinationTablePtrOutputWithContext(ctx context.Context) JobCopyDestinationTablePtrOutput

type JobCopyInput

type JobCopyInput interface {
	pulumi.Input

	ToJobCopyOutput() JobCopyOutput
	ToJobCopyOutputWithContext(context.Context) JobCopyOutput
}

JobCopyInput is an input type that accepts JobCopyArgs and JobCopyOutput values. You can construct a concrete instance of `JobCopyInput` via:

JobCopyArgs{...}

type JobCopyOutput

type JobCopyOutput struct{ *pulumi.OutputState }

func (JobCopyOutput) CreateDisposition

func (o JobCopyOutput) CreateDisposition() pulumi.StringPtrOutput

Specifies whether the job is allowed to create new tables. The following values are supported: CREATE_IF_NEEDED: If the table does not exist, BigQuery creates the table. CREATE_NEVER: The table must already exist. If it does not, a 'notFound' error is returned in the job result. Creation, truncation and append actions occur as one atomic update upon job completion Default value is `CREATE_IF_NEEDED`. Possible values are `CREATE_IF_NEEDED` and `CREATE_NEVER`.

func (JobCopyOutput) DestinationEncryptionConfiguration

func (o JobCopyOutput) DestinationEncryptionConfiguration() JobCopyDestinationEncryptionConfigurationPtrOutput

Custom encryption configuration (e.g., Cloud KMS keys) Structure is documented below.

func (JobCopyOutput) DestinationTable

func (o JobCopyOutput) DestinationTable() JobCopyDestinationTablePtrOutput

The destination table. Structure is documented below.

func (JobCopyOutput) ElementType

func (JobCopyOutput) ElementType() reflect.Type

func (JobCopyOutput) SourceTables

Source tables to copy. Structure is documented below.

func (JobCopyOutput) ToJobCopyOutput

func (o JobCopyOutput) ToJobCopyOutput() JobCopyOutput

func (JobCopyOutput) ToJobCopyOutputWithContext

func (o JobCopyOutput) ToJobCopyOutputWithContext(ctx context.Context) JobCopyOutput

func (JobCopyOutput) ToJobCopyPtrOutput

func (o JobCopyOutput) ToJobCopyPtrOutput() JobCopyPtrOutput

func (JobCopyOutput) ToJobCopyPtrOutputWithContext

func (o JobCopyOutput) ToJobCopyPtrOutputWithContext(ctx context.Context) JobCopyPtrOutput

func (JobCopyOutput) WriteDisposition

func (o JobCopyOutput) WriteDisposition() pulumi.StringPtrOutput

Specifies the action that occurs if the destination table already exists. The following values are supported: WRITE_TRUNCATE: If the table already exists, BigQuery overwrites the table data and uses the schema from the query result. WRITE_APPEND: If the table already exists, BigQuery appends the data to the table. WRITE_EMPTY: If the table already exists and contains data, a 'duplicate' error is returned in the job result. Each action is atomic and only occurs if BigQuery is able to complete the job successfully. Creation, truncation and append actions occur as one atomic update upon job completion. Default value is `WRITE_EMPTY`. Possible values are `WRITE_TRUNCATE`, `WRITE_APPEND`, and `WRITE_EMPTY`.

type JobCopyPtrInput

type JobCopyPtrInput interface {
	pulumi.Input

	ToJobCopyPtrOutput() JobCopyPtrOutput
	ToJobCopyPtrOutputWithContext(context.Context) JobCopyPtrOutput
}

JobCopyPtrInput is an input type that accepts JobCopyArgs, JobCopyPtr and JobCopyPtrOutput values. You can construct a concrete instance of `JobCopyPtrInput` via:

        JobCopyArgs{...}

or:

        nil

func JobCopyPtr

func JobCopyPtr(v *JobCopyArgs) JobCopyPtrInput

type JobCopyPtrOutput

type JobCopyPtrOutput struct{ *pulumi.OutputState }

func (JobCopyPtrOutput) CreateDisposition

func (o JobCopyPtrOutput) CreateDisposition() pulumi.StringPtrOutput

Specifies whether the job is allowed to create new tables. The following values are supported: CREATE_IF_NEEDED: If the table does not exist, BigQuery creates the table. CREATE_NEVER: The table must already exist. If it does not, a 'notFound' error is returned in the job result. Creation, truncation and append actions occur as one atomic update upon job completion Default value is `CREATE_IF_NEEDED`. Possible values are `CREATE_IF_NEEDED` and `CREATE_NEVER`.

func (JobCopyPtrOutput) DestinationEncryptionConfiguration

func (o JobCopyPtrOutput) DestinationEncryptionConfiguration() JobCopyDestinationEncryptionConfigurationPtrOutput

Custom encryption configuration (e.g., Cloud KMS keys) Structure is documented below.

func (JobCopyPtrOutput) DestinationTable

The destination table. Structure is documented below.

func (JobCopyPtrOutput) Elem

func (JobCopyPtrOutput) ElementType

func (JobCopyPtrOutput) ElementType() reflect.Type

func (JobCopyPtrOutput) SourceTables

Source tables to copy. Structure is documented below.

func (JobCopyPtrOutput) ToJobCopyPtrOutput

func (o JobCopyPtrOutput) ToJobCopyPtrOutput() JobCopyPtrOutput

func (JobCopyPtrOutput) ToJobCopyPtrOutputWithContext

func (o JobCopyPtrOutput) ToJobCopyPtrOutputWithContext(ctx context.Context) JobCopyPtrOutput

func (JobCopyPtrOutput) WriteDisposition

func (o JobCopyPtrOutput) WriteDisposition() pulumi.StringPtrOutput

Specifies the action that occurs if the destination table already exists. The following values are supported: WRITE_TRUNCATE: If the table already exists, BigQuery overwrites the table data and uses the schema from the query result. WRITE_APPEND: If the table already exists, BigQuery appends the data to the table. WRITE_EMPTY: If the table already exists and contains data, a 'duplicate' error is returned in the job result. Each action is atomic and only occurs if BigQuery is able to complete the job successfully. Creation, truncation and append actions occur as one atomic update upon job completion. Default value is `WRITE_EMPTY`. Possible values are `WRITE_TRUNCATE`, `WRITE_APPEND`, and `WRITE_EMPTY`.

type JobCopySourceTable

type JobCopySourceTable struct {
	// The ID of the dataset containing this model.
	DatasetId *string `pulumi:"datasetId"`
	// The ID of the project containing this model.
	ProjectId *string `pulumi:"projectId"`
	// The table. Can be specified `{{table_id}}` if `projectId` and `datasetId` are also set,
	// or of the form `projects/{{project}}/datasets/{{dataset_id}}/tables/{{table_id}}` if not.
	TableId string `pulumi:"tableId"`
}

type JobCopySourceTableArgs

type JobCopySourceTableArgs struct {
	// The ID of the dataset containing this model.
	DatasetId pulumi.StringPtrInput `pulumi:"datasetId"`
	// The ID of the project containing this model.
	ProjectId pulumi.StringPtrInput `pulumi:"projectId"`
	// The table. Can be specified `{{table_id}}` if `projectId` and `datasetId` are also set,
	// or of the form `projects/{{project}}/datasets/{{dataset_id}}/tables/{{table_id}}` if not.
	TableId pulumi.StringInput `pulumi:"tableId"`
}

func (JobCopySourceTableArgs) ElementType

func (JobCopySourceTableArgs) ElementType() reflect.Type

func (JobCopySourceTableArgs) ToJobCopySourceTableOutput

func (i JobCopySourceTableArgs) ToJobCopySourceTableOutput() JobCopySourceTableOutput

func (JobCopySourceTableArgs) ToJobCopySourceTableOutputWithContext

func (i JobCopySourceTableArgs) ToJobCopySourceTableOutputWithContext(ctx context.Context) JobCopySourceTableOutput

type JobCopySourceTableArray

type JobCopySourceTableArray []JobCopySourceTableInput

func (JobCopySourceTableArray) ElementType

func (JobCopySourceTableArray) ElementType() reflect.Type

func (JobCopySourceTableArray) ToJobCopySourceTableArrayOutput

func (i JobCopySourceTableArray) ToJobCopySourceTableArrayOutput() JobCopySourceTableArrayOutput

func (JobCopySourceTableArray) ToJobCopySourceTableArrayOutputWithContext

func (i JobCopySourceTableArray) ToJobCopySourceTableArrayOutputWithContext(ctx context.Context) JobCopySourceTableArrayOutput

type JobCopySourceTableArrayInput

type JobCopySourceTableArrayInput interface {
	pulumi.Input

	ToJobCopySourceTableArrayOutput() JobCopySourceTableArrayOutput
	ToJobCopySourceTableArrayOutputWithContext(context.Context) JobCopySourceTableArrayOutput
}

JobCopySourceTableArrayInput is an input type that accepts JobCopySourceTableArray and JobCopySourceTableArrayOutput values. You can construct a concrete instance of `JobCopySourceTableArrayInput` via:

JobCopySourceTableArray{ JobCopySourceTableArgs{...} }

type JobCopySourceTableArrayOutput

type JobCopySourceTableArrayOutput struct{ *pulumi.OutputState }

func (JobCopySourceTableArrayOutput) ElementType

func (JobCopySourceTableArrayOutput) Index

func (JobCopySourceTableArrayOutput) ToJobCopySourceTableArrayOutput

func (o JobCopySourceTableArrayOutput) ToJobCopySourceTableArrayOutput() JobCopySourceTableArrayOutput

func (JobCopySourceTableArrayOutput) ToJobCopySourceTableArrayOutputWithContext

func (o JobCopySourceTableArrayOutput) ToJobCopySourceTableArrayOutputWithContext(ctx context.Context) JobCopySourceTableArrayOutput

type JobCopySourceTableInput

type JobCopySourceTableInput interface {
	pulumi.Input

	ToJobCopySourceTableOutput() JobCopySourceTableOutput
	ToJobCopySourceTableOutputWithContext(context.Context) JobCopySourceTableOutput
}

JobCopySourceTableInput is an input type that accepts JobCopySourceTableArgs and JobCopySourceTableOutput values. You can construct a concrete instance of `JobCopySourceTableInput` via:

JobCopySourceTableArgs{...}

type JobCopySourceTableOutput

type JobCopySourceTableOutput struct{ *pulumi.OutputState }

func (JobCopySourceTableOutput) DatasetId

The ID of the dataset containing this model.

func (JobCopySourceTableOutput) ElementType

func (JobCopySourceTableOutput) ElementType() reflect.Type

func (JobCopySourceTableOutput) ProjectId

The ID of the project containing this model.

func (JobCopySourceTableOutput) TableId

The table. Can be specified `{{table_id}}` if `projectId` and `datasetId` are also set, or of the form `projects/{{project}}/datasets/{{dataset_id}}/tables/{{table_id}}` if not.

func (JobCopySourceTableOutput) ToJobCopySourceTableOutput

func (o JobCopySourceTableOutput) ToJobCopySourceTableOutput() JobCopySourceTableOutput

func (JobCopySourceTableOutput) ToJobCopySourceTableOutputWithContext

func (o JobCopySourceTableOutput) ToJobCopySourceTableOutputWithContext(ctx context.Context) JobCopySourceTableOutput

type JobExtract

type JobExtract struct {
	// The compression type to use for exported files. Possible values include GZIP, DEFLATE, SNAPPY, and NONE.
	// The default value is NONE. DEFLATE and SNAPPY are only supported for Avro.
	Compression *string `pulumi:"compression"`
	// The exported file format. Possible values include CSV, NEWLINE_DELIMITED_JSON and AVRO for tables and SAVED_MODEL for models.
	// The default value for tables is CSV. Tables with nested or repeated fields cannot be exported as CSV.
	// The default value for models is SAVED_MODEL.
	DestinationFormat *string `pulumi:"destinationFormat"`
	// A list of fully-qualified Google Cloud Storage URIs where the extracted table should be written.
	DestinationUris []string `pulumi:"destinationUris"`
	// When extracting data in CSV format, this defines the delimiter to use between fields in the exported data.
	// Default is ','
	FieldDelimiter *string `pulumi:"fieldDelimiter"`
	// Whether to print out a header row in the results. Default is true.
	PrintHeader *bool `pulumi:"printHeader"`
	// A reference to the model being exported.
	// Structure is documented below.
	SourceModel *JobExtractSourceModel `pulumi:"sourceModel"`
	// A reference to the table being exported.
	// Structure is documented below.
	SourceTable *JobExtractSourceTable `pulumi:"sourceTable"`
	// Whether to use logical types when extracting to AVRO format.
	UseAvroLogicalTypes *bool `pulumi:"useAvroLogicalTypes"`
}

type JobExtractArgs

type JobExtractArgs struct {
	// The compression type to use for exported files. Possible values include GZIP, DEFLATE, SNAPPY, and NONE.
	// The default value is NONE. DEFLATE and SNAPPY are only supported for Avro.
	Compression pulumi.StringPtrInput `pulumi:"compression"`
	// The exported file format. Possible values include CSV, NEWLINE_DELIMITED_JSON and AVRO for tables and SAVED_MODEL for models.
	// The default value for tables is CSV. Tables with nested or repeated fields cannot be exported as CSV.
	// The default value for models is SAVED_MODEL.
	DestinationFormat pulumi.StringPtrInput `pulumi:"destinationFormat"`
	// A list of fully-qualified Google Cloud Storage URIs where the extracted table should be written.
	DestinationUris pulumi.StringArrayInput `pulumi:"destinationUris"`
	// When extracting data in CSV format, this defines the delimiter to use between fields in the exported data.
	// Default is ','
	FieldDelimiter pulumi.StringPtrInput `pulumi:"fieldDelimiter"`
	// Whether to print out a header row in the results. Default is true.
	PrintHeader pulumi.BoolPtrInput `pulumi:"printHeader"`
	// A reference to the model being exported.
	// Structure is documented below.
	SourceModel JobExtractSourceModelPtrInput `pulumi:"sourceModel"`
	// A reference to the table being exported.
	// Structure is documented below.
	SourceTable JobExtractSourceTablePtrInput `pulumi:"sourceTable"`
	// Whether to use logical types when extracting to AVRO format.
	UseAvroLogicalTypes pulumi.BoolPtrInput `pulumi:"useAvroLogicalTypes"`
}

func (JobExtractArgs) ElementType

func (JobExtractArgs) ElementType() reflect.Type

func (JobExtractArgs) ToJobExtractOutput

func (i JobExtractArgs) ToJobExtractOutput() JobExtractOutput

func (JobExtractArgs) ToJobExtractOutputWithContext

func (i JobExtractArgs) ToJobExtractOutputWithContext(ctx context.Context) JobExtractOutput

func (JobExtractArgs) ToJobExtractPtrOutput

func (i JobExtractArgs) ToJobExtractPtrOutput() JobExtractPtrOutput

func (JobExtractArgs) ToJobExtractPtrOutputWithContext

func (i JobExtractArgs) ToJobExtractPtrOutputWithContext(ctx context.Context) JobExtractPtrOutput

type JobExtractInput

type JobExtractInput interface {
	pulumi.Input

	ToJobExtractOutput() JobExtractOutput
	ToJobExtractOutputWithContext(context.Context) JobExtractOutput
}

JobExtractInput is an input type that accepts JobExtractArgs and JobExtractOutput values. You can construct a concrete instance of `JobExtractInput` via:

JobExtractArgs{...}

type JobExtractOutput

type JobExtractOutput struct{ *pulumi.OutputState }

func (JobExtractOutput) Compression

func (o JobExtractOutput) Compression() pulumi.StringPtrOutput

The compression type to use for exported files. Possible values include GZIP, DEFLATE, SNAPPY, and NONE. The default value is NONE. DEFLATE and SNAPPY are only supported for Avro.

func (JobExtractOutput) DestinationFormat

func (o JobExtractOutput) DestinationFormat() pulumi.StringPtrOutput

The exported file format. Possible values include CSV, NEWLINE_DELIMITED_JSON and AVRO for tables and SAVED_MODEL for models. The default value for tables is CSV. Tables with nested or repeated fields cannot be exported as CSV. The default value for models is SAVED_MODEL.

func (JobExtractOutput) DestinationUris

func (o JobExtractOutput) DestinationUris() pulumi.StringArrayOutput

A list of fully-qualified Google Cloud Storage URIs where the extracted table should be written.

func (JobExtractOutput) ElementType

func (JobExtractOutput) ElementType() reflect.Type

func (JobExtractOutput) FieldDelimiter

func (o JobExtractOutput) FieldDelimiter() pulumi.StringPtrOutput

When extracting data in CSV format, this defines the delimiter to use between fields in the exported data. Default is ','

func (JobExtractOutput) PrintHeader

func (o JobExtractOutput) PrintHeader() pulumi.BoolPtrOutput

Whether to print out a header row in the results. Default is true.

func (JobExtractOutput) SourceModel

A reference to the model being exported. Structure is documented below.

func (JobExtractOutput) SourceTable

A reference to the table being exported. Structure is documented below.

func (JobExtractOutput) ToJobExtractOutput

func (o JobExtractOutput) ToJobExtractOutput() JobExtractOutput

func (JobExtractOutput) ToJobExtractOutputWithContext

func (o JobExtractOutput) ToJobExtractOutputWithContext(ctx context.Context) JobExtractOutput

func (JobExtractOutput) ToJobExtractPtrOutput

func (o JobExtractOutput) ToJobExtractPtrOutput() JobExtractPtrOutput

func (JobExtractOutput) ToJobExtractPtrOutputWithContext

func (o JobExtractOutput) ToJobExtractPtrOutputWithContext(ctx context.Context) JobExtractPtrOutput

func (JobExtractOutput) UseAvroLogicalTypes

func (o JobExtractOutput) UseAvroLogicalTypes() pulumi.BoolPtrOutput

Whether to use logical types when extracting to AVRO format.

type JobExtractPtrInput

type JobExtractPtrInput interface {
	pulumi.Input

	ToJobExtractPtrOutput() JobExtractPtrOutput
	ToJobExtractPtrOutputWithContext(context.Context) JobExtractPtrOutput
}

JobExtractPtrInput is an input type that accepts JobExtractArgs, JobExtractPtr and JobExtractPtrOutput values. You can construct a concrete instance of `JobExtractPtrInput` via:

        JobExtractArgs{...}

or:

        nil

func JobExtractPtr

func JobExtractPtr(v *JobExtractArgs) JobExtractPtrInput

type JobExtractPtrOutput

type JobExtractPtrOutput struct{ *pulumi.OutputState }

func (JobExtractPtrOutput) Compression

func (o JobExtractPtrOutput) Compression() pulumi.StringPtrOutput

The compression type to use for exported files. Possible values include GZIP, DEFLATE, SNAPPY, and NONE. The default value is NONE. DEFLATE and SNAPPY are only supported for Avro.

func (JobExtractPtrOutput) DestinationFormat

func (o JobExtractPtrOutput) DestinationFormat() pulumi.StringPtrOutput

The exported file format. Possible values include CSV, NEWLINE_DELIMITED_JSON and AVRO for tables and SAVED_MODEL for models. The default value for tables is CSV. Tables with nested or repeated fields cannot be exported as CSV. The default value for models is SAVED_MODEL.

func (JobExtractPtrOutput) DestinationUris

func (o JobExtractPtrOutput) DestinationUris() pulumi.StringArrayOutput

A list of fully-qualified Google Cloud Storage URIs where the extracted table should be written.

func (JobExtractPtrOutput) Elem

func (JobExtractPtrOutput) ElementType

func (JobExtractPtrOutput) ElementType() reflect.Type

func (JobExtractPtrOutput) FieldDelimiter

func (o JobExtractPtrOutput) FieldDelimiter() pulumi.StringPtrOutput

When extracting data in CSV format, this defines the delimiter to use between fields in the exported data. Default is ','

func (JobExtractPtrOutput) PrintHeader

func (o JobExtractPtrOutput) PrintHeader() pulumi.BoolPtrOutput

Whether to print out a header row in the results. Default is true.

func (JobExtractPtrOutput) SourceModel

A reference to the model being exported. Structure is documented below.

func (JobExtractPtrOutput) SourceTable

A reference to the table being exported. Structure is documented below.

func (JobExtractPtrOutput) ToJobExtractPtrOutput

func (o JobExtractPtrOutput) ToJobExtractPtrOutput() JobExtractPtrOutput

func (JobExtractPtrOutput) ToJobExtractPtrOutputWithContext

func (o JobExtractPtrOutput) ToJobExtractPtrOutputWithContext(ctx context.Context) JobExtractPtrOutput

func (JobExtractPtrOutput) UseAvroLogicalTypes

func (o JobExtractPtrOutput) UseAvroLogicalTypes() pulumi.BoolPtrOutput

Whether to use logical types when extracting to AVRO format.

type JobExtractSourceModel

type JobExtractSourceModel struct {
	// The ID of the dataset containing this model.
	DatasetId string `pulumi:"datasetId"`
	// The ID of the model.
	ModelId string `pulumi:"modelId"`
	// The ID of the project containing this model.
	ProjectId string `pulumi:"projectId"`
}

type JobExtractSourceModelArgs

type JobExtractSourceModelArgs struct {
	// The ID of the dataset containing this model.
	DatasetId pulumi.StringInput `pulumi:"datasetId"`
	// The ID of the model.
	ModelId pulumi.StringInput `pulumi:"modelId"`
	// The ID of the project containing this model.
	ProjectId pulumi.StringInput `pulumi:"projectId"`
}

func (JobExtractSourceModelArgs) ElementType

func (JobExtractSourceModelArgs) ElementType() reflect.Type

func (JobExtractSourceModelArgs) ToJobExtractSourceModelOutput

func (i JobExtractSourceModelArgs) ToJobExtractSourceModelOutput() JobExtractSourceModelOutput

func (JobExtractSourceModelArgs) ToJobExtractSourceModelOutputWithContext

func (i JobExtractSourceModelArgs) ToJobExtractSourceModelOutputWithContext(ctx context.Context) JobExtractSourceModelOutput

func (JobExtractSourceModelArgs) ToJobExtractSourceModelPtrOutput

func (i JobExtractSourceModelArgs) ToJobExtractSourceModelPtrOutput() JobExtractSourceModelPtrOutput

func (JobExtractSourceModelArgs) ToJobExtractSourceModelPtrOutputWithContext

func (i JobExtractSourceModelArgs) ToJobExtractSourceModelPtrOutputWithContext(ctx context.Context) JobExtractSourceModelPtrOutput

type JobExtractSourceModelInput

type JobExtractSourceModelInput interface {
	pulumi.Input

	ToJobExtractSourceModelOutput() JobExtractSourceModelOutput
	ToJobExtractSourceModelOutputWithContext(context.Context) JobExtractSourceModelOutput
}

JobExtractSourceModelInput is an input type that accepts JobExtractSourceModelArgs and JobExtractSourceModelOutput values. You can construct a concrete instance of `JobExtractSourceModelInput` via:

JobExtractSourceModelArgs{...}

type JobExtractSourceModelOutput

type JobExtractSourceModelOutput struct{ *pulumi.OutputState }

func (JobExtractSourceModelOutput) DatasetId

The ID of the dataset containing this model.

func (JobExtractSourceModelOutput) ElementType

func (JobExtractSourceModelOutput) ModelId

The ID of the model.

func (JobExtractSourceModelOutput) ProjectId

The ID of the project containing this model.

func (JobExtractSourceModelOutput) ToJobExtractSourceModelOutput

func (o JobExtractSourceModelOutput) ToJobExtractSourceModelOutput() JobExtractSourceModelOutput

func (JobExtractSourceModelOutput) ToJobExtractSourceModelOutputWithContext

func (o JobExtractSourceModelOutput) ToJobExtractSourceModelOutputWithContext(ctx context.Context) JobExtractSourceModelOutput

func (JobExtractSourceModelOutput) ToJobExtractSourceModelPtrOutput

func (o JobExtractSourceModelOutput) ToJobExtractSourceModelPtrOutput() JobExtractSourceModelPtrOutput

func (JobExtractSourceModelOutput) ToJobExtractSourceModelPtrOutputWithContext

func (o JobExtractSourceModelOutput) ToJobExtractSourceModelPtrOutputWithContext(ctx context.Context) JobExtractSourceModelPtrOutput

type JobExtractSourceModelPtrInput

type JobExtractSourceModelPtrInput interface {
	pulumi.Input

	ToJobExtractSourceModelPtrOutput() JobExtractSourceModelPtrOutput
	ToJobExtractSourceModelPtrOutputWithContext(context.Context) JobExtractSourceModelPtrOutput
}

JobExtractSourceModelPtrInput is an input type that accepts JobExtractSourceModelArgs, JobExtractSourceModelPtr and JobExtractSourceModelPtrOutput values. You can construct a concrete instance of `JobExtractSourceModelPtrInput` via:

        JobExtractSourceModelArgs{...}

or:

        nil

type JobExtractSourceModelPtrOutput

type JobExtractSourceModelPtrOutput struct{ *pulumi.OutputState }

func (JobExtractSourceModelPtrOutput) DatasetId

The ID of the dataset containing this model.

func (JobExtractSourceModelPtrOutput) Elem

func (JobExtractSourceModelPtrOutput) ElementType

func (JobExtractSourceModelPtrOutput) ModelId

The ID of the model.

func (JobExtractSourceModelPtrOutput) ProjectId

The ID of the project containing this model.

func (JobExtractSourceModelPtrOutput) ToJobExtractSourceModelPtrOutput

func (o JobExtractSourceModelPtrOutput) ToJobExtractSourceModelPtrOutput() JobExtractSourceModelPtrOutput

func (JobExtractSourceModelPtrOutput) ToJobExtractSourceModelPtrOutputWithContext

func (o JobExtractSourceModelPtrOutput) ToJobExtractSourceModelPtrOutputWithContext(ctx context.Context) JobExtractSourceModelPtrOutput

type JobExtractSourceTable

type JobExtractSourceTable struct {
	// The ID of the dataset containing this model.
	DatasetId *string `pulumi:"datasetId"`
	// The ID of the project containing this model.
	ProjectId *string `pulumi:"projectId"`
	// The table. Can be specified `{{table_id}}` if `projectId` and `datasetId` are also set,
	// or of the form `projects/{{project}}/datasets/{{dataset_id}}/tables/{{table_id}}` if not.
	TableId string `pulumi:"tableId"`
}

type JobExtractSourceTableArgs

type JobExtractSourceTableArgs struct {
	// The ID of the dataset containing this model.
	DatasetId pulumi.StringPtrInput `pulumi:"datasetId"`
	// The ID of the project containing this model.
	ProjectId pulumi.StringPtrInput `pulumi:"projectId"`
	// The table. Can be specified `{{table_id}}` if `projectId` and `datasetId` are also set,
	// or of the form `projects/{{project}}/datasets/{{dataset_id}}/tables/{{table_id}}` if not.
	TableId pulumi.StringInput `pulumi:"tableId"`
}

func (JobExtractSourceTableArgs) ElementType

func (JobExtractSourceTableArgs) ElementType() reflect.Type

func (JobExtractSourceTableArgs) ToJobExtractSourceTableOutput

func (i JobExtractSourceTableArgs) ToJobExtractSourceTableOutput() JobExtractSourceTableOutput

func (JobExtractSourceTableArgs) ToJobExtractSourceTableOutputWithContext

func (i JobExtractSourceTableArgs) ToJobExtractSourceTableOutputWithContext(ctx context.Context) JobExtractSourceTableOutput

func (JobExtractSourceTableArgs) ToJobExtractSourceTablePtrOutput

func (i JobExtractSourceTableArgs) ToJobExtractSourceTablePtrOutput() JobExtractSourceTablePtrOutput

func (JobExtractSourceTableArgs) ToJobExtractSourceTablePtrOutputWithContext

func (i JobExtractSourceTableArgs) ToJobExtractSourceTablePtrOutputWithContext(ctx context.Context) JobExtractSourceTablePtrOutput

type JobExtractSourceTableInput

type JobExtractSourceTableInput interface {
	pulumi.Input

	ToJobExtractSourceTableOutput() JobExtractSourceTableOutput
	ToJobExtractSourceTableOutputWithContext(context.Context) JobExtractSourceTableOutput
}

JobExtractSourceTableInput is an input type that accepts JobExtractSourceTableArgs and JobExtractSourceTableOutput values. You can construct a concrete instance of `JobExtractSourceTableInput` via:

JobExtractSourceTableArgs{...}

type JobExtractSourceTableOutput

type JobExtractSourceTableOutput struct{ *pulumi.OutputState }

func (JobExtractSourceTableOutput) DatasetId

The ID of the dataset containing this model.

func (JobExtractSourceTableOutput) ElementType

func (JobExtractSourceTableOutput) ProjectId

The ID of the project containing this model.

func (JobExtractSourceTableOutput) TableId

The table. Can be specified `{{table_id}}` if `projectId` and `datasetId` are also set, or of the form `projects/{{project}}/datasets/{{dataset_id}}/tables/{{table_id}}` if not.

func (JobExtractSourceTableOutput) ToJobExtractSourceTableOutput

func (o JobExtractSourceTableOutput) ToJobExtractSourceTableOutput() JobExtractSourceTableOutput

func (JobExtractSourceTableOutput) ToJobExtractSourceTableOutputWithContext

func (o JobExtractSourceTableOutput) ToJobExtractSourceTableOutputWithContext(ctx context.Context) JobExtractSourceTableOutput

func (JobExtractSourceTableOutput) ToJobExtractSourceTablePtrOutput

func (o JobExtractSourceTableOutput) ToJobExtractSourceTablePtrOutput() JobExtractSourceTablePtrOutput

func (JobExtractSourceTableOutput) ToJobExtractSourceTablePtrOutputWithContext

func (o JobExtractSourceTableOutput) ToJobExtractSourceTablePtrOutputWithContext(ctx context.Context) JobExtractSourceTablePtrOutput

type JobExtractSourceTablePtrInput

type JobExtractSourceTablePtrInput interface {
	pulumi.Input

	ToJobExtractSourceTablePtrOutput() JobExtractSourceTablePtrOutput
	ToJobExtractSourceTablePtrOutputWithContext(context.Context) JobExtractSourceTablePtrOutput
}

JobExtractSourceTablePtrInput is an input type that accepts JobExtractSourceTableArgs, JobExtractSourceTablePtr and JobExtractSourceTablePtrOutput values. You can construct a concrete instance of `JobExtractSourceTablePtrInput` via:

        JobExtractSourceTableArgs{...}

or:

        nil

type JobExtractSourceTablePtrOutput

type JobExtractSourceTablePtrOutput struct{ *pulumi.OutputState }

func (JobExtractSourceTablePtrOutput) DatasetId

The ID of the dataset containing this model.

func (JobExtractSourceTablePtrOutput) Elem

func (JobExtractSourceTablePtrOutput) ElementType

func (JobExtractSourceTablePtrOutput) ProjectId

The ID of the project containing this model.

func (JobExtractSourceTablePtrOutput) TableId

The table. Can be specified `{{table_id}}` if `projectId` and `datasetId` are also set, or of the form `projects/{{project}}/datasets/{{dataset_id}}/tables/{{table_id}}` if not.

func (JobExtractSourceTablePtrOutput) ToJobExtractSourceTablePtrOutput

func (o JobExtractSourceTablePtrOutput) ToJobExtractSourceTablePtrOutput() JobExtractSourceTablePtrOutput

func (JobExtractSourceTablePtrOutput) ToJobExtractSourceTablePtrOutputWithContext

func (o JobExtractSourceTablePtrOutput) ToJobExtractSourceTablePtrOutputWithContext(ctx context.Context) JobExtractSourceTablePtrOutput

type JobLoad

type JobLoad struct {
	// Accept rows that are missing trailing optional columns. The missing values are treated as nulls.
	// If false, records with missing trailing columns are treated as bad records, and if there are too many bad records,
	// an invalid error is returned in the job result. The default value is false. Only applicable to CSV, ignored for other formats.
	AllowJaggedRows *bool `pulumi:"allowJaggedRows"`
	// Indicates if BigQuery should allow quoted data sections that contain newline characters in a CSV file.
	// The default value is false.
	AllowQuotedNewlines *bool `pulumi:"allowQuotedNewlines"`
	// Indicates if we should automatically infer the options and schema for CSV and JSON sources.
	Autodetect *bool `pulumi:"autodetect"`
	// Specifies whether the job is allowed to create new tables. The following values are supported:
	// CREATE_IF_NEEDED: If the table does not exist, BigQuery creates the table.
	// CREATE_NEVER: The table must already exist. If it does not, a 'notFound' error is returned in the job result.
	// Creation, truncation and append actions occur as one atomic update upon job completion
	// Default value is `CREATE_IF_NEEDED`.
	// Possible values are `CREATE_IF_NEEDED` and `CREATE_NEVER`.
	CreateDisposition *string `pulumi:"createDisposition"`
	// Custom encryption configuration (e.g., Cloud KMS keys)
	// Structure is documented below.
	DestinationEncryptionConfiguration *JobLoadDestinationEncryptionConfiguration `pulumi:"destinationEncryptionConfiguration"`
	// The destination table.
	// Structure is documented below.
	DestinationTable JobLoadDestinationTable `pulumi:"destinationTable"`
	// The character encoding of the data. The supported values are UTF-8 or ISO-8859-1.
	// The default value is UTF-8. BigQuery decodes the data after the raw, binary data
	// has been split using the values of the quote and fieldDelimiter properties.
	Encoding *string `pulumi:"encoding"`
	// When extracting data in CSV format, this defines the delimiter to use between fields in the exported data.
	// Default is ','
	FieldDelimiter *string `pulumi:"fieldDelimiter"`
	// Indicates if BigQuery should allow extra values that are not represented in the table schema.
	// If true, the extra values are ignored. If false, records with extra columns are treated as bad records,
	// and if there are too many bad records, an invalid error is returned in the job result.
	// The default value is false. The sourceFormat property determines what BigQuery treats as an extra value:
	// CSV: Trailing columns
	// JSON: Named values that don't match any column names
	IgnoreUnknownValues *bool `pulumi:"ignoreUnknownValues"`
	// The maximum number of bad records that BigQuery can ignore when running the job. If the number of bad records exceeds this value,
	// an invalid error is returned in the job result. The default value is 0, which requires that all records are valid.
	MaxBadRecords *int `pulumi:"maxBadRecords"`
	// Specifies a string that represents a null value in a CSV file. The default value is the empty string. If you set this
	// property to a custom value, BigQuery throws an error if an
	// empty string is present for all data types except for STRING and BYTE. For STRING and BYTE columns, BigQuery interprets the empty string as
	// an empty value.
	NullMarker *string `pulumi:"nullMarker"`
	// If sourceFormat is set to "DATASTORE_BACKUP", indicates which entity properties to load into BigQuery from a Cloud Datastore backup.
	// Property names are case sensitive and must be top-level properties. If no properties are specified, BigQuery loads all properties.
	// If any named property isn't found in the Cloud Datastore backup, an invalid error is returned in the job result.
	ProjectionFields []string `pulumi:"projectionFields"`
	// The value that is used to quote data sections in a CSV file. BigQuery converts the string to ISO-8859-1 encoding,
	// and then uses the first byte of the encoded string to split the data in its raw, binary state.
	// The default value is a double-quote ('"'). If your data does not contain quoted sections, set the property value to an empty string.
	// If your data contains quoted newline characters, you must also set the allowQuotedNewlines property to true.
	Quote *string `pulumi:"quote"`
	// Allows the schema of the destination table to be updated as a side effect of the load job if a schema is autodetected or
	// supplied in the job configuration. Schema update options are supported in two cases: when writeDisposition is WRITE_APPEND;
	// when writeDisposition is WRITE_TRUNCATE and the destination table is a partition of a table, specified by partition decorators.
	// For normal tables, WRITE_TRUNCATE will always overwrite the schema. One or more of the following values are specified:
	// ALLOW_FIELD_ADDITION: allow adding a nullable field to the schema.
	// ALLOW_FIELD_RELAXATION: allow relaxing a required field in the original schema to nullable.
	SchemaUpdateOptions []string `pulumi:"schemaUpdateOptions"`
	// The number of rows at the top of a CSV file that BigQuery will skip when loading the data.
	// The default value is 0. This property is useful if you have header rows in the file that should be skipped.
	// When autodetect is on, the behavior is the following:
	// skipLeadingRows unspecified - Autodetect tries to detect headers in the first row. If they are not detected,
	// the row is read as data. Otherwise data is read starting from the second row.
	// skipLeadingRows is 0 - Instructs autodetect that there are no headers and data should be read starting from the first row.
	// skipLeadingRows = N > 0 - Autodetect skips N-1 rows and tries to detect headers in row N. If headers are not detected,
	// row N is just skipped. Otherwise row N is used to extract column names for the detected schema.
	SkipLeadingRows *int `pulumi:"skipLeadingRows"`
	// The format of the data files. For CSV files, specify "CSV". For datastore backups, specify "DATASTORE_BACKUP".
	// For newline-delimited JSON, specify "NEWLINE_DELIMITED_JSON". For Avro, specify "AVRO". For parquet, specify "PARQUET".
	// For orc, specify "ORC". The default value is CSV.
	SourceFormat *string `pulumi:"sourceFormat"`
	// The fully-qualified URIs that point to your data in Google Cloud.
	// For Google Cloud Storage URIs: Each URI can contain one '*' wildcard character
	// and it must come after the 'bucket' name. Size limits related to load jobs apply
	// to external data sources. For Google Cloud Bigtable URIs: Exactly one URI can be
	// specified and it has be a fully specified and valid HTTPS URL for a Google Cloud Bigtable table.
	// For Google Cloud Datastore backups: Exactly one URI can be specified. Also, the '*' wildcard character is not allowed.
	SourceUris []string `pulumi:"sourceUris"`
	// Time-based partitioning specification for the destination table.
	// Structure is documented below.
	TimePartitioning *JobLoadTimePartitioning `pulumi:"timePartitioning"`
	// Specifies the action that occurs if the destination table already exists. The following values are supported:
	// WRITE_TRUNCATE: If the table already exists, BigQuery overwrites the table data and uses the schema from the query result.
	// WRITE_APPEND: If the table already exists, BigQuery appends the data to the table.
	// WRITE_EMPTY: If the table already exists and contains data, a 'duplicate' error is returned in the job result.
	// Each action is atomic and only occurs if BigQuery is able to complete the job successfully.
	// Creation, truncation and append actions occur as one atomic update upon job completion.
	// Default value is `WRITE_EMPTY`.
	// Possible values are `WRITE_TRUNCATE`, `WRITE_APPEND`, and `WRITE_EMPTY`.
	WriteDisposition *string `pulumi:"writeDisposition"`
}

type JobLoadArgs

type JobLoadArgs struct {
	// Accept rows that are missing trailing optional columns. The missing values are treated as nulls.
	// If false, records with missing trailing columns are treated as bad records, and if there are too many bad records,
	// an invalid error is returned in the job result. The default value is false. Only applicable to CSV, ignored for other formats.
	AllowJaggedRows pulumi.BoolPtrInput `pulumi:"allowJaggedRows"`
	// Indicates if BigQuery should allow quoted data sections that contain newline characters in a CSV file.
	// The default value is false.
	AllowQuotedNewlines pulumi.BoolPtrInput `pulumi:"allowQuotedNewlines"`
	// Indicates if we should automatically infer the options and schema for CSV and JSON sources.
	Autodetect pulumi.BoolPtrInput `pulumi:"autodetect"`
	// Specifies whether the job is allowed to create new tables. The following values are supported:
	// CREATE_IF_NEEDED: If the table does not exist, BigQuery creates the table.
	// CREATE_NEVER: The table must already exist. If it does not, a 'notFound' error is returned in the job result.
	// Creation, truncation and append actions occur as one atomic update upon job completion
	// Default value is `CREATE_IF_NEEDED`.
	// Possible values are `CREATE_IF_NEEDED` and `CREATE_NEVER`.
	CreateDisposition pulumi.StringPtrInput `pulumi:"createDisposition"`
	// Custom encryption configuration (e.g., Cloud KMS keys)
	// Structure is documented below.
	DestinationEncryptionConfiguration JobLoadDestinationEncryptionConfigurationPtrInput `pulumi:"destinationEncryptionConfiguration"`
	// The destination table.
	// Structure is documented below.
	DestinationTable JobLoadDestinationTableInput `pulumi:"destinationTable"`
	// The character encoding of the data. The supported values are UTF-8 or ISO-8859-1.
	// The default value is UTF-8. BigQuery decodes the data after the raw, binary data
	// has been split using the values of the quote and fieldDelimiter properties.
	Encoding pulumi.StringPtrInput `pulumi:"encoding"`
	// When extracting data in CSV format, this defines the delimiter to use between fields in the exported data.
	// Default is ','
	FieldDelimiter pulumi.StringPtrInput `pulumi:"fieldDelimiter"`
	// Indicates if BigQuery should allow extra values that are not represented in the table schema.
	// If true, the extra values are ignored. If false, records with extra columns are treated as bad records,
	// and if there are too many bad records, an invalid error is returned in the job result.
	// The default value is false. The sourceFormat property determines what BigQuery treats as an extra value:
	// CSV: Trailing columns
	// JSON: Named values that don't match any column names
	IgnoreUnknownValues pulumi.BoolPtrInput `pulumi:"ignoreUnknownValues"`
	// The maximum number of bad records that BigQuery can ignore when running the job. If the number of bad records exceeds this value,
	// an invalid error is returned in the job result. The default value is 0, which requires that all records are valid.
	MaxBadRecords pulumi.IntPtrInput `pulumi:"maxBadRecords"`
	// Specifies a string that represents a null value in a CSV file. The default value is the empty string. If you set this
	// property to a custom value, BigQuery throws an error if an
	// empty string is present for all data types except for STRING and BYTE. For STRING and BYTE columns, BigQuery interprets the empty string as
	// an empty value.
	NullMarker pulumi.StringPtrInput `pulumi:"nullMarker"`
	// If sourceFormat is set to "DATASTORE_BACKUP", indicates which entity properties to load into BigQuery from a Cloud Datastore backup.
	// Property names are case sensitive and must be top-level properties. If no properties are specified, BigQuery loads all properties.
	// If any named property isn't found in the Cloud Datastore backup, an invalid error is returned in the job result.
	ProjectionFields pulumi.StringArrayInput `pulumi:"projectionFields"`
	// The value that is used to quote data sections in a CSV file. BigQuery converts the string to ISO-8859-1 encoding,
	// and then uses the first byte of the encoded string to split the data in its raw, binary state.
	// The default value is a double-quote ('"'). If your data does not contain quoted sections, set the property value to an empty string.
	// If your data contains quoted newline characters, you must also set the allowQuotedNewlines property to true.
	Quote pulumi.StringPtrInput `pulumi:"quote"`
	// Allows the schema of the destination table to be updated as a side effect of the load job if a schema is autodetected or
	// supplied in the job configuration. Schema update options are supported in two cases: when writeDisposition is WRITE_APPEND;
	// when writeDisposition is WRITE_TRUNCATE and the destination table is a partition of a table, specified by partition decorators.
	// For normal tables, WRITE_TRUNCATE will always overwrite the schema. One or more of the following values are specified:
	// ALLOW_FIELD_ADDITION: allow adding a nullable field to the schema.
	// ALLOW_FIELD_RELAXATION: allow relaxing a required field in the original schema to nullable.
	SchemaUpdateOptions pulumi.StringArrayInput `pulumi:"schemaUpdateOptions"`
	// The number of rows at the top of a CSV file that BigQuery will skip when loading the data.
	// The default value is 0. This property is useful if you have header rows in the file that should be skipped.
	// When autodetect is on, the behavior is the following:
	// skipLeadingRows unspecified - Autodetect tries to detect headers in the first row. If they are not detected,
	// the row is read as data. Otherwise data is read starting from the second row.
	// skipLeadingRows is 0 - Instructs autodetect that there are no headers and data should be read starting from the first row.
	// skipLeadingRows = N > 0 - Autodetect skips N-1 rows and tries to detect headers in row N. If headers are not detected,
	// row N is just skipped. Otherwise row N is used to extract column names for the detected schema.
	SkipLeadingRows pulumi.IntPtrInput `pulumi:"skipLeadingRows"`
	// The format of the data files. For CSV files, specify "CSV". For datastore backups, specify "DATASTORE_BACKUP".
	// For newline-delimited JSON, specify "NEWLINE_DELIMITED_JSON". For Avro, specify "AVRO". For parquet, specify "PARQUET".
	// For orc, specify "ORC". The default value is CSV.
	SourceFormat pulumi.StringPtrInput `pulumi:"sourceFormat"`
	// The fully-qualified URIs that point to your data in Google Cloud.
	// For Google Cloud Storage URIs: Each URI can contain one '*' wildcard character
	// and it must come after the 'bucket' name. Size limits related to load jobs apply
	// to external data sources. For Google Cloud Bigtable URIs: Exactly one URI can be
	// specified and it has be a fully specified and valid HTTPS URL for a Google Cloud Bigtable table.
	// For Google Cloud Datastore backups: Exactly one URI can be specified. Also, the '*' wildcard character is not allowed.
	SourceUris pulumi.StringArrayInput `pulumi:"sourceUris"`
	// Time-based partitioning specification for the destination table.
	// Structure is documented below.
	TimePartitioning JobLoadTimePartitioningPtrInput `pulumi:"timePartitioning"`
	// Specifies the action that occurs if the destination table already exists. The following values are supported:
	// WRITE_TRUNCATE: If the table already exists, BigQuery overwrites the table data and uses the schema from the query result.
	// WRITE_APPEND: If the table already exists, BigQuery appends the data to the table.
	// WRITE_EMPTY: If the table already exists and contains data, a 'duplicate' error is returned in the job result.
	// Each action is atomic and only occurs if BigQuery is able to complete the job successfully.
	// Creation, truncation and append actions occur as one atomic update upon job completion.
	// Default value is `WRITE_EMPTY`.
	// Possible values are `WRITE_TRUNCATE`, `WRITE_APPEND`, and `WRITE_EMPTY`.
	WriteDisposition pulumi.StringPtrInput `pulumi:"writeDisposition"`
}

func (JobLoadArgs) ElementType

func (JobLoadArgs) ElementType() reflect.Type

func (JobLoadArgs) ToJobLoadOutput

func (i JobLoadArgs) ToJobLoadOutput() JobLoadOutput

func (JobLoadArgs) ToJobLoadOutputWithContext

func (i JobLoadArgs) ToJobLoadOutputWithContext(ctx context.Context) JobLoadOutput

func (JobLoadArgs) ToJobLoadPtrOutput

func (i JobLoadArgs) ToJobLoadPtrOutput() JobLoadPtrOutput

func (JobLoadArgs) ToJobLoadPtrOutputWithContext

func (i JobLoadArgs) ToJobLoadPtrOutputWithContext(ctx context.Context) JobLoadPtrOutput

type JobLoadDestinationEncryptionConfiguration

type JobLoadDestinationEncryptionConfiguration struct {
	// Describes the Cloud KMS encryption key that will be used to protect destination BigQuery table.
	// The BigQuery Service Account associated with your project requires access to this encryption key.
	KmsKeyName string `pulumi:"kmsKeyName"`
}

type JobLoadDestinationEncryptionConfigurationArgs

type JobLoadDestinationEncryptionConfigurationArgs struct {
	// Describes the Cloud KMS encryption key that will be used to protect destination BigQuery table.
	// The BigQuery Service Account associated with your project requires access to this encryption key.
	KmsKeyName pulumi.StringInput `pulumi:"kmsKeyName"`
}

func (JobLoadDestinationEncryptionConfigurationArgs) ElementType

func (JobLoadDestinationEncryptionConfigurationArgs) ToJobLoadDestinationEncryptionConfigurationOutput

func (i JobLoadDestinationEncryptionConfigurationArgs) ToJobLoadDestinationEncryptionConfigurationOutput() JobLoadDestinationEncryptionConfigurationOutput

func (JobLoadDestinationEncryptionConfigurationArgs) ToJobLoadDestinationEncryptionConfigurationOutputWithContext

func (i JobLoadDestinationEncryptionConfigurationArgs) ToJobLoadDestinationEncryptionConfigurationOutputWithContext(ctx context.Context) JobLoadDestinationEncryptionConfigurationOutput

func (JobLoadDestinationEncryptionConfigurationArgs) ToJobLoadDestinationEncryptionConfigurationPtrOutput

func (i JobLoadDestinationEncryptionConfigurationArgs) ToJobLoadDestinationEncryptionConfigurationPtrOutput() JobLoadDestinationEncryptionConfigurationPtrOutput

func (JobLoadDestinationEncryptionConfigurationArgs) ToJobLoadDestinationEncryptionConfigurationPtrOutputWithContext

func (i JobLoadDestinationEncryptionConfigurationArgs) ToJobLoadDestinationEncryptionConfigurationPtrOutputWithContext(ctx context.Context) JobLoadDestinationEncryptionConfigurationPtrOutput

type JobLoadDestinationEncryptionConfigurationInput

type JobLoadDestinationEncryptionConfigurationInput interface {
	pulumi.Input

	ToJobLoadDestinationEncryptionConfigurationOutput() JobLoadDestinationEncryptionConfigurationOutput
	ToJobLoadDestinationEncryptionConfigurationOutputWithContext(context.Context) JobLoadDestinationEncryptionConfigurationOutput
}

JobLoadDestinationEncryptionConfigurationInput is an input type that accepts JobLoadDestinationEncryptionConfigurationArgs and JobLoadDestinationEncryptionConfigurationOutput values. You can construct a concrete instance of `JobLoadDestinationEncryptionConfigurationInput` via:

JobLoadDestinationEncryptionConfigurationArgs{...}

type JobLoadDestinationEncryptionConfigurationOutput

type JobLoadDestinationEncryptionConfigurationOutput struct{ *pulumi.OutputState }

func (JobLoadDestinationEncryptionConfigurationOutput) ElementType

func (JobLoadDestinationEncryptionConfigurationOutput) KmsKeyName

Describes the Cloud KMS encryption key that will be used to protect destination BigQuery table. The BigQuery Service Account associated with your project requires access to this encryption key.

func (JobLoadDestinationEncryptionConfigurationOutput) ToJobLoadDestinationEncryptionConfigurationOutput

func (o JobLoadDestinationEncryptionConfigurationOutput) ToJobLoadDestinationEncryptionConfigurationOutput() JobLoadDestinationEncryptionConfigurationOutput

func (JobLoadDestinationEncryptionConfigurationOutput) ToJobLoadDestinationEncryptionConfigurationOutputWithContext

func (o JobLoadDestinationEncryptionConfigurationOutput) ToJobLoadDestinationEncryptionConfigurationOutputWithContext(ctx context.Context) JobLoadDestinationEncryptionConfigurationOutput

func (JobLoadDestinationEncryptionConfigurationOutput) ToJobLoadDestinationEncryptionConfigurationPtrOutput

func (o JobLoadDestinationEncryptionConfigurationOutput) ToJobLoadDestinationEncryptionConfigurationPtrOutput() JobLoadDestinationEncryptionConfigurationPtrOutput

func (JobLoadDestinationEncryptionConfigurationOutput) ToJobLoadDestinationEncryptionConfigurationPtrOutputWithContext

func (o JobLoadDestinationEncryptionConfigurationOutput) ToJobLoadDestinationEncryptionConfigurationPtrOutputWithContext(ctx context.Context) JobLoadDestinationEncryptionConfigurationPtrOutput

type JobLoadDestinationEncryptionConfigurationPtrInput

type JobLoadDestinationEncryptionConfigurationPtrInput interface {
	pulumi.Input

	ToJobLoadDestinationEncryptionConfigurationPtrOutput() JobLoadDestinationEncryptionConfigurationPtrOutput
	ToJobLoadDestinationEncryptionConfigurationPtrOutputWithContext(context.Context) JobLoadDestinationEncryptionConfigurationPtrOutput
}

JobLoadDestinationEncryptionConfigurationPtrInput is an input type that accepts JobLoadDestinationEncryptionConfigurationArgs, JobLoadDestinationEncryptionConfigurationPtr and JobLoadDestinationEncryptionConfigurationPtrOutput values. You can construct a concrete instance of `JobLoadDestinationEncryptionConfigurationPtrInput` via:

        JobLoadDestinationEncryptionConfigurationArgs{...}

or:

        nil

type JobLoadDestinationEncryptionConfigurationPtrOutput

type JobLoadDestinationEncryptionConfigurationPtrOutput struct{ *pulumi.OutputState }

func (JobLoadDestinationEncryptionConfigurationPtrOutput) Elem

func (JobLoadDestinationEncryptionConfigurationPtrOutput) ElementType

func (JobLoadDestinationEncryptionConfigurationPtrOutput) KmsKeyName

Describes the Cloud KMS encryption key that will be used to protect destination BigQuery table. The BigQuery Service Account associated with your project requires access to this encryption key.

func (JobLoadDestinationEncryptionConfigurationPtrOutput) ToJobLoadDestinationEncryptionConfigurationPtrOutput

func (o JobLoadDestinationEncryptionConfigurationPtrOutput) ToJobLoadDestinationEncryptionConfigurationPtrOutput() JobLoadDestinationEncryptionConfigurationPtrOutput

func (JobLoadDestinationEncryptionConfigurationPtrOutput) ToJobLoadDestinationEncryptionConfigurationPtrOutputWithContext

func (o JobLoadDestinationEncryptionConfigurationPtrOutput) ToJobLoadDestinationEncryptionConfigurationPtrOutputWithContext(ctx context.Context) JobLoadDestinationEncryptionConfigurationPtrOutput

type JobLoadDestinationTable

type JobLoadDestinationTable struct {
	// The ID of the dataset containing this model.
	DatasetId *string `pulumi:"datasetId"`
	// The ID of the project containing this model.
	ProjectId *string `pulumi:"projectId"`
	// The table. Can be specified `{{table_id}}` if `projectId` and `datasetId` are also set,
	// or of the form `projects/{{project}}/datasets/{{dataset_id}}/tables/{{table_id}}` if not.
	TableId string `pulumi:"tableId"`
}

type JobLoadDestinationTableArgs

type JobLoadDestinationTableArgs struct {
	// The ID of the dataset containing this model.
	DatasetId pulumi.StringPtrInput `pulumi:"datasetId"`
	// The ID of the project containing this model.
	ProjectId pulumi.StringPtrInput `pulumi:"projectId"`
	// The table. Can be specified `{{table_id}}` if `projectId` and `datasetId` are also set,
	// or of the form `projects/{{project}}/datasets/{{dataset_id}}/tables/{{table_id}}` if not.
	TableId pulumi.StringInput `pulumi:"tableId"`
}

func (JobLoadDestinationTableArgs) ElementType

func (JobLoadDestinationTableArgs) ToJobLoadDestinationTableOutput

func (i JobLoadDestinationTableArgs) ToJobLoadDestinationTableOutput() JobLoadDestinationTableOutput

func (JobLoadDestinationTableArgs) ToJobLoadDestinationTableOutputWithContext

func (i JobLoadDestinationTableArgs) ToJobLoadDestinationTableOutputWithContext(ctx context.Context) JobLoadDestinationTableOutput

func (JobLoadDestinationTableArgs) ToJobLoadDestinationTablePtrOutput

func (i JobLoadDestinationTableArgs) ToJobLoadDestinationTablePtrOutput() JobLoadDestinationTablePtrOutput

func (JobLoadDestinationTableArgs) ToJobLoadDestinationTablePtrOutputWithContext

func (i JobLoadDestinationTableArgs) ToJobLoadDestinationTablePtrOutputWithContext(ctx context.Context) JobLoadDestinationTablePtrOutput

type JobLoadDestinationTableInput

type JobLoadDestinationTableInput interface {
	pulumi.Input

	ToJobLoadDestinationTableOutput() JobLoadDestinationTableOutput
	ToJobLoadDestinationTableOutputWithContext(context.Context) JobLoadDestinationTableOutput
}

JobLoadDestinationTableInput is an input type that accepts JobLoadDestinationTableArgs and JobLoadDestinationTableOutput values. You can construct a concrete instance of `JobLoadDestinationTableInput` via:

JobLoadDestinationTableArgs{...}

type JobLoadDestinationTableOutput

type JobLoadDestinationTableOutput struct{ *pulumi.OutputState }

func (JobLoadDestinationTableOutput) DatasetId

The ID of the dataset containing this model.

func (JobLoadDestinationTableOutput) ElementType

func (JobLoadDestinationTableOutput) ProjectId

The ID of the project containing this model.

func (JobLoadDestinationTableOutput) TableId

The table. Can be specified `{{table_id}}` if `projectId` and `datasetId` are also set, or of the form `projects/{{project}}/datasets/{{dataset_id}}/tables/{{table_id}}` if not.

func (JobLoadDestinationTableOutput) ToJobLoadDestinationTableOutput

func (o JobLoadDestinationTableOutput) ToJobLoadDestinationTableOutput() JobLoadDestinationTableOutput

func (JobLoadDestinationTableOutput) ToJobLoadDestinationTableOutputWithContext

func (o JobLoadDestinationTableOutput) ToJobLoadDestinationTableOutputWithContext(ctx context.Context) JobLoadDestinationTableOutput

func (JobLoadDestinationTableOutput) ToJobLoadDestinationTablePtrOutput

func (o JobLoadDestinationTableOutput) ToJobLoadDestinationTablePtrOutput() JobLoadDestinationTablePtrOutput

func (JobLoadDestinationTableOutput) ToJobLoadDestinationTablePtrOutputWithContext

func (o JobLoadDestinationTableOutput) ToJobLoadDestinationTablePtrOutputWithContext(ctx context.Context) JobLoadDestinationTablePtrOutput

type JobLoadDestinationTablePtrInput

type JobLoadDestinationTablePtrInput interface {
	pulumi.Input

	ToJobLoadDestinationTablePtrOutput() JobLoadDestinationTablePtrOutput
	ToJobLoadDestinationTablePtrOutputWithContext(context.Context) JobLoadDestinationTablePtrOutput
}

JobLoadDestinationTablePtrInput is an input type that accepts JobLoadDestinationTableArgs, JobLoadDestinationTablePtr and JobLoadDestinationTablePtrOutput values. You can construct a concrete instance of `JobLoadDestinationTablePtrInput` via:

        JobLoadDestinationTableArgs{...}

or:

        nil

type JobLoadDestinationTablePtrOutput

type JobLoadDestinationTablePtrOutput struct{ *pulumi.OutputState }

func (JobLoadDestinationTablePtrOutput) DatasetId

The ID of the dataset containing this model.

func (JobLoadDestinationTablePtrOutput) Elem

func (JobLoadDestinationTablePtrOutput) ElementType

func (JobLoadDestinationTablePtrOutput) ProjectId

The ID of the project containing this model.

func (JobLoadDestinationTablePtrOutput) TableId

The table. Can be specified `{{table_id}}` if `projectId` and `datasetId` are also set, or of the form `projects/{{project}}/datasets/{{dataset_id}}/tables/{{table_id}}` if not.

func (JobLoadDestinationTablePtrOutput) ToJobLoadDestinationTablePtrOutput

func (o JobLoadDestinationTablePtrOutput) ToJobLoadDestinationTablePtrOutput() JobLoadDestinationTablePtrOutput

func (JobLoadDestinationTablePtrOutput) ToJobLoadDestinationTablePtrOutputWithContext

func (o JobLoadDestinationTablePtrOutput) ToJobLoadDestinationTablePtrOutputWithContext(ctx context.Context) JobLoadDestinationTablePtrOutput

type JobLoadInput

type JobLoadInput interface {
	pulumi.Input

	ToJobLoadOutput() JobLoadOutput
	ToJobLoadOutputWithContext(context.Context) JobLoadOutput
}

JobLoadInput is an input type that accepts JobLoadArgs and JobLoadOutput values. You can construct a concrete instance of `JobLoadInput` via:

JobLoadArgs{...}

type JobLoadOutput

type JobLoadOutput struct{ *pulumi.OutputState }

func (JobLoadOutput) AllowJaggedRows

func (o JobLoadOutput) AllowJaggedRows() pulumi.BoolPtrOutput

Accept rows that are missing trailing optional columns. The missing values are treated as nulls. If false, records with missing trailing columns are treated as bad records, and if there are too many bad records, an invalid error is returned in the job result. The default value is false. Only applicable to CSV, ignored for other formats.

func (JobLoadOutput) AllowQuotedNewlines

func (o JobLoadOutput) AllowQuotedNewlines() pulumi.BoolPtrOutput

Indicates if BigQuery should allow quoted data sections that contain newline characters in a CSV file. The default value is false.

func (JobLoadOutput) Autodetect

func (o JobLoadOutput) Autodetect() pulumi.BoolPtrOutput

Indicates if we should automatically infer the options and schema for CSV and JSON sources.

func (JobLoadOutput) CreateDisposition

func (o JobLoadOutput) CreateDisposition() pulumi.StringPtrOutput

Specifies whether the job is allowed to create new tables. The following values are supported: CREATE_IF_NEEDED: If the table does not exist, BigQuery creates the table. CREATE_NEVER: The table must already exist. If it does not, a 'notFound' error is returned in the job result. Creation, truncation and append actions occur as one atomic update upon job completion Default value is `CREATE_IF_NEEDED`. Possible values are `CREATE_IF_NEEDED` and `CREATE_NEVER`.

func (JobLoadOutput) DestinationEncryptionConfiguration

func (o JobLoadOutput) DestinationEncryptionConfiguration() JobLoadDestinationEncryptionConfigurationPtrOutput

Custom encryption configuration (e.g., Cloud KMS keys) Structure is documented below.

func (JobLoadOutput) DestinationTable

func (o JobLoadOutput) DestinationTable() JobLoadDestinationTableOutput

The destination table. Structure is documented below.

func (JobLoadOutput) ElementType

func (JobLoadOutput) ElementType() reflect.Type

func (JobLoadOutput) Encoding

func (o JobLoadOutput) Encoding() pulumi.StringPtrOutput

The character encoding of the data. The supported values are UTF-8 or ISO-8859-1. The default value is UTF-8. BigQuery decodes the data after the raw, binary data has been split using the values of the quote and fieldDelimiter properties.

func (JobLoadOutput) FieldDelimiter

func (o JobLoadOutput) FieldDelimiter() pulumi.StringPtrOutput

When extracting data in CSV format, this defines the delimiter to use between fields in the exported data. Default is ','

func (JobLoadOutput) IgnoreUnknownValues

func (o JobLoadOutput) IgnoreUnknownValues() pulumi.BoolPtrOutput

Indicates if BigQuery should allow extra values that are not represented in the table schema. If true, the extra values are ignored. If false, records with extra columns are treated as bad records, and if there are too many bad records, an invalid error is returned in the job result. The default value is false. The sourceFormat property determines what BigQuery treats as an extra value: CSV: Trailing columns JSON: Named values that don't match any column names

func (JobLoadOutput) MaxBadRecords

func (o JobLoadOutput) MaxBadRecords() pulumi.IntPtrOutput

The maximum number of bad records that BigQuery can ignore when running the job. If the number of bad records exceeds this value, an invalid error is returned in the job result. The default value is 0, which requires that all records are valid.

func (JobLoadOutput) NullMarker

func (o JobLoadOutput) NullMarker() pulumi.StringPtrOutput

Specifies a string that represents a null value in a CSV file. The default value is the empty string. If you set this property to a custom value, BigQuery throws an error if an empty string is present for all data types except for STRING and BYTE. For STRING and BYTE columns, BigQuery interprets the empty string as an empty value.

func (JobLoadOutput) ProjectionFields

func (o JobLoadOutput) ProjectionFields() pulumi.StringArrayOutput

If sourceFormat is set to "DATASTORE_BACKUP", indicates which entity properties to load into BigQuery from a Cloud Datastore backup. Property names are case sensitive and must be top-level properties. If no properties are specified, BigQuery loads all properties. If any named property isn't found in the Cloud Datastore backup, an invalid error is returned in the job result.

func (JobLoadOutput) Quote

The value that is used to quote data sections in a CSV file. BigQuery converts the string to ISO-8859-1 encoding, and then uses the first byte of the encoded string to split the data in its raw, binary state. The default value is a double-quote ('"'). If your data does not contain quoted sections, set the property value to an empty string. If your data contains quoted newline characters, you must also set the allowQuotedNewlines property to true.

func (JobLoadOutput) SchemaUpdateOptions

func (o JobLoadOutput) SchemaUpdateOptions() pulumi.StringArrayOutput

Allows the schema of the destination table to be updated as a side effect of the load job if a schema is autodetected or supplied in the job configuration. Schema update options are supported in two cases: when writeDisposition is WRITE_APPEND; when writeDisposition is WRITE_TRUNCATE and the destination table is a partition of a table, specified by partition decorators. For normal tables, WRITE_TRUNCATE will always overwrite the schema. One or more of the following values are specified: ALLOW_FIELD_ADDITION: allow adding a nullable field to the schema. ALLOW_FIELD_RELAXATION: allow relaxing a required field in the original schema to nullable.

func (JobLoadOutput) SkipLeadingRows

func (o JobLoadOutput) SkipLeadingRows() pulumi.IntPtrOutput

The number of rows at the top of a CSV file that BigQuery will skip when loading the data. The default value is 0. This property is useful if you have header rows in the file that should be skipped. When autodetect is on, the behavior is the following: skipLeadingRows unspecified - Autodetect tries to detect headers in the first row. If they are not detected, the row is read as data. Otherwise data is read starting from the second row. skipLeadingRows is 0 - Instructs autodetect that there are no headers and data should be read starting from the first row. skipLeadingRows = N > 0 - Autodetect skips N-1 rows and tries to detect headers in row N. If headers are not detected, row N is just skipped. Otherwise row N is used to extract column names for the detected schema.

func (JobLoadOutput) SourceFormat

func (o JobLoadOutput) SourceFormat() pulumi.StringPtrOutput

The format of the data files. For CSV files, specify "CSV". For datastore backups, specify "DATASTORE_BACKUP". For newline-delimited JSON, specify "NEWLINE_DELIMITED_JSON". For Avro, specify "AVRO". For parquet, specify "PARQUET". For orc, specify "ORC". The default value is CSV.

func (JobLoadOutput) SourceUris

func (o JobLoadOutput) SourceUris() pulumi.StringArrayOutput

The fully-qualified URIs that point to your data in Google Cloud. For Google Cloud Storage URIs: Each URI can contain one '*' wildcard character and it must come after the 'bucket' name. Size limits related to load jobs apply to external data sources. For Google Cloud Bigtable URIs: Exactly one URI can be specified and it has be a fully specified and valid HTTPS URL for a Google Cloud Bigtable table. For Google Cloud Datastore backups: Exactly one URI can be specified. Also, the '*' wildcard character is not allowed.

func (JobLoadOutput) TimePartitioning

func (o JobLoadOutput) TimePartitioning() JobLoadTimePartitioningPtrOutput

Time-based partitioning specification for the destination table. Structure is documented below.

func (JobLoadOutput) ToJobLoadOutput

func (o JobLoadOutput) ToJobLoadOutput() JobLoadOutput

func (JobLoadOutput) ToJobLoadOutputWithContext

func (o JobLoadOutput) ToJobLoadOutputWithContext(ctx context.Context) JobLoadOutput

func (JobLoadOutput) ToJobLoadPtrOutput

func (o JobLoadOutput) ToJobLoadPtrOutput() JobLoadPtrOutput

func (JobLoadOutput) ToJobLoadPtrOutputWithContext

func (o JobLoadOutput) ToJobLoadPtrOutputWithContext(ctx context.Context) JobLoadPtrOutput

func (JobLoadOutput) WriteDisposition

func (o JobLoadOutput) WriteDisposition() pulumi.StringPtrOutput

Specifies the action that occurs if the destination table already exists. The following values are supported: WRITE_TRUNCATE: If the table already exists, BigQuery overwrites the table data and uses the schema from the query result. WRITE_APPEND: If the table already exists, BigQuery appends the data to the table. WRITE_EMPTY: If the table already exists and contains data, a 'duplicate' error is returned in the job result. Each action is atomic and only occurs if BigQuery is able to complete the job successfully. Creation, truncation and append actions occur as one atomic update upon job completion. Default value is `WRITE_EMPTY`. Possible values are `WRITE_TRUNCATE`, `WRITE_APPEND`, and `WRITE_EMPTY`.

type JobLoadPtrInput

type JobLoadPtrInput interface {
	pulumi.Input

	ToJobLoadPtrOutput() JobLoadPtrOutput
	ToJobLoadPtrOutputWithContext(context.Context) JobLoadPtrOutput
}

JobLoadPtrInput is an input type that accepts JobLoadArgs, JobLoadPtr and JobLoadPtrOutput values. You can construct a concrete instance of `JobLoadPtrInput` via:

        JobLoadArgs{...}

or:

        nil

func JobLoadPtr

func JobLoadPtr(v *JobLoadArgs) JobLoadPtrInput

type JobLoadPtrOutput

type JobLoadPtrOutput struct{ *pulumi.OutputState }

func (JobLoadPtrOutput) AllowJaggedRows

func (o JobLoadPtrOutput) AllowJaggedRows() pulumi.BoolPtrOutput

Accept rows that are missing trailing optional columns. The missing values are treated as nulls. If false, records with missing trailing columns are treated as bad records, and if there are too many bad records, an invalid error is returned in the job result. The default value is false. Only applicable to CSV, ignored for other formats.

func (JobLoadPtrOutput) AllowQuotedNewlines

func (o JobLoadPtrOutput) AllowQuotedNewlines() pulumi.BoolPtrOutput

Indicates if BigQuery should allow quoted data sections that contain newline characters in a CSV file. The default value is false.

func (JobLoadPtrOutput) Autodetect

func (o JobLoadPtrOutput) Autodetect() pulumi.BoolPtrOutput

Indicates if we should automatically infer the options and schema for CSV and JSON sources.

func (JobLoadPtrOutput) CreateDisposition

func (o JobLoadPtrOutput) CreateDisposition() pulumi.StringPtrOutput

Specifies whether the job is allowed to create new tables. The following values are supported: CREATE_IF_NEEDED: If the table does not exist, BigQuery creates the table. CREATE_NEVER: The table must already exist. If it does not, a 'notFound' error is returned in the job result. Creation, truncation and append actions occur as one atomic update upon job completion Default value is `CREATE_IF_NEEDED`. Possible values are `CREATE_IF_NEEDED` and `CREATE_NEVER`.

func (JobLoadPtrOutput) DestinationEncryptionConfiguration

func (o JobLoadPtrOutput) DestinationEncryptionConfiguration() JobLoadDestinationEncryptionConfigurationPtrOutput

Custom encryption configuration (e.g., Cloud KMS keys) Structure is documented below.

func (JobLoadPtrOutput) DestinationTable

The destination table. Structure is documented below.

func (JobLoadPtrOutput) Elem

func (JobLoadPtrOutput) ElementType

func (JobLoadPtrOutput) ElementType() reflect.Type

func (JobLoadPtrOutput) Encoding

The character encoding of the data. The supported values are UTF-8 or ISO-8859-1. The default value is UTF-8. BigQuery decodes the data after the raw, binary data has been split using the values of the quote and fieldDelimiter properties.

func (JobLoadPtrOutput) FieldDelimiter

func (o JobLoadPtrOutput) FieldDelimiter() pulumi.StringPtrOutput

When extracting data in CSV format, this defines the delimiter to use between fields in the exported data. Default is ','

func (JobLoadPtrOutput) IgnoreUnknownValues

func (o JobLoadPtrOutput) IgnoreUnknownValues() pulumi.BoolPtrOutput

Indicates if BigQuery should allow extra values that are not represented in the table schema. If true, the extra values are ignored. If false, records with extra columns are treated as bad records, and if there are too many bad records, an invalid error is returned in the job result. The default value is false. The sourceFormat property determines what BigQuery treats as an extra value: CSV: Trailing columns JSON: Named values that don't match any column names

func (JobLoadPtrOutput) MaxBadRecords

func (o JobLoadPtrOutput) MaxBadRecords() pulumi.IntPtrOutput

The maximum number of bad records that BigQuery can ignore when running the job. If the number of bad records exceeds this value, an invalid error is returned in the job result. The default value is 0, which requires that all records are valid.

func (JobLoadPtrOutput) NullMarker

func (o JobLoadPtrOutput) NullMarker() pulumi.StringPtrOutput

Specifies a string that represents a null value in a CSV file. The default value is the empty string. If you set this property to a custom value, BigQuery throws an error if an empty string is present for all data types except for STRING and BYTE. For STRING and BYTE columns, BigQuery interprets the empty string as an empty value.

func (JobLoadPtrOutput) ProjectionFields

func (o JobLoadPtrOutput) ProjectionFields() pulumi.StringArrayOutput

If sourceFormat is set to "DATASTORE_BACKUP", indicates which entity properties to load into BigQuery from a Cloud Datastore backup. Property names are case sensitive and must be top-level properties. If no properties are specified, BigQuery loads all properties. If any named property isn't found in the Cloud Datastore backup, an invalid error is returned in the job result.

func (JobLoadPtrOutput) Quote

The value that is used to quote data sections in a CSV file. BigQuery converts the string to ISO-8859-1 encoding, and then uses the first byte of the encoded string to split the data in its raw, binary state. The default value is a double-quote ('"'). If your data does not contain quoted sections, set the property value to an empty string. If your data contains quoted newline characters, you must also set the allowQuotedNewlines property to true.

func (JobLoadPtrOutput) SchemaUpdateOptions

func (o JobLoadPtrOutput) SchemaUpdateOptions() pulumi.StringArrayOutput

Allows the schema of the destination table to be updated as a side effect of the load job if a schema is autodetected or supplied in the job configuration. Schema update options are supported in two cases: when writeDisposition is WRITE_APPEND; when writeDisposition is WRITE_TRUNCATE and the destination table is a partition of a table, specified by partition decorators. For normal tables, WRITE_TRUNCATE will always overwrite the schema. One or more of the following values are specified: ALLOW_FIELD_ADDITION: allow adding a nullable field to the schema. ALLOW_FIELD_RELAXATION: allow relaxing a required field in the original schema to nullable.

func (JobLoadPtrOutput) SkipLeadingRows

func (o JobLoadPtrOutput) SkipLeadingRows() pulumi.IntPtrOutput

The number of rows at the top of a CSV file that BigQuery will skip when loading the data. The default value is 0. This property is useful if you have header rows in the file that should be skipped. When autodetect is on, the behavior is the following: skipLeadingRows unspecified - Autodetect tries to detect headers in the first row. If they are not detected, the row is read as data. Otherwise data is read starting from the second row. skipLeadingRows is 0 - Instructs autodetect that there are no headers and data should be read starting from the first row. skipLeadingRows = N > 0 - Autodetect skips N-1 rows and tries to detect headers in row N. If headers are not detected, row N is just skipped. Otherwise row N is used to extract column names for the detected schema.

func (JobLoadPtrOutput) SourceFormat

func (o JobLoadPtrOutput) SourceFormat() pulumi.StringPtrOutput

The format of the data files. For CSV files, specify "CSV". For datastore backups, specify "DATASTORE_BACKUP". For newline-delimited JSON, specify "NEWLINE_DELIMITED_JSON". For Avro, specify "AVRO". For parquet, specify "PARQUET". For orc, specify "ORC". The default value is CSV.

func (JobLoadPtrOutput) SourceUris

func (o JobLoadPtrOutput) SourceUris() pulumi.StringArrayOutput

The fully-qualified URIs that point to your data in Google Cloud. For Google Cloud Storage URIs: Each URI can contain one '*' wildcard character and it must come after the 'bucket' name. Size limits related to load jobs apply to external data sources. For Google Cloud Bigtable URIs: Exactly one URI can be specified and it has be a fully specified and valid HTTPS URL for a Google Cloud Bigtable table. For Google Cloud Datastore backups: Exactly one URI can be specified. Also, the '*' wildcard character is not allowed.

func (JobLoadPtrOutput) TimePartitioning

Time-based partitioning specification for the destination table. Structure is documented below.

func (JobLoadPtrOutput) ToJobLoadPtrOutput

func (o JobLoadPtrOutput) ToJobLoadPtrOutput() JobLoadPtrOutput

func (JobLoadPtrOutput) ToJobLoadPtrOutputWithContext

func (o JobLoadPtrOutput) ToJobLoadPtrOutputWithContext(ctx context.Context) JobLoadPtrOutput

func (JobLoadPtrOutput) WriteDisposition

func (o JobLoadPtrOutput) WriteDisposition() pulumi.StringPtrOutput

Specifies the action that occurs if the destination table already exists. The following values are supported: WRITE_TRUNCATE: If the table already exists, BigQuery overwrites the table data and uses the schema from the query result. WRITE_APPEND: If the table already exists, BigQuery appends the data to the table. WRITE_EMPTY: If the table already exists and contains data, a 'duplicate' error is returned in the job result. Each action is atomic and only occurs if BigQuery is able to complete the job successfully. Creation, truncation and append actions occur as one atomic update upon job completion. Default value is `WRITE_EMPTY`. Possible values are `WRITE_TRUNCATE`, `WRITE_APPEND`, and `WRITE_EMPTY`.

type JobLoadTimePartitioning

type JobLoadTimePartitioning struct {
	// Number of milliseconds for which to keep the storage for a partition. A wrapper is used here because 0 is an invalid value.
	ExpirationMs *string `pulumi:"expirationMs"`
	// If not set, the table is partitioned by pseudo column '_PARTITIONTIME'; if set, the table is partitioned by this field.
	// The field must be a top-level TIMESTAMP or DATE field. Its mode must be NULLABLE or REQUIRED.
	// A wrapper is used here because an empty string is an invalid value.
	Field *string `pulumi:"field"`
	// The only type supported is DAY, which will generate one partition per day. Providing an empty string used to cause an error,
	// but in OnePlatform the field will be treated as unset.
	Type string `pulumi:"type"`
}

type JobLoadTimePartitioningArgs

type JobLoadTimePartitioningArgs struct {
	// Number of milliseconds for which to keep the storage for a partition. A wrapper is used here because 0 is an invalid value.
	ExpirationMs pulumi.StringPtrInput `pulumi:"expirationMs"`
	// If not set, the table is partitioned by pseudo column '_PARTITIONTIME'; if set, the table is partitioned by this field.
	// The field must be a top-level TIMESTAMP or DATE field. Its mode must be NULLABLE or REQUIRED.
	// A wrapper is used here because an empty string is an invalid value.
	Field pulumi.StringPtrInput `pulumi:"field"`
	// The only type supported is DAY, which will generate one partition per day. Providing an empty string used to cause an error,
	// but in OnePlatform the field will be treated as unset.
	Type pulumi.StringInput `pulumi:"type"`
}

func (JobLoadTimePartitioningArgs) ElementType

func (JobLoadTimePartitioningArgs) ToJobLoadTimePartitioningOutput

func (i JobLoadTimePartitioningArgs) ToJobLoadTimePartitioningOutput() JobLoadTimePartitioningOutput

func (JobLoadTimePartitioningArgs) ToJobLoadTimePartitioningOutputWithContext

func (i JobLoadTimePartitioningArgs) ToJobLoadTimePartitioningOutputWithContext(ctx context.Context) JobLoadTimePartitioningOutput

func (JobLoadTimePartitioningArgs) ToJobLoadTimePartitioningPtrOutput

func (i JobLoadTimePartitioningArgs) ToJobLoadTimePartitioningPtrOutput() JobLoadTimePartitioningPtrOutput

func (JobLoadTimePartitioningArgs) ToJobLoadTimePartitioningPtrOutputWithContext

func (i JobLoadTimePartitioningArgs) ToJobLoadTimePartitioningPtrOutputWithContext(ctx context.Context) JobLoadTimePartitioningPtrOutput

type JobLoadTimePartitioningInput

type JobLoadTimePartitioningInput interface {
	pulumi.Input

	ToJobLoadTimePartitioningOutput() JobLoadTimePartitioningOutput
	ToJobLoadTimePartitioningOutputWithContext(context.Context) JobLoadTimePartitioningOutput
}

JobLoadTimePartitioningInput is an input type that accepts JobLoadTimePartitioningArgs and JobLoadTimePartitioningOutput values. You can construct a concrete instance of `JobLoadTimePartitioningInput` via:

JobLoadTimePartitioningArgs{...}

type JobLoadTimePartitioningOutput

type JobLoadTimePartitioningOutput struct{ *pulumi.OutputState }

func (JobLoadTimePartitioningOutput) ElementType

func (JobLoadTimePartitioningOutput) ExpirationMs

Number of milliseconds for which to keep the storage for a partition. A wrapper is used here because 0 is an invalid value.

func (JobLoadTimePartitioningOutput) Field

If not set, the table is partitioned by pseudo column '_PARTITIONTIME'; if set, the table is partitioned by this field. The field must be a top-level TIMESTAMP or DATE field. Its mode must be NULLABLE or REQUIRED. A wrapper is used here because an empty string is an invalid value.

func (JobLoadTimePartitioningOutput) ToJobLoadTimePartitioningOutput

func (o JobLoadTimePartitioningOutput) ToJobLoadTimePartitioningOutput() JobLoadTimePartitioningOutput

func (JobLoadTimePartitioningOutput) ToJobLoadTimePartitioningOutputWithContext

func (o JobLoadTimePartitioningOutput) ToJobLoadTimePartitioningOutputWithContext(ctx context.Context) JobLoadTimePartitioningOutput

func (JobLoadTimePartitioningOutput) ToJobLoadTimePartitioningPtrOutput

func (o JobLoadTimePartitioningOutput) ToJobLoadTimePartitioningPtrOutput() JobLoadTimePartitioningPtrOutput

func (JobLoadTimePartitioningOutput) ToJobLoadTimePartitioningPtrOutputWithContext

func (o JobLoadTimePartitioningOutput) ToJobLoadTimePartitioningPtrOutputWithContext(ctx context.Context) JobLoadTimePartitioningPtrOutput

func (JobLoadTimePartitioningOutput) Type

The only type supported is DAY, which will generate one partition per day. Providing an empty string used to cause an error, but in OnePlatform the field will be treated as unset.

type JobLoadTimePartitioningPtrInput

type JobLoadTimePartitioningPtrInput interface {
	pulumi.Input

	ToJobLoadTimePartitioningPtrOutput() JobLoadTimePartitioningPtrOutput
	ToJobLoadTimePartitioningPtrOutputWithContext(context.Context) JobLoadTimePartitioningPtrOutput
}

JobLoadTimePartitioningPtrInput is an input type that accepts JobLoadTimePartitioningArgs, JobLoadTimePartitioningPtr and JobLoadTimePartitioningPtrOutput values. You can construct a concrete instance of `JobLoadTimePartitioningPtrInput` via:

        JobLoadTimePartitioningArgs{...}

or:

        nil

type JobLoadTimePartitioningPtrOutput

type JobLoadTimePartitioningPtrOutput struct{ *pulumi.OutputState }

func (JobLoadTimePartitioningPtrOutput) Elem

func (JobLoadTimePartitioningPtrOutput) ElementType

func (JobLoadTimePartitioningPtrOutput) ExpirationMs

Number of milliseconds for which to keep the storage for a partition. A wrapper is used here because 0 is an invalid value.

func (JobLoadTimePartitioningPtrOutput) Field

If not set, the table is partitioned by pseudo column '_PARTITIONTIME'; if set, the table is partitioned by this field. The field must be a top-level TIMESTAMP or DATE field. Its mode must be NULLABLE or REQUIRED. A wrapper is used here because an empty string is an invalid value.

func (JobLoadTimePartitioningPtrOutput) ToJobLoadTimePartitioningPtrOutput

func (o JobLoadTimePartitioningPtrOutput) ToJobLoadTimePartitioningPtrOutput() JobLoadTimePartitioningPtrOutput

func (JobLoadTimePartitioningPtrOutput) ToJobLoadTimePartitioningPtrOutputWithContext

func (o JobLoadTimePartitioningPtrOutput) ToJobLoadTimePartitioningPtrOutputWithContext(ctx context.Context) JobLoadTimePartitioningPtrOutput

func (JobLoadTimePartitioningPtrOutput) Type

The only type supported is DAY, which will generate one partition per day. Providing an empty string used to cause an error, but in OnePlatform the field will be treated as unset.

type JobQuery

type JobQuery struct {
	// If true and query uses legacy SQL dialect, allows the query to produce arbitrarily large result tables at a slight cost in performance.
	// Requires destinationTable to be set. For standard SQL queries, this flag is ignored and large results are always allowed.
	// However, you must still set destinationTable when result size exceeds the allowed maximum response size.
	AllowLargeResults *bool `pulumi:"allowLargeResults"`
	// Specifies whether the job is allowed to create new tables. The following values are supported:
	// CREATE_IF_NEEDED: If the table does not exist, BigQuery creates the table.
	// CREATE_NEVER: The table must already exist. If it does not, a 'notFound' error is returned in the job result.
	// Creation, truncation and append actions occur as one atomic update upon job completion
	// Default value is `CREATE_IF_NEEDED`.
	// Possible values are `CREATE_IF_NEEDED` and `CREATE_NEVER`.
	CreateDisposition *string `pulumi:"createDisposition"`
	// Specifies the default dataset to use for unqualified table names in the query. Note that this does not alter behavior of unqualified dataset names.
	// Structure is documented below.
	DefaultDataset *JobQueryDefaultDataset `pulumi:"defaultDataset"`
	// Custom encryption configuration (e.g., Cloud KMS keys)
	// Structure is documented below.
	DestinationEncryptionConfiguration *JobQueryDestinationEncryptionConfiguration `pulumi:"destinationEncryptionConfiguration"`
	// The destination table.
	// Structure is documented below.
	DestinationTable *JobQueryDestinationTable `pulumi:"destinationTable"`
	// If true and query uses legacy SQL dialect, flattens all nested and repeated fields in the query results.
	// allowLargeResults must be true if this is set to false. For standard SQL queries, this flag is ignored and results are never flattened.
	FlattenResults *bool `pulumi:"flattenResults"`
	// Limits the billing tier for this job. Queries that have resource usage beyond this tier will fail (without incurring a charge).
	// If unspecified, this will be set to your project default.
	MaximumBillingTier *int `pulumi:"maximumBillingTier"`
	// Limits the bytes billed for this job. Queries that will have bytes billed beyond this limit will fail (without incurring a charge).
	// If unspecified, this will be set to your project default.
	MaximumBytesBilled *string `pulumi:"maximumBytesBilled"`
	// Standard SQL only. Set to POSITIONAL to use positional (?) query parameters or to NAMED to use named (@myparam) query parameters in this query.
	ParameterMode *string `pulumi:"parameterMode"`
	// Specifies a priority for the query.
	// Default value is `INTERACTIVE`.
	// Possible values are `INTERACTIVE` and `BATCH`.
	Priority *string `pulumi:"priority"`
	// Configures a query job.
	// Structure is documented below.
	Query string `pulumi:"query"`
	// Allows the schema of the destination table to be updated as a side effect of the load job if a schema is autodetected or
	// supplied in the job configuration. Schema update options are supported in two cases: when writeDisposition is WRITE_APPEND;
	// when writeDisposition is WRITE_TRUNCATE and the destination table is a partition of a table, specified by partition decorators.
	// For normal tables, WRITE_TRUNCATE will always overwrite the schema. One or more of the following values are specified:
	// ALLOW_FIELD_ADDITION: allow adding a nullable field to the schema.
	// ALLOW_FIELD_RELAXATION: allow relaxing a required field in the original schema to nullable.
	SchemaUpdateOptions []string `pulumi:"schemaUpdateOptions"`
	// Options controlling the execution of scripts.
	// Structure is documented below.
	ScriptOptions *JobQueryScriptOptions `pulumi:"scriptOptions"`
	// Specifies whether to use BigQuery's legacy SQL dialect for this query. The default value is true.
	// If set to false, the query will use BigQuery's standard SQL.
	UseLegacySql *bool `pulumi:"useLegacySql"`
	// Whether to look for the result in the query cache. The query cache is a best-effort cache that will be flushed whenever
	// tables in the query are modified. Moreover, the query cache is only available when a query does not have a destination table specified.
	// The default value is true.
	UseQueryCache *bool `pulumi:"useQueryCache"`
	// Describes user-defined function resources used in the query.
	// Structure is documented below.
	UserDefinedFunctionResources []JobQueryUserDefinedFunctionResource `pulumi:"userDefinedFunctionResources"`
	// Specifies the action that occurs if the destination table already exists. The following values are supported:
	// WRITE_TRUNCATE: If the table already exists, BigQuery overwrites the table data and uses the schema from the query result.
	// WRITE_APPEND: If the table already exists, BigQuery appends the data to the table.
	// WRITE_EMPTY: If the table already exists and contains data, a 'duplicate' error is returned in the job result.
	// Each action is atomic and only occurs if BigQuery is able to complete the job successfully.
	// Creation, truncation and append actions occur as one atomic update upon job completion.
	// Default value is `WRITE_EMPTY`.
	// Possible values are `WRITE_TRUNCATE`, `WRITE_APPEND`, and `WRITE_EMPTY`.
	WriteDisposition *string `pulumi:"writeDisposition"`
}

type JobQueryArgs

type JobQueryArgs struct {
	// If true and query uses legacy SQL dialect, allows the query to produce arbitrarily large result tables at a slight cost in performance.
	// Requires destinationTable to be set. For standard SQL queries, this flag is ignored and large results are always allowed.
	// However, you must still set destinationTable when result size exceeds the allowed maximum response size.
	AllowLargeResults pulumi.BoolPtrInput `pulumi:"allowLargeResults"`
	// Specifies whether the job is allowed to create new tables. The following values are supported:
	// CREATE_IF_NEEDED: If the table does not exist, BigQuery creates the table.
	// CREATE_NEVER: The table must already exist. If it does not, a 'notFound' error is returned in the job result.
	// Creation, truncation and append actions occur as one atomic update upon job completion
	// Default value is `CREATE_IF_NEEDED`.
	// Possible values are `CREATE_IF_NEEDED` and `CREATE_NEVER`.
	CreateDisposition pulumi.StringPtrInput `pulumi:"createDisposition"`
	// Specifies the default dataset to use for unqualified table names in the query. Note that this does not alter behavior of unqualified dataset names.
	// Structure is documented below.
	DefaultDataset JobQueryDefaultDatasetPtrInput `pulumi:"defaultDataset"`
	// Custom encryption configuration (e.g., Cloud KMS keys)
	// Structure is documented below.
	DestinationEncryptionConfiguration JobQueryDestinationEncryptionConfigurationPtrInput `pulumi:"destinationEncryptionConfiguration"`
	// The destination table.
	// Structure is documented below.
	DestinationTable JobQueryDestinationTablePtrInput `pulumi:"destinationTable"`
	// If true and query uses legacy SQL dialect, flattens all nested and repeated fields in the query results.
	// allowLargeResults must be true if this is set to false. For standard SQL queries, this flag is ignored and results are never flattened.
	FlattenResults pulumi.BoolPtrInput `pulumi:"flattenResults"`
	// Limits the billing tier for this job. Queries that have resource usage beyond this tier will fail (without incurring a charge).
	// If unspecified, this will be set to your project default.
	MaximumBillingTier pulumi.IntPtrInput `pulumi:"maximumBillingTier"`
	// Limits the bytes billed for this job. Queries that will have bytes billed beyond this limit will fail (without incurring a charge).
	// If unspecified, this will be set to your project default.
	MaximumBytesBilled pulumi.StringPtrInput `pulumi:"maximumBytesBilled"`
	// Standard SQL only. Set to POSITIONAL to use positional (?) query parameters or to NAMED to use named (@myparam) query parameters in this query.
	ParameterMode pulumi.StringPtrInput `pulumi:"parameterMode"`
	// Specifies a priority for the query.
	// Default value is `INTERACTIVE`.
	// Possible values are `INTERACTIVE` and `BATCH`.
	Priority pulumi.StringPtrInput `pulumi:"priority"`
	// Configures a query job.
	// Structure is documented below.
	Query pulumi.StringInput `pulumi:"query"`
	// Allows the schema of the destination table to be updated as a side effect of the load job if a schema is autodetected or
	// supplied in the job configuration. Schema update options are supported in two cases: when writeDisposition is WRITE_APPEND;
	// when writeDisposition is WRITE_TRUNCATE and the destination table is a partition of a table, specified by partition decorators.
	// For normal tables, WRITE_TRUNCATE will always overwrite the schema. One or more of the following values are specified:
	// ALLOW_FIELD_ADDITION: allow adding a nullable field to the schema.
	// ALLOW_FIELD_RELAXATION: allow relaxing a required field in the original schema to nullable.
	SchemaUpdateOptions pulumi.StringArrayInput `pulumi:"schemaUpdateOptions"`
	// Options controlling the execution of scripts.
	// Structure is documented below.
	ScriptOptions JobQueryScriptOptionsPtrInput `pulumi:"scriptOptions"`
	// Specifies whether to use BigQuery's legacy SQL dialect for this query. The default value is true.
	// If set to false, the query will use BigQuery's standard SQL.
	UseLegacySql pulumi.BoolPtrInput `pulumi:"useLegacySql"`
	// Whether to look for the result in the query cache. The query cache is a best-effort cache that will be flushed whenever
	// tables in the query are modified. Moreover, the query cache is only available when a query does not have a destination table specified.
	// The default value is true.
	UseQueryCache pulumi.BoolPtrInput `pulumi:"useQueryCache"`
	// Describes user-defined function resources used in the query.
	// Structure is documented below.
	UserDefinedFunctionResources JobQueryUserDefinedFunctionResourceArrayInput `pulumi:"userDefinedFunctionResources"`
	// Specifies the action that occurs if the destination table already exists. The following values are supported:
	// WRITE_TRUNCATE: If the table already exists, BigQuery overwrites the table data and uses the schema from the query result.
	// WRITE_APPEND: If the table already exists, BigQuery appends the data to the table.
	// WRITE_EMPTY: If the table already exists and contains data, a 'duplicate' error is returned in the job result.
	// Each action is atomic and only occurs if BigQuery is able to complete the job successfully.
	// Creation, truncation and append actions occur as one atomic update upon job completion.
	// Default value is `WRITE_EMPTY`.
	// Possible values are `WRITE_TRUNCATE`, `WRITE_APPEND`, and `WRITE_EMPTY`.
	WriteDisposition pulumi.StringPtrInput `pulumi:"writeDisposition"`
}

func (JobQueryArgs) ElementType

func (JobQueryArgs) ElementType() reflect.Type

func (JobQueryArgs) ToJobQueryOutput

func (i JobQueryArgs) ToJobQueryOutput() JobQueryOutput

func (JobQueryArgs) ToJobQueryOutputWithContext

func (i JobQueryArgs) ToJobQueryOutputWithContext(ctx context.Context) JobQueryOutput

func (JobQueryArgs) ToJobQueryPtrOutput

func (i JobQueryArgs) ToJobQueryPtrOutput() JobQueryPtrOutput

func (JobQueryArgs) ToJobQueryPtrOutputWithContext

func (i JobQueryArgs) ToJobQueryPtrOutputWithContext(ctx context.Context) JobQueryPtrOutput

type JobQueryDefaultDataset

type JobQueryDefaultDataset struct {
	// The ID of the dataset containing this model.
	DatasetId string `pulumi:"datasetId"`
	// The ID of the project containing this model.
	ProjectId *string `pulumi:"projectId"`
}

type JobQueryDefaultDatasetArgs

type JobQueryDefaultDatasetArgs struct {
	// The ID of the dataset containing this model.
	DatasetId pulumi.StringInput `pulumi:"datasetId"`
	// The ID of the project containing this model.
	ProjectId pulumi.StringPtrInput `pulumi:"projectId"`
}

func (JobQueryDefaultDatasetArgs) ElementType

func (JobQueryDefaultDatasetArgs) ElementType() reflect.Type

func (JobQueryDefaultDatasetArgs) ToJobQueryDefaultDatasetOutput

func (i JobQueryDefaultDatasetArgs) ToJobQueryDefaultDatasetOutput() JobQueryDefaultDatasetOutput

func (JobQueryDefaultDatasetArgs) ToJobQueryDefaultDatasetOutputWithContext

func (i JobQueryDefaultDatasetArgs) ToJobQueryDefaultDatasetOutputWithContext(ctx context.Context) JobQueryDefaultDatasetOutput

func (JobQueryDefaultDatasetArgs) ToJobQueryDefaultDatasetPtrOutput

func (i JobQueryDefaultDatasetArgs) ToJobQueryDefaultDatasetPtrOutput() JobQueryDefaultDatasetPtrOutput

func (JobQueryDefaultDatasetArgs) ToJobQueryDefaultDatasetPtrOutputWithContext

func (i JobQueryDefaultDatasetArgs) ToJobQueryDefaultDatasetPtrOutputWithContext(ctx context.Context) JobQueryDefaultDatasetPtrOutput

type JobQueryDefaultDatasetInput

type JobQueryDefaultDatasetInput interface {
	pulumi.Input

	ToJobQueryDefaultDatasetOutput() JobQueryDefaultDatasetOutput
	ToJobQueryDefaultDatasetOutputWithContext(context.Context) JobQueryDefaultDatasetOutput
}

JobQueryDefaultDatasetInput is an input type that accepts JobQueryDefaultDatasetArgs and JobQueryDefaultDatasetOutput values. You can construct a concrete instance of `JobQueryDefaultDatasetInput` via:

JobQueryDefaultDatasetArgs{...}

type JobQueryDefaultDatasetOutput

type JobQueryDefaultDatasetOutput struct{ *pulumi.OutputState }

func (JobQueryDefaultDatasetOutput) DatasetId

The ID of the dataset containing this model.

func (JobQueryDefaultDatasetOutput) ElementType

func (JobQueryDefaultDatasetOutput) ProjectId

The ID of the project containing this model.

func (JobQueryDefaultDatasetOutput) ToJobQueryDefaultDatasetOutput

func (o JobQueryDefaultDatasetOutput) ToJobQueryDefaultDatasetOutput() JobQueryDefaultDatasetOutput

func (JobQueryDefaultDatasetOutput) ToJobQueryDefaultDatasetOutputWithContext

func (o JobQueryDefaultDatasetOutput) ToJobQueryDefaultDatasetOutputWithContext(ctx context.Context) JobQueryDefaultDatasetOutput

func (JobQueryDefaultDatasetOutput) ToJobQueryDefaultDatasetPtrOutput

func (o JobQueryDefaultDatasetOutput) ToJobQueryDefaultDatasetPtrOutput() JobQueryDefaultDatasetPtrOutput

func (JobQueryDefaultDatasetOutput) ToJobQueryDefaultDatasetPtrOutputWithContext

func (o JobQueryDefaultDatasetOutput) ToJobQueryDefaultDatasetPtrOutputWithContext(ctx context.Context) JobQueryDefaultDatasetPtrOutput

type JobQueryDefaultDatasetPtrInput

type JobQueryDefaultDatasetPtrInput interface {
	pulumi.Input

	ToJobQueryDefaultDatasetPtrOutput() JobQueryDefaultDatasetPtrOutput
	ToJobQueryDefaultDatasetPtrOutputWithContext(context.Context) JobQueryDefaultDatasetPtrOutput
}

JobQueryDefaultDatasetPtrInput is an input type that accepts JobQueryDefaultDatasetArgs, JobQueryDefaultDatasetPtr and JobQueryDefaultDatasetPtrOutput values. You can construct a concrete instance of `JobQueryDefaultDatasetPtrInput` via:

        JobQueryDefaultDatasetArgs{...}

or:

        nil

type JobQueryDefaultDatasetPtrOutput

type JobQueryDefaultDatasetPtrOutput struct{ *pulumi.OutputState }

func (JobQueryDefaultDatasetPtrOutput) DatasetId

The ID of the dataset containing this model.

func (JobQueryDefaultDatasetPtrOutput) Elem

func (JobQueryDefaultDatasetPtrOutput) ElementType

func (JobQueryDefaultDatasetPtrOutput) ProjectId

The ID of the project containing this model.

func (JobQueryDefaultDatasetPtrOutput) ToJobQueryDefaultDatasetPtrOutput

func (o JobQueryDefaultDatasetPtrOutput) ToJobQueryDefaultDatasetPtrOutput() JobQueryDefaultDatasetPtrOutput

func (JobQueryDefaultDatasetPtrOutput) ToJobQueryDefaultDatasetPtrOutputWithContext

func (o JobQueryDefaultDatasetPtrOutput) ToJobQueryDefaultDatasetPtrOutputWithContext(ctx context.Context) JobQueryDefaultDatasetPtrOutput

type JobQueryDestinationEncryptionConfiguration

type JobQueryDestinationEncryptionConfiguration struct {
	// Describes the Cloud KMS encryption key that will be used to protect destination BigQuery table.
	// The BigQuery Service Account associated with your project requires access to this encryption key.
	KmsKeyName string `pulumi:"kmsKeyName"`
}

type JobQueryDestinationEncryptionConfigurationArgs

type JobQueryDestinationEncryptionConfigurationArgs struct {
	// Describes the Cloud KMS encryption key that will be used to protect destination BigQuery table.
	// The BigQuery Service Account associated with your project requires access to this encryption key.
	KmsKeyName pulumi.StringInput `pulumi:"kmsKeyName"`
}

func (JobQueryDestinationEncryptionConfigurationArgs) ElementType

func (JobQueryDestinationEncryptionConfigurationArgs) ToJobQueryDestinationEncryptionConfigurationOutput

func (i JobQueryDestinationEncryptionConfigurationArgs) ToJobQueryDestinationEncryptionConfigurationOutput() JobQueryDestinationEncryptionConfigurationOutput

func (JobQueryDestinationEncryptionConfigurationArgs) ToJobQueryDestinationEncryptionConfigurationOutputWithContext

func (i JobQueryDestinationEncryptionConfigurationArgs) ToJobQueryDestinationEncryptionConfigurationOutputWithContext(ctx context.Context) JobQueryDestinationEncryptionConfigurationOutput

func (JobQueryDestinationEncryptionConfigurationArgs) ToJobQueryDestinationEncryptionConfigurationPtrOutput

func (i JobQueryDestinationEncryptionConfigurationArgs) ToJobQueryDestinationEncryptionConfigurationPtrOutput() JobQueryDestinationEncryptionConfigurationPtrOutput

func (JobQueryDestinationEncryptionConfigurationArgs) ToJobQueryDestinationEncryptionConfigurationPtrOutputWithContext

func (i JobQueryDestinationEncryptionConfigurationArgs) ToJobQueryDestinationEncryptionConfigurationPtrOutputWithContext(ctx context.Context) JobQueryDestinationEncryptionConfigurationPtrOutput

type JobQueryDestinationEncryptionConfigurationInput

type JobQueryDestinationEncryptionConfigurationInput interface {
	pulumi.Input

	ToJobQueryDestinationEncryptionConfigurationOutput() JobQueryDestinationEncryptionConfigurationOutput
	ToJobQueryDestinationEncryptionConfigurationOutputWithContext(context.Context) JobQueryDestinationEncryptionConfigurationOutput
}

JobQueryDestinationEncryptionConfigurationInput is an input type that accepts JobQueryDestinationEncryptionConfigurationArgs and JobQueryDestinationEncryptionConfigurationOutput values. You can construct a concrete instance of `JobQueryDestinationEncryptionConfigurationInput` via:

JobQueryDestinationEncryptionConfigurationArgs{...}

type JobQueryDestinationEncryptionConfigurationOutput

type JobQueryDestinationEncryptionConfigurationOutput struct{ *pulumi.OutputState }

func (JobQueryDestinationEncryptionConfigurationOutput) ElementType

func (JobQueryDestinationEncryptionConfigurationOutput) KmsKeyName

Describes the Cloud KMS encryption key that will be used to protect destination BigQuery table. The BigQuery Service Account associated with your project requires access to this encryption key.

func (JobQueryDestinationEncryptionConfigurationOutput) ToJobQueryDestinationEncryptionConfigurationOutput

func (o JobQueryDestinationEncryptionConfigurationOutput) ToJobQueryDestinationEncryptionConfigurationOutput() JobQueryDestinationEncryptionConfigurationOutput

func (JobQueryDestinationEncryptionConfigurationOutput) ToJobQueryDestinationEncryptionConfigurationOutputWithContext

func (o JobQueryDestinationEncryptionConfigurationOutput) ToJobQueryDestinationEncryptionConfigurationOutputWithContext(ctx context.Context) JobQueryDestinationEncryptionConfigurationOutput

func (JobQueryDestinationEncryptionConfigurationOutput) ToJobQueryDestinationEncryptionConfigurationPtrOutput

func (o JobQueryDestinationEncryptionConfigurationOutput) ToJobQueryDestinationEncryptionConfigurationPtrOutput() JobQueryDestinationEncryptionConfigurationPtrOutput

func (JobQueryDestinationEncryptionConfigurationOutput) ToJobQueryDestinationEncryptionConfigurationPtrOutputWithContext

func (o JobQueryDestinationEncryptionConfigurationOutput) ToJobQueryDestinationEncryptionConfigurationPtrOutputWithContext(ctx context.Context) JobQueryDestinationEncryptionConfigurationPtrOutput

type JobQueryDestinationEncryptionConfigurationPtrInput

type JobQueryDestinationEncryptionConfigurationPtrInput interface {
	pulumi.Input

	ToJobQueryDestinationEncryptionConfigurationPtrOutput() JobQueryDestinationEncryptionConfigurationPtrOutput
	ToJobQueryDestinationEncryptionConfigurationPtrOutputWithContext(context.Context) JobQueryDestinationEncryptionConfigurationPtrOutput
}

JobQueryDestinationEncryptionConfigurationPtrInput is an input type that accepts JobQueryDestinationEncryptionConfigurationArgs, JobQueryDestinationEncryptionConfigurationPtr and JobQueryDestinationEncryptionConfigurationPtrOutput values. You can construct a concrete instance of `JobQueryDestinationEncryptionConfigurationPtrInput` via:

        JobQueryDestinationEncryptionConfigurationArgs{...}

or:

        nil

type JobQueryDestinationEncryptionConfigurationPtrOutput

type JobQueryDestinationEncryptionConfigurationPtrOutput struct{ *pulumi.OutputState }

func (JobQueryDestinationEncryptionConfigurationPtrOutput) Elem

func (JobQueryDestinationEncryptionConfigurationPtrOutput) ElementType

func (JobQueryDestinationEncryptionConfigurationPtrOutput) KmsKeyName

Describes the Cloud KMS encryption key that will be used to protect destination BigQuery table. The BigQuery Service Account associated with your project requires access to this encryption key.

func (JobQueryDestinationEncryptionConfigurationPtrOutput) ToJobQueryDestinationEncryptionConfigurationPtrOutput

func (o JobQueryDestinationEncryptionConfigurationPtrOutput) ToJobQueryDestinationEncryptionConfigurationPtrOutput() JobQueryDestinationEncryptionConfigurationPtrOutput

func (JobQueryDestinationEncryptionConfigurationPtrOutput) ToJobQueryDestinationEncryptionConfigurationPtrOutputWithContext

func (o JobQueryDestinationEncryptionConfigurationPtrOutput) ToJobQueryDestinationEncryptionConfigurationPtrOutputWithContext(ctx context.Context) JobQueryDestinationEncryptionConfigurationPtrOutput

type JobQueryDestinationTable

type JobQueryDestinationTable struct {
	// The ID of the dataset containing this model.
	DatasetId *string `pulumi:"datasetId"`
	// The ID of the project containing this model.
	ProjectId *string `pulumi:"projectId"`
	// The table. Can be specified `{{table_id}}` if `projectId` and `datasetId` are also set,
	// or of the form `projects/{{project}}/datasets/{{dataset_id}}/tables/{{table_id}}` if not.
	TableId string `pulumi:"tableId"`
}

type JobQueryDestinationTableArgs

type JobQueryDestinationTableArgs struct {
	// The ID of the dataset containing this model.
	DatasetId pulumi.StringPtrInput `pulumi:"datasetId"`
	// The ID of the project containing this model.
	ProjectId pulumi.StringPtrInput `pulumi:"projectId"`
	// The table. Can be specified `{{table_id}}` if `projectId` and `datasetId` are also set,
	// or of the form `projects/{{project}}/datasets/{{dataset_id}}/tables/{{table_id}}` if not.
	TableId pulumi.StringInput `pulumi:"tableId"`
}

func (JobQueryDestinationTableArgs) ElementType

func (JobQueryDestinationTableArgs) ToJobQueryDestinationTableOutput

func (i JobQueryDestinationTableArgs) ToJobQueryDestinationTableOutput() JobQueryDestinationTableOutput

func (JobQueryDestinationTableArgs) ToJobQueryDestinationTableOutputWithContext

func (i JobQueryDestinationTableArgs) ToJobQueryDestinationTableOutputWithContext(ctx context.Context) JobQueryDestinationTableOutput

func (JobQueryDestinationTableArgs) ToJobQueryDestinationTablePtrOutput

func (i JobQueryDestinationTableArgs) ToJobQueryDestinationTablePtrOutput() JobQueryDestinationTablePtrOutput

func (JobQueryDestinationTableArgs) ToJobQueryDestinationTablePtrOutputWithContext

func (i JobQueryDestinationTableArgs) ToJobQueryDestinationTablePtrOutputWithContext(ctx context.Context) JobQueryDestinationTablePtrOutput

type JobQueryDestinationTableInput

type JobQueryDestinationTableInput interface {
	pulumi.Input

	ToJobQueryDestinationTableOutput() JobQueryDestinationTableOutput
	ToJobQueryDestinationTableOutputWithContext(context.Context) JobQueryDestinationTableOutput
}

JobQueryDestinationTableInput is an input type that accepts JobQueryDestinationTableArgs and JobQueryDestinationTableOutput values. You can construct a concrete instance of `JobQueryDestinationTableInput` via:

JobQueryDestinationTableArgs{...}

type JobQueryDestinationTableOutput

type JobQueryDestinationTableOutput struct{ *pulumi.OutputState }

func (JobQueryDestinationTableOutput) DatasetId

The ID of the dataset containing this model.

func (JobQueryDestinationTableOutput) ElementType

func (JobQueryDestinationTableOutput) ProjectId

The ID of the project containing this model.

func (JobQueryDestinationTableOutput) TableId

The table. Can be specified `{{table_id}}` if `projectId` and `datasetId` are also set, or of the form `projects/{{project}}/datasets/{{dataset_id}}/tables/{{table_id}}` if not.

func (JobQueryDestinationTableOutput) ToJobQueryDestinationTableOutput

func (o JobQueryDestinationTableOutput) ToJobQueryDestinationTableOutput() JobQueryDestinationTableOutput

func (JobQueryDestinationTableOutput) ToJobQueryDestinationTableOutputWithContext

func (o JobQueryDestinationTableOutput) ToJobQueryDestinationTableOutputWithContext(ctx context.Context) JobQueryDestinationTableOutput

func (JobQueryDestinationTableOutput) ToJobQueryDestinationTablePtrOutput

func (o JobQueryDestinationTableOutput) ToJobQueryDestinationTablePtrOutput() JobQueryDestinationTablePtrOutput

func (JobQueryDestinationTableOutput) ToJobQueryDestinationTablePtrOutputWithContext

func (o JobQueryDestinationTableOutput) ToJobQueryDestinationTablePtrOutputWithContext(ctx context.Context) JobQueryDestinationTablePtrOutput

type JobQueryDestinationTablePtrInput

type JobQueryDestinationTablePtrInput interface {
	pulumi.Input

	ToJobQueryDestinationTablePtrOutput() JobQueryDestinationTablePtrOutput
	ToJobQueryDestinationTablePtrOutputWithContext(context.Context) JobQueryDestinationTablePtrOutput
}

JobQueryDestinationTablePtrInput is an input type that accepts JobQueryDestinationTableArgs, JobQueryDestinationTablePtr and JobQueryDestinationTablePtrOutput values. You can construct a concrete instance of `JobQueryDestinationTablePtrInput` via:

        JobQueryDestinationTableArgs{...}

or:

        nil

type JobQueryDestinationTablePtrOutput

type JobQueryDestinationTablePtrOutput struct{ *pulumi.OutputState }

func (JobQueryDestinationTablePtrOutput) DatasetId

The ID of the dataset containing this model.

func (JobQueryDestinationTablePtrOutput) Elem

func (JobQueryDestinationTablePtrOutput) ElementType

func (JobQueryDestinationTablePtrOutput) ProjectId

The ID of the project containing this model.

func (JobQueryDestinationTablePtrOutput) TableId

The table. Can be specified `{{table_id}}` if `projectId` and `datasetId` are also set, or of the form `projects/{{project}}/datasets/{{dataset_id}}/tables/{{table_id}}` if not.

func (JobQueryDestinationTablePtrOutput) ToJobQueryDestinationTablePtrOutput

func (o JobQueryDestinationTablePtrOutput) ToJobQueryDestinationTablePtrOutput() JobQueryDestinationTablePtrOutput

func (JobQueryDestinationTablePtrOutput) ToJobQueryDestinationTablePtrOutputWithContext

func (o JobQueryDestinationTablePtrOutput) ToJobQueryDestinationTablePtrOutputWithContext(ctx context.Context) JobQueryDestinationTablePtrOutput

type JobQueryInput

type JobQueryInput interface {
	pulumi.Input

	ToJobQueryOutput() JobQueryOutput
	ToJobQueryOutputWithContext(context.Context) JobQueryOutput
}

JobQueryInput is an input type that accepts JobQueryArgs and JobQueryOutput values. You can construct a concrete instance of `JobQueryInput` via:

JobQueryArgs{...}

type JobQueryOutput

type JobQueryOutput struct{ *pulumi.OutputState }

func (JobQueryOutput) AllowLargeResults

func (o JobQueryOutput) AllowLargeResults() pulumi.BoolPtrOutput

If true and query uses legacy SQL dialect, allows the query to produce arbitrarily large result tables at a slight cost in performance. Requires destinationTable to be set. For standard SQL queries, this flag is ignored and large results are always allowed. However, you must still set destinationTable when result size exceeds the allowed maximum response size.

func (JobQueryOutput) CreateDisposition

func (o JobQueryOutput) CreateDisposition() pulumi.StringPtrOutput

Specifies whether the job is allowed to create new tables. The following values are supported: CREATE_IF_NEEDED: If the table does not exist, BigQuery creates the table. CREATE_NEVER: The table must already exist. If it does not, a 'notFound' error is returned in the job result. Creation, truncation and append actions occur as one atomic update upon job completion Default value is `CREATE_IF_NEEDED`. Possible values are `CREATE_IF_NEEDED` and `CREATE_NEVER`.

func (JobQueryOutput) DefaultDataset

Specifies the default dataset to use for unqualified table names in the query. Note that this does not alter behavior of unqualified dataset names. Structure is documented below.

func (JobQueryOutput) DestinationEncryptionConfiguration

func (o JobQueryOutput) DestinationEncryptionConfiguration() JobQueryDestinationEncryptionConfigurationPtrOutput

Custom encryption configuration (e.g., Cloud KMS keys) Structure is documented below.

func (JobQueryOutput) DestinationTable

The destination table. Structure is documented below.

func (JobQueryOutput) ElementType

func (JobQueryOutput) ElementType() reflect.Type

func (JobQueryOutput) FlattenResults

func (o JobQueryOutput) FlattenResults() pulumi.BoolPtrOutput

If true and query uses legacy SQL dialect, flattens all nested and repeated fields in the query results. allowLargeResults must be true if this is set to false. For standard SQL queries, this flag is ignored and results are never flattened.

func (JobQueryOutput) MaximumBillingTier

func (o JobQueryOutput) MaximumBillingTier() pulumi.IntPtrOutput

Limits the billing tier for this job. Queries that have resource usage beyond this tier will fail (without incurring a charge). If unspecified, this will be set to your project default.

func (JobQueryOutput) MaximumBytesBilled

func (o JobQueryOutput) MaximumBytesBilled() pulumi.StringPtrOutput

Limits the bytes billed for this job. Queries that will have bytes billed beyond this limit will fail (without incurring a charge). If unspecified, this will be set to your project default.

func (JobQueryOutput) ParameterMode

func (o JobQueryOutput) ParameterMode() pulumi.StringPtrOutput

Standard SQL only. Set to POSITIONAL to use positional (?) query parameters or to NAMED to use named (@myparam) query parameters in this query.

func (JobQueryOutput) Priority

func (o JobQueryOutput) Priority() pulumi.StringPtrOutput

Specifies a priority for the query. Default value is `INTERACTIVE`. Possible values are `INTERACTIVE` and `BATCH`.

func (JobQueryOutput) Query

Configures a query job. Structure is documented below.

func (JobQueryOutput) SchemaUpdateOptions

func (o JobQueryOutput) SchemaUpdateOptions() pulumi.StringArrayOutput

Allows the schema of the destination table to be updated as a side effect of the load job if a schema is autodetected or supplied in the job configuration. Schema update options are supported in two cases: when writeDisposition is WRITE_APPEND; when writeDisposition is WRITE_TRUNCATE and the destination table is a partition of a table, specified by partition decorators. For normal tables, WRITE_TRUNCATE will always overwrite the schema. One or more of the following values are specified: ALLOW_FIELD_ADDITION: allow adding a nullable field to the schema. ALLOW_FIELD_RELAXATION: allow relaxing a required field in the original schema to nullable.

func (JobQueryOutput) ScriptOptions

Options controlling the execution of scripts. Structure is documented below.

func (JobQueryOutput) ToJobQueryOutput

func (o JobQueryOutput) ToJobQueryOutput() JobQueryOutput

func (JobQueryOutput) ToJobQueryOutputWithContext

func (o JobQueryOutput) ToJobQueryOutputWithContext(ctx context.Context) JobQueryOutput

func (JobQueryOutput) ToJobQueryPtrOutput

func (o JobQueryOutput) ToJobQueryPtrOutput() JobQueryPtrOutput

func (JobQueryOutput) ToJobQueryPtrOutputWithContext

func (o JobQueryOutput) ToJobQueryPtrOutputWithContext(ctx context.Context) JobQueryPtrOutput

func (JobQueryOutput) UseLegacySql

func (o JobQueryOutput) UseLegacySql() pulumi.BoolPtrOutput

Specifies whether to use BigQuery's legacy SQL dialect for this query. The default value is true. If set to false, the query will use BigQuery's standard SQL.

func (JobQueryOutput) UseQueryCache

func (o JobQueryOutput) UseQueryCache() pulumi.BoolPtrOutput

Whether to look for the result in the query cache. The query cache is a best-effort cache that will be flushed whenever tables in the query are modified. Moreover, the query cache is only available when a query does not have a destination table specified. The default value is true.

func (JobQueryOutput) UserDefinedFunctionResources

func (o JobQueryOutput) UserDefinedFunctionResources() JobQueryUserDefinedFunctionResourceArrayOutput

Describes user-defined function resources used in the query. Structure is documented below.

func (JobQueryOutput) WriteDisposition

func (o JobQueryOutput) WriteDisposition() pulumi.StringPtrOutput

Specifies the action that occurs if the destination table already exists. The following values are supported: WRITE_TRUNCATE: If the table already exists, BigQuery overwrites the table data and uses the schema from the query result. WRITE_APPEND: If the table already exists, BigQuery appends the data to the table. WRITE_EMPTY: If the table already exists and contains data, a 'duplicate' error is returned in the job result. Each action is atomic and only occurs if BigQuery is able to complete the job successfully. Creation, truncation and append actions occur as one atomic update upon job completion. Default value is `WRITE_EMPTY`. Possible values are `WRITE_TRUNCATE`, `WRITE_APPEND`, and `WRITE_EMPTY`.

type JobQueryPtrInput

type JobQueryPtrInput interface {
	pulumi.Input

	ToJobQueryPtrOutput() JobQueryPtrOutput
	ToJobQueryPtrOutputWithContext(context.Context) JobQueryPtrOutput
}

JobQueryPtrInput is an input type that accepts JobQueryArgs, JobQueryPtr and JobQueryPtrOutput values. You can construct a concrete instance of `JobQueryPtrInput` via:

        JobQueryArgs{...}

or:

        nil

func JobQueryPtr

func JobQueryPtr(v *JobQueryArgs) JobQueryPtrInput

type JobQueryPtrOutput

type JobQueryPtrOutput struct{ *pulumi.OutputState }

func (JobQueryPtrOutput) AllowLargeResults

func (o JobQueryPtrOutput) AllowLargeResults() pulumi.BoolPtrOutput

If true and query uses legacy SQL dialect, allows the query to produce arbitrarily large result tables at a slight cost in performance. Requires destinationTable to be set. For standard SQL queries, this flag is ignored and large results are always allowed. However, you must still set destinationTable when result size exceeds the allowed maximum response size.

func (JobQueryPtrOutput) CreateDisposition

func (o JobQueryPtrOutput) CreateDisposition() pulumi.StringPtrOutput

Specifies whether the job is allowed to create new tables. The following values are supported: CREATE_IF_NEEDED: If the table does not exist, BigQuery creates the table. CREATE_NEVER: The table must already exist. If it does not, a 'notFound' error is returned in the job result. Creation, truncation and append actions occur as one atomic update upon job completion Default value is `CREATE_IF_NEEDED`. Possible values are `CREATE_IF_NEEDED` and `CREATE_NEVER`.

func (JobQueryPtrOutput) DefaultDataset

Specifies the default dataset to use for unqualified table names in the query. Note that this does not alter behavior of unqualified dataset names. Structure is documented below.

func (JobQueryPtrOutput) DestinationEncryptionConfiguration

func (o JobQueryPtrOutput) DestinationEncryptionConfiguration() JobQueryDestinationEncryptionConfigurationPtrOutput

Custom encryption configuration (e.g., Cloud KMS keys) Structure is documented below.

func (JobQueryPtrOutput) DestinationTable

The destination table. Structure is documented below.

func (JobQueryPtrOutput) Elem

func (JobQueryPtrOutput) ElementType

func (JobQueryPtrOutput) ElementType() reflect.Type

func (JobQueryPtrOutput) FlattenResults

func (o JobQueryPtrOutput) FlattenResults() pulumi.BoolPtrOutput

If true and query uses legacy SQL dialect, flattens all nested and repeated fields in the query results. allowLargeResults must be true if this is set to false. For standard SQL queries, this flag is ignored and results are never flattened.

func (JobQueryPtrOutput) MaximumBillingTier

func (o JobQueryPtrOutput) MaximumBillingTier() pulumi.IntPtrOutput

Limits the billing tier for this job. Queries that have resource usage beyond this tier will fail (without incurring a charge). If unspecified, this will be set to your project default.

func (JobQueryPtrOutput) MaximumBytesBilled

func (o JobQueryPtrOutput) MaximumBytesBilled() pulumi.StringPtrOutput

Limits the bytes billed for this job. Queries that will have bytes billed beyond this limit will fail (without incurring a charge). If unspecified, this will be set to your project default.

func (JobQueryPtrOutput) ParameterMode

func (o JobQueryPtrOutput) ParameterMode() pulumi.StringPtrOutput

Standard SQL only. Set to POSITIONAL to use positional (?) query parameters or to NAMED to use named (@myparam) query parameters in this query.

func (JobQueryPtrOutput) Priority

Specifies a priority for the query. Default value is `INTERACTIVE`. Possible values are `INTERACTIVE` and `BATCH`.

func (JobQueryPtrOutput) Query

Configures a query job. Structure is documented below.

func (JobQueryPtrOutput) SchemaUpdateOptions

func (o JobQueryPtrOutput) SchemaUpdateOptions() pulumi.StringArrayOutput

Allows the schema of the destination table to be updated as a side effect of the load job if a schema is autodetected or supplied in the job configuration. Schema update options are supported in two cases: when writeDisposition is WRITE_APPEND; when writeDisposition is WRITE_TRUNCATE and the destination table is a partition of a table, specified by partition decorators. For normal tables, WRITE_TRUNCATE will always overwrite the schema. One or more of the following values are specified: ALLOW_FIELD_ADDITION: allow adding a nullable field to the schema. ALLOW_FIELD_RELAXATION: allow relaxing a required field in the original schema to nullable.

func (JobQueryPtrOutput) ScriptOptions

Options controlling the execution of scripts. Structure is documented below.

func (JobQueryPtrOutput) ToJobQueryPtrOutput

func (o JobQueryPtrOutput) ToJobQueryPtrOutput() JobQueryPtrOutput

func (JobQueryPtrOutput) ToJobQueryPtrOutputWithContext

func (o JobQueryPtrOutput) ToJobQueryPtrOutputWithContext(ctx context.Context) JobQueryPtrOutput

func (JobQueryPtrOutput) UseLegacySql

func (o JobQueryPtrOutput) UseLegacySql() pulumi.BoolPtrOutput

Specifies whether to use BigQuery's legacy SQL dialect for this query. The default value is true. If set to false, the query will use BigQuery's standard SQL.

func (JobQueryPtrOutput) UseQueryCache

func (o JobQueryPtrOutput) UseQueryCache() pulumi.BoolPtrOutput

Whether to look for the result in the query cache. The query cache is a best-effort cache that will be flushed whenever tables in the query are modified. Moreover, the query cache is only available when a query does not have a destination table specified. The default value is true.

func (JobQueryPtrOutput) UserDefinedFunctionResources

func (o JobQueryPtrOutput) UserDefinedFunctionResources() JobQueryUserDefinedFunctionResourceArrayOutput

Describes user-defined function resources used in the query. Structure is documented below.

func (JobQueryPtrOutput) WriteDisposition

func (o JobQueryPtrOutput) WriteDisposition() pulumi.StringPtrOutput

Specifies the action that occurs if the destination table already exists. The following values are supported: WRITE_TRUNCATE: If the table already exists, BigQuery overwrites the table data and uses the schema from the query result. WRITE_APPEND: If the table already exists, BigQuery appends the data to the table. WRITE_EMPTY: If the table already exists and contains data, a 'duplicate' error is returned in the job result. Each action is atomic and only occurs if BigQuery is able to complete the job successfully. Creation, truncation and append actions occur as one atomic update upon job completion. Default value is `WRITE_EMPTY`. Possible values are `WRITE_TRUNCATE`, `WRITE_APPEND`, and `WRITE_EMPTY`.

type JobQueryScriptOptions

type JobQueryScriptOptions struct {
	// Determines which statement in the script represents the "key result",
	// used to populate the schema and query results of the script job.
	// Possible values are `LAST` and `FIRST_SELECT`.
	KeyResultStatement *string `pulumi:"keyResultStatement"`
	// Limit on the number of bytes billed per statement. Exceeding this budget results in an error.
	StatementByteBudget *string `pulumi:"statementByteBudget"`
	// Timeout period for each statement in a script.
	StatementTimeoutMs *string `pulumi:"statementTimeoutMs"`
}

type JobQueryScriptOptionsArgs

type JobQueryScriptOptionsArgs struct {
	// Determines which statement in the script represents the "key result",
	// used to populate the schema and query results of the script job.
	// Possible values are `LAST` and `FIRST_SELECT`.
	KeyResultStatement pulumi.StringPtrInput `pulumi:"keyResultStatement"`
	// Limit on the number of bytes billed per statement. Exceeding this budget results in an error.
	StatementByteBudget pulumi.StringPtrInput `pulumi:"statementByteBudget"`
	// Timeout period for each statement in a script.
	StatementTimeoutMs pulumi.StringPtrInput `pulumi:"statementTimeoutMs"`
}

func (JobQueryScriptOptionsArgs) ElementType

func (JobQueryScriptOptionsArgs) ElementType() reflect.Type

func (JobQueryScriptOptionsArgs) ToJobQueryScriptOptionsOutput

func (i JobQueryScriptOptionsArgs) ToJobQueryScriptOptionsOutput() JobQueryScriptOptionsOutput

func (JobQueryScriptOptionsArgs) ToJobQueryScriptOptionsOutputWithContext

func (i JobQueryScriptOptionsArgs) ToJobQueryScriptOptionsOutputWithContext(ctx context.Context) JobQueryScriptOptionsOutput

func (JobQueryScriptOptionsArgs) ToJobQueryScriptOptionsPtrOutput

func (i JobQueryScriptOptionsArgs) ToJobQueryScriptOptionsPtrOutput() JobQueryScriptOptionsPtrOutput

func (JobQueryScriptOptionsArgs) ToJobQueryScriptOptionsPtrOutputWithContext

func (i JobQueryScriptOptionsArgs) ToJobQueryScriptOptionsPtrOutputWithContext(ctx context.Context) JobQueryScriptOptionsPtrOutput

type JobQueryScriptOptionsInput

type JobQueryScriptOptionsInput interface {
	pulumi.Input

	ToJobQueryScriptOptionsOutput() JobQueryScriptOptionsOutput
	ToJobQueryScriptOptionsOutputWithContext(context.Context) JobQueryScriptOptionsOutput
}

JobQueryScriptOptionsInput is an input type that accepts JobQueryScriptOptionsArgs and JobQueryScriptOptionsOutput values. You can construct a concrete instance of `JobQueryScriptOptionsInput` via:

JobQueryScriptOptionsArgs{...}

type JobQueryScriptOptionsOutput

type JobQueryScriptOptionsOutput struct{ *pulumi.OutputState }

func (JobQueryScriptOptionsOutput) ElementType

func (JobQueryScriptOptionsOutput) KeyResultStatement

func (o JobQueryScriptOptionsOutput) KeyResultStatement() pulumi.StringPtrOutput

Determines which statement in the script represents the "key result", used to populate the schema and query results of the script job. Possible values are `LAST` and `FIRST_SELECT`.

func (JobQueryScriptOptionsOutput) StatementByteBudget

func (o JobQueryScriptOptionsOutput) StatementByteBudget() pulumi.StringPtrOutput

Limit on the number of bytes billed per statement. Exceeding this budget results in an error.

func (JobQueryScriptOptionsOutput) StatementTimeoutMs

func (o JobQueryScriptOptionsOutput) StatementTimeoutMs() pulumi.StringPtrOutput

Timeout period for each statement in a script.

func (JobQueryScriptOptionsOutput) ToJobQueryScriptOptionsOutput

func (o JobQueryScriptOptionsOutput) ToJobQueryScriptOptionsOutput() JobQueryScriptOptionsOutput

func (JobQueryScriptOptionsOutput) ToJobQueryScriptOptionsOutputWithContext

func (o JobQueryScriptOptionsOutput) ToJobQueryScriptOptionsOutputWithContext(ctx context.Context) JobQueryScriptOptionsOutput

func (JobQueryScriptOptionsOutput) ToJobQueryScriptOptionsPtrOutput

func (o JobQueryScriptOptionsOutput) ToJobQueryScriptOptionsPtrOutput() JobQueryScriptOptionsPtrOutput

func (JobQueryScriptOptionsOutput) ToJobQueryScriptOptionsPtrOutputWithContext

func (o JobQueryScriptOptionsOutput) ToJobQueryScriptOptionsPtrOutputWithContext(ctx context.Context) JobQueryScriptOptionsPtrOutput

type JobQueryScriptOptionsPtrInput

type JobQueryScriptOptionsPtrInput interface {
	pulumi.Input

	ToJobQueryScriptOptionsPtrOutput() JobQueryScriptOptionsPtrOutput
	ToJobQueryScriptOptionsPtrOutputWithContext(context.Context) JobQueryScriptOptionsPtrOutput
}

JobQueryScriptOptionsPtrInput is an input type that accepts JobQueryScriptOptionsArgs, JobQueryScriptOptionsPtr and JobQueryScriptOptionsPtrOutput values. You can construct a concrete instance of `JobQueryScriptOptionsPtrInput` via:

        JobQueryScriptOptionsArgs{...}

or:

        nil

type JobQueryScriptOptionsPtrOutput

type JobQueryScriptOptionsPtrOutput struct{ *pulumi.OutputState }

func (JobQueryScriptOptionsPtrOutput) Elem

func (JobQueryScriptOptionsPtrOutput) ElementType

func (JobQueryScriptOptionsPtrOutput) KeyResultStatement

func (o JobQueryScriptOptionsPtrOutput) KeyResultStatement() pulumi.StringPtrOutput

Determines which statement in the script represents the "key result", used to populate the schema and query results of the script job. Possible values are `LAST` and `FIRST_SELECT`.

func (JobQueryScriptOptionsPtrOutput) StatementByteBudget

func (o JobQueryScriptOptionsPtrOutput) StatementByteBudget() pulumi.StringPtrOutput

Limit on the number of bytes billed per statement. Exceeding this budget results in an error.

func (JobQueryScriptOptionsPtrOutput) StatementTimeoutMs

func (o JobQueryScriptOptionsPtrOutput) StatementTimeoutMs() pulumi.StringPtrOutput

Timeout period for each statement in a script.

func (JobQueryScriptOptionsPtrOutput) ToJobQueryScriptOptionsPtrOutput

func (o JobQueryScriptOptionsPtrOutput) ToJobQueryScriptOptionsPtrOutput() JobQueryScriptOptionsPtrOutput

func (JobQueryScriptOptionsPtrOutput) ToJobQueryScriptOptionsPtrOutputWithContext

func (o JobQueryScriptOptionsPtrOutput) ToJobQueryScriptOptionsPtrOutputWithContext(ctx context.Context) JobQueryScriptOptionsPtrOutput

type JobQueryUserDefinedFunctionResource

type JobQueryUserDefinedFunctionResource struct {
	// An inline resource that contains code for a user-defined function (UDF).
	// Providing a inline code resource is equivalent to providing a URI for a file containing the same code.
	InlineCode *string `pulumi:"inlineCode"`
	// A code resource to load from a Google Cloud Storage URI (gs://bucket/path).
	ResourceUri *string `pulumi:"resourceUri"`
}

type JobQueryUserDefinedFunctionResourceArgs

type JobQueryUserDefinedFunctionResourceArgs struct {
	// An inline resource that contains code for a user-defined function (UDF).
	// Providing a inline code resource is equivalent to providing a URI for a file containing the same code.
	InlineCode pulumi.StringPtrInput `pulumi:"inlineCode"`
	// A code resource to load from a Google Cloud Storage URI (gs://bucket/path).
	ResourceUri pulumi.StringPtrInput `pulumi:"resourceUri"`
}

func (JobQueryUserDefinedFunctionResourceArgs) ElementType

func (JobQueryUserDefinedFunctionResourceArgs) ToJobQueryUserDefinedFunctionResourceOutput

func (i JobQueryUserDefinedFunctionResourceArgs) ToJobQueryUserDefinedFunctionResourceOutput() JobQueryUserDefinedFunctionResourceOutput

func (JobQueryUserDefinedFunctionResourceArgs) ToJobQueryUserDefinedFunctionResourceOutputWithContext

func (i JobQueryUserDefinedFunctionResourceArgs) ToJobQueryUserDefinedFunctionResourceOutputWithContext(ctx context.Context) JobQueryUserDefinedFunctionResourceOutput

type JobQueryUserDefinedFunctionResourceArray

type JobQueryUserDefinedFunctionResourceArray []JobQueryUserDefinedFunctionResourceInput

func (JobQueryUserDefinedFunctionResourceArray) ElementType

func (JobQueryUserDefinedFunctionResourceArray) ToJobQueryUserDefinedFunctionResourceArrayOutput

func (i JobQueryUserDefinedFunctionResourceArray) ToJobQueryUserDefinedFunctionResourceArrayOutput() JobQueryUserDefinedFunctionResourceArrayOutput

func (JobQueryUserDefinedFunctionResourceArray) ToJobQueryUserDefinedFunctionResourceArrayOutputWithContext

func (i JobQueryUserDefinedFunctionResourceArray) ToJobQueryUserDefinedFunctionResourceArrayOutputWithContext(ctx context.Context) JobQueryUserDefinedFunctionResourceArrayOutput

type JobQueryUserDefinedFunctionResourceArrayInput

type JobQueryUserDefinedFunctionResourceArrayInput interface {
	pulumi.Input

	ToJobQueryUserDefinedFunctionResourceArrayOutput() JobQueryUserDefinedFunctionResourceArrayOutput
	ToJobQueryUserDefinedFunctionResourceArrayOutputWithContext(context.Context) JobQueryUserDefinedFunctionResourceArrayOutput
}

JobQueryUserDefinedFunctionResourceArrayInput is an input type that accepts JobQueryUserDefinedFunctionResourceArray and JobQueryUserDefinedFunctionResourceArrayOutput values. You can construct a concrete instance of `JobQueryUserDefinedFunctionResourceArrayInput` via:

JobQueryUserDefinedFunctionResourceArray{ JobQueryUserDefinedFunctionResourceArgs{...} }

type JobQueryUserDefinedFunctionResourceArrayOutput

type JobQueryUserDefinedFunctionResourceArrayOutput struct{ *pulumi.OutputState }

func (JobQueryUserDefinedFunctionResourceArrayOutput) ElementType

func (JobQueryUserDefinedFunctionResourceArrayOutput) Index

func (JobQueryUserDefinedFunctionResourceArrayOutput) ToJobQueryUserDefinedFunctionResourceArrayOutput

func (o JobQueryUserDefinedFunctionResourceArrayOutput) ToJobQueryUserDefinedFunctionResourceArrayOutput() JobQueryUserDefinedFunctionResourceArrayOutput

func (JobQueryUserDefinedFunctionResourceArrayOutput) ToJobQueryUserDefinedFunctionResourceArrayOutputWithContext

func (o JobQueryUserDefinedFunctionResourceArrayOutput) ToJobQueryUserDefinedFunctionResourceArrayOutputWithContext(ctx context.Context) JobQueryUserDefinedFunctionResourceArrayOutput

type JobQueryUserDefinedFunctionResourceInput

type JobQueryUserDefinedFunctionResourceInput interface {
	pulumi.Input

	ToJobQueryUserDefinedFunctionResourceOutput() JobQueryUserDefinedFunctionResourceOutput
	ToJobQueryUserDefinedFunctionResourceOutputWithContext(context.Context) JobQueryUserDefinedFunctionResourceOutput
}

JobQueryUserDefinedFunctionResourceInput is an input type that accepts JobQueryUserDefinedFunctionResourceArgs and JobQueryUserDefinedFunctionResourceOutput values. You can construct a concrete instance of `JobQueryUserDefinedFunctionResourceInput` via:

JobQueryUserDefinedFunctionResourceArgs{...}

type JobQueryUserDefinedFunctionResourceOutput

type JobQueryUserDefinedFunctionResourceOutput struct{ *pulumi.OutputState }

func (JobQueryUserDefinedFunctionResourceOutput) ElementType

func (JobQueryUserDefinedFunctionResourceOutput) InlineCode

An inline resource that contains code for a user-defined function (UDF). Providing a inline code resource is equivalent to providing a URI for a file containing the same code.

func (JobQueryUserDefinedFunctionResourceOutput) ResourceUri

A code resource to load from a Google Cloud Storage URI (gs://bucket/path).

func (JobQueryUserDefinedFunctionResourceOutput) ToJobQueryUserDefinedFunctionResourceOutput

func (o JobQueryUserDefinedFunctionResourceOutput) ToJobQueryUserDefinedFunctionResourceOutput() JobQueryUserDefinedFunctionResourceOutput

func (JobQueryUserDefinedFunctionResourceOutput) ToJobQueryUserDefinedFunctionResourceOutputWithContext

func (o JobQueryUserDefinedFunctionResourceOutput) ToJobQueryUserDefinedFunctionResourceOutputWithContext(ctx context.Context) JobQueryUserDefinedFunctionResourceOutput

type JobState

type JobState struct {
	// Copies a table.
	// Structure is documented below.
	Copy JobCopyPtrInput
	// Configures an extract job.
	// Structure is documented below.
	Extract JobExtractPtrInput
	// The ID of the job. The ID must contain only letters (a-z, A-Z), numbers (0-9), underscores (_), or dashes (-). The maximum length is 1,024 characters.
	JobId pulumi.StringPtrInput
	// Job timeout in milliseconds. If this time limit is exceeded, BigQuery may attempt to terminate the job.
	JobTimeoutMs pulumi.StringPtrInput
	// The type of the job.
	JobType pulumi.StringPtrInput
	// The labels associated with this job. You can use these to organize and group your jobs.
	Labels pulumi.StringMapInput
	// Configures a load job.
	// Structure is documented below.
	Load JobLoadPtrInput
	// The geographic location of the job. The default value is US.
	Location pulumi.StringPtrInput
	// The ID of the project in which the resource belongs.
	// If it is not provided, the provider project is used.
	Project pulumi.StringPtrInput
	// Configures a query job.
	// Structure is documented below.
	Query JobQueryPtrInput
	// Email address of the user who ran the job.
	UserEmail pulumi.StringPtrInput
}

func (JobState) ElementType

func (JobState) ElementType() reflect.Type

type Reservation

type Reservation struct {
	pulumi.CustomResourceState

	// If false, any query using this reservation will use idle slots from other reservations within
	// the same admin project. If true, a query using this reservation will execute with the slot
	// capacity specified above at most.
	IgnoreIdleSlots pulumi.BoolPtrOutput `pulumi:"ignoreIdleSlots"`
	// The geographic location where the transfer config should reside.
	// Examples: US, EU, asia-northeast1. The default value is US.
	Location pulumi.StringPtrOutput `pulumi:"location"`
	// The name of the reservation. This field must only contain alphanumeric characters or dash.
	Name pulumi.StringOutput `pulumi:"name"`
	// The ID of the project in which the resource belongs.
	// If it is not provided, the provider project is used.
	Project pulumi.StringOutput `pulumi:"project"`
	// Minimum slots available to this reservation. A slot is a unit of computational power in BigQuery, and serves as the
	// unit of parallelism. Queries using this reservation might use more slots during runtime if ignoreIdleSlots is set to false.
	SlotCapacity pulumi.IntOutput `pulumi:"slotCapacity"`
}

A reservation is a mechanism used to guarantee BigQuery slots to users.

To get more information about Reservation, see:

* [API documentation](https://cloud.google.com/bigquery/docs/reference/reservations/rest/v1beta1/projects.locations.reservations/create) * How-to Guides

## Example Usage

func GetReservation

func GetReservation(ctx *pulumi.Context,
	name string, id pulumi.IDInput, state *ReservationState, opts ...pulumi.ResourceOption) (*Reservation, error)

GetReservation gets an existing Reservation resource's state with the given name, ID, and optional state properties that are used to uniquely qualify the lookup (nil if not required).

func NewReservation

func NewReservation(ctx *pulumi.Context,
	name string, args *ReservationArgs, opts ...pulumi.ResourceOption) (*Reservation, error)

NewReservation registers a new resource with the given unique name, arguments, and options.

type ReservationArgs

type ReservationArgs struct {
	// If false, any query using this reservation will use idle slots from other reservations within
	// the same admin project. If true, a query using this reservation will execute with the slot
	// capacity specified above at most.
	IgnoreIdleSlots pulumi.BoolPtrInput
	// The geographic location where the transfer config should reside.
	// Examples: US, EU, asia-northeast1. The default value is US.
	Location pulumi.StringPtrInput
	// The name of the reservation. This field must only contain alphanumeric characters or dash.
	Name pulumi.StringPtrInput
	// The ID of the project in which the resource belongs.
	// If it is not provided, the provider project is used.
	Project pulumi.StringPtrInput
	// Minimum slots available to this reservation. A slot is a unit of computational power in BigQuery, and serves as the
	// unit of parallelism. Queries using this reservation might use more slots during runtime if ignoreIdleSlots is set to false.
	SlotCapacity pulumi.IntInput
}

The set of arguments for constructing a Reservation resource.

func (ReservationArgs) ElementType

func (ReservationArgs) ElementType() reflect.Type

type ReservationState

type ReservationState struct {
	// If false, any query using this reservation will use idle slots from other reservations within
	// the same admin project. If true, a query using this reservation will execute with the slot
	// capacity specified above at most.
	IgnoreIdleSlots pulumi.BoolPtrInput
	// The geographic location where the transfer config should reside.
	// Examples: US, EU, asia-northeast1. The default value is US.
	Location pulumi.StringPtrInput
	// The name of the reservation. This field must only contain alphanumeric characters or dash.
	Name pulumi.StringPtrInput
	// The ID of the project in which the resource belongs.
	// If it is not provided, the provider project is used.
	Project pulumi.StringPtrInput
	// Minimum slots available to this reservation. A slot is a unit of computational power in BigQuery, and serves as the
	// unit of parallelism. Queries using this reservation might use more slots during runtime if ignoreIdleSlots is set to false.
	SlotCapacity pulumi.IntPtrInput
}

func (ReservationState) ElementType

func (ReservationState) ElementType() reflect.Type

type Table

type Table struct {
	pulumi.CustomResourceState

	// Specifies column names to use for data clustering.
	// Up to four top-level columns are allowed, and should be specified in
	// descending priority order.
	Clusterings pulumi.StringArrayOutput `pulumi:"clusterings"`
	// The time when this table was created, in milliseconds since the epoch.
	CreationTime pulumi.IntOutput `pulumi:"creationTime"`
	// The dataset ID to create the table in.
	// Changing this forces a new resource to be created.
	DatasetId pulumi.StringOutput `pulumi:"datasetId"`
	// The field description.
	Description pulumi.StringPtrOutput `pulumi:"description"`
	// Specifies how the table should be encrypted.
	// If left blank, the table will be encrypted with a Google-managed key; that process
	// is transparent to the user.  Structure is documented below.
	EncryptionConfiguration TableEncryptionConfigurationPtrOutput `pulumi:"encryptionConfiguration"`
	// A hash of the resource.
	Etag pulumi.StringOutput `pulumi:"etag"`
	// The time when this table expires, in
	// milliseconds since the epoch. If not present, the table will persist
	// indefinitely. Expired tables will be deleted and their storage
	// reclaimed.
	ExpirationTime pulumi.IntOutput `pulumi:"expirationTime"`
	// Describes the data format,
	// location, and other properties of a table stored outside of BigQuery.
	// By defining these properties, the data source can then be queried as
	// if it were a standard BigQuery table. Structure is documented below.
	ExternalDataConfiguration TableExternalDataConfigurationPtrOutput `pulumi:"externalDataConfiguration"`
	// A descriptive name for the table.
	FriendlyName pulumi.StringPtrOutput `pulumi:"friendlyName"`
	// A mapping of labels to assign to the resource.
	Labels pulumi.StringMapOutput `pulumi:"labels"`
	// The time when this table was last modified, in milliseconds since the epoch.
	LastModifiedTime pulumi.IntOutput `pulumi:"lastModifiedTime"`
	// The geographic location where the table resides. This value is inherited from the dataset.
	Location pulumi.StringOutput `pulumi:"location"`
	// If specified, configures this table as a materialized view.
	// Structure is documented below.
	MaterializedView TableMaterializedViewPtrOutput `pulumi:"materializedView"`
	// The size of this table in bytes, excluding any data in the streaming buffer.
	NumBytes pulumi.IntOutput `pulumi:"numBytes"`
	// The number of bytes in the table that are considered "long-term storage".
	NumLongTermBytes pulumi.IntOutput `pulumi:"numLongTermBytes"`
	// The number of rows of data in this table, excluding any data in the streaming buffer.
	NumRows pulumi.IntOutput `pulumi:"numRows"`
	// The ID of the project in which the resource belongs. If it
	// is not provided, the provider project is used.
	Project pulumi.StringOutput `pulumi:"project"`
	// If specified, configures range-based
	// partitioning for this table. Structure is documented below.
	RangePartitioning TableRangePartitioningPtrOutput `pulumi:"rangePartitioning"`
	// A JSON schema for the external table. Schema is required
	// for CSV and JSON formats if autodetect is not on. Schema is disallowed
	// for Google Cloud Bigtable, Cloud Datastore backups, Avro, ORC and Parquet formats.
	// ~>**NOTE:** Because this field expects a JSON string, any changes to the
	// string will create a diff, even if the JSON itself hasn't changed.
	// Furthermore drift for this field cannot not be detected because BigQuery
	// only uses this schema to compute the effective schema for the table, therefore
	// any changes on the configured value will force the table to be recreated.
	// This schema is effectively only applied when creating a table from an external
	// datasource, after creation the computed schema will be stored in
	// `google_bigquery_table.schema`
	Schema pulumi.StringOutput `pulumi:"schema"`
	// The URI of the created resource.
	SelfLink pulumi.StringOutput `pulumi:"selfLink"`
	// A unique ID for the resource.
	// Changing this forces a new resource to be created.
	TableId pulumi.StringOutput `pulumi:"tableId"`
	// If specified, configures time-based
	// partitioning for this table. Structure is documented below.
	TimePartitioning TableTimePartitioningPtrOutput `pulumi:"timePartitioning"`
	// The supported types are DAY, HOUR, MONTH, and YEAR,
	// which will generate one partition per day, hour, month, and year, respectively.
	Type pulumi.StringOutput `pulumi:"type"`
	// If specified, configures this table as a view.
	// Structure is documented below.
	View TableViewPtrOutput `pulumi:"view"`
}

Creates a table resource in a dataset for Google BigQuery. For more information see [the official documentation](https://cloud.google.com/bigquery/docs/) and [API](https://cloud.google.com/bigquery/docs/reference/rest/v2/tables).

func GetTable

func GetTable(ctx *pulumi.Context,
	name string, id pulumi.IDInput, state *TableState, opts ...pulumi.ResourceOption) (*Table, error)

GetTable gets an existing Table resource's state with the given name, ID, and optional state properties that are used to uniquely qualify the lookup (nil if not required).

func NewTable

func NewTable(ctx *pulumi.Context,
	name string, args *TableArgs, opts ...pulumi.ResourceOption) (*Table, error)

NewTable registers a new resource with the given unique name, arguments, and options.

type TableArgs

type TableArgs struct {
	// Specifies column names to use for data clustering.
	// Up to four top-level columns are allowed, and should be specified in
	// descending priority order.
	Clusterings pulumi.StringArrayInput
	// The dataset ID to create the table in.
	// Changing this forces a new resource to be created.
	DatasetId pulumi.StringInput
	// The field description.
	Description pulumi.StringPtrInput
	// Specifies how the table should be encrypted.
	// If left blank, the table will be encrypted with a Google-managed key; that process
	// is transparent to the user.  Structure is documented below.
	EncryptionConfiguration TableEncryptionConfigurationPtrInput
	// The time when this table expires, in
	// milliseconds since the epoch. If not present, the table will persist
	// indefinitely. Expired tables will be deleted and their storage
	// reclaimed.
	ExpirationTime pulumi.IntPtrInput
	// Describes the data format,
	// location, and other properties of a table stored outside of BigQuery.
	// By defining these properties, the data source can then be queried as
	// if it were a standard BigQuery table. Structure is documented below.
	ExternalDataConfiguration TableExternalDataConfigurationPtrInput
	// A descriptive name for the table.
	FriendlyName pulumi.StringPtrInput
	// A mapping of labels to assign to the resource.
	Labels pulumi.StringMapInput
	// If specified, configures this table as a materialized view.
	// Structure is documented below.
	MaterializedView TableMaterializedViewPtrInput
	// The ID of the project in which the resource belongs. If it
	// is not provided, the provider project is used.
	Project pulumi.StringPtrInput
	// If specified, configures range-based
	// partitioning for this table. Structure is documented below.
	RangePartitioning TableRangePartitioningPtrInput
	// A JSON schema for the external table. Schema is required
	// for CSV and JSON formats if autodetect is not on. Schema is disallowed
	// for Google Cloud Bigtable, Cloud Datastore backups, Avro, ORC and Parquet formats.
	// ~>**NOTE:** Because this field expects a JSON string, any changes to the
	// string will create a diff, even if the JSON itself hasn't changed.
	// Furthermore drift for this field cannot not be detected because BigQuery
	// only uses this schema to compute the effective schema for the table, therefore
	// any changes on the configured value will force the table to be recreated.
	// This schema is effectively only applied when creating a table from an external
	// datasource, after creation the computed schema will be stored in
	// `google_bigquery_table.schema`
	Schema pulumi.StringPtrInput
	// A unique ID for the resource.
	// Changing this forces a new resource to be created.
	TableId pulumi.StringInput
	// If specified, configures time-based
	// partitioning for this table. Structure is documented below.
	TimePartitioning TableTimePartitioningPtrInput
	// If specified, configures this table as a view.
	// Structure is documented below.
	View TableViewPtrInput
}

The set of arguments for constructing a Table resource.

func (TableArgs) ElementType

func (TableArgs) ElementType() reflect.Type

type TableEncryptionConfiguration

type TableEncryptionConfiguration struct {
	// The self link or full name of a key which should be used to
	// encrypt this table.  Note that the default bigquery service account will need to have
	// encrypt/decrypt permissions on this key - you may want to see the
	// `bigquery.getDefaultServiceAccount` datasource and the
	// `kms.CryptoKeyIAMBinding` resource.
	KmsKeyName string `pulumi:"kmsKeyName"`
}

type TableEncryptionConfigurationArgs

type TableEncryptionConfigurationArgs struct {
	// The self link or full name of a key which should be used to
	// encrypt this table.  Note that the default bigquery service account will need to have
	// encrypt/decrypt permissions on this key - you may want to see the
	// `bigquery.getDefaultServiceAccount` datasource and the
	// `kms.CryptoKeyIAMBinding` resource.
	KmsKeyName pulumi.StringInput `pulumi:"kmsKeyName"`
}

func (TableEncryptionConfigurationArgs) ElementType

func (TableEncryptionConfigurationArgs) ToTableEncryptionConfigurationOutput

func (i TableEncryptionConfigurationArgs) ToTableEncryptionConfigurationOutput() TableEncryptionConfigurationOutput

func (TableEncryptionConfigurationArgs) ToTableEncryptionConfigurationOutputWithContext

func (i TableEncryptionConfigurationArgs) ToTableEncryptionConfigurationOutputWithContext(ctx context.Context) TableEncryptionConfigurationOutput

func (TableEncryptionConfigurationArgs) ToTableEncryptionConfigurationPtrOutput

func (i TableEncryptionConfigurationArgs) ToTableEncryptionConfigurationPtrOutput() TableEncryptionConfigurationPtrOutput

func (TableEncryptionConfigurationArgs) ToTableEncryptionConfigurationPtrOutputWithContext

func (i TableEncryptionConfigurationArgs) ToTableEncryptionConfigurationPtrOutputWithContext(ctx context.Context) TableEncryptionConfigurationPtrOutput

type TableEncryptionConfigurationInput

type TableEncryptionConfigurationInput interface {
	pulumi.Input

	ToTableEncryptionConfigurationOutput() TableEncryptionConfigurationOutput
	ToTableEncryptionConfigurationOutputWithContext(context.Context) TableEncryptionConfigurationOutput
}

TableEncryptionConfigurationInput is an input type that accepts TableEncryptionConfigurationArgs and TableEncryptionConfigurationOutput values. You can construct a concrete instance of `TableEncryptionConfigurationInput` via:

TableEncryptionConfigurationArgs{...}

type TableEncryptionConfigurationOutput

type TableEncryptionConfigurationOutput struct{ *pulumi.OutputState }

func (TableEncryptionConfigurationOutput) ElementType

func (TableEncryptionConfigurationOutput) KmsKeyName

The self link or full name of a key which should be used to encrypt this table. Note that the default bigquery service account will need to have encrypt/decrypt permissions on this key - you may want to see the `bigquery.getDefaultServiceAccount` datasource and the `kms.CryptoKeyIAMBinding` resource.

func (TableEncryptionConfigurationOutput) ToTableEncryptionConfigurationOutput

func (o TableEncryptionConfigurationOutput) ToTableEncryptionConfigurationOutput() TableEncryptionConfigurationOutput

func (TableEncryptionConfigurationOutput) ToTableEncryptionConfigurationOutputWithContext

func (o TableEncryptionConfigurationOutput) ToTableEncryptionConfigurationOutputWithContext(ctx context.Context) TableEncryptionConfigurationOutput

func (TableEncryptionConfigurationOutput) ToTableEncryptionConfigurationPtrOutput

func (o TableEncryptionConfigurationOutput) ToTableEncryptionConfigurationPtrOutput() TableEncryptionConfigurationPtrOutput

func (TableEncryptionConfigurationOutput) ToTableEncryptionConfigurationPtrOutputWithContext

func (o TableEncryptionConfigurationOutput) ToTableEncryptionConfigurationPtrOutputWithContext(ctx context.Context) TableEncryptionConfigurationPtrOutput

type TableEncryptionConfigurationPtrInput

type TableEncryptionConfigurationPtrInput interface {
	pulumi.Input

	ToTableEncryptionConfigurationPtrOutput() TableEncryptionConfigurationPtrOutput
	ToTableEncryptionConfigurationPtrOutputWithContext(context.Context) TableEncryptionConfigurationPtrOutput
}

TableEncryptionConfigurationPtrInput is an input type that accepts TableEncryptionConfigurationArgs, TableEncryptionConfigurationPtr and TableEncryptionConfigurationPtrOutput values. You can construct a concrete instance of `TableEncryptionConfigurationPtrInput` via:

        TableEncryptionConfigurationArgs{...}

or:

        nil

type TableEncryptionConfigurationPtrOutput

type TableEncryptionConfigurationPtrOutput struct{ *pulumi.OutputState }

func (TableEncryptionConfigurationPtrOutput) Elem

func (TableEncryptionConfigurationPtrOutput) ElementType

func (TableEncryptionConfigurationPtrOutput) KmsKeyName

The self link or full name of a key which should be used to encrypt this table. Note that the default bigquery service account will need to have encrypt/decrypt permissions on this key - you may want to see the `bigquery.getDefaultServiceAccount` datasource and the `kms.CryptoKeyIAMBinding` resource.

func (TableEncryptionConfigurationPtrOutput) ToTableEncryptionConfigurationPtrOutput

func (o TableEncryptionConfigurationPtrOutput) ToTableEncryptionConfigurationPtrOutput() TableEncryptionConfigurationPtrOutput

func (TableEncryptionConfigurationPtrOutput) ToTableEncryptionConfigurationPtrOutputWithContext

func (o TableEncryptionConfigurationPtrOutput) ToTableEncryptionConfigurationPtrOutputWithContext(ctx context.Context) TableEncryptionConfigurationPtrOutput

type TableExternalDataConfiguration

type TableExternalDataConfiguration struct {
	// - Let BigQuery try to autodetect the schema
	//   and format of the table.
	Autodetect bool `pulumi:"autodetect"`
	// The compression type of the data source.
	// Valid values are "NONE" or "GZIP".
	Compression *string `pulumi:"compression"`
	// Additional properties to set if
	// `sourceFormat` is set to "CSV". Structure is documented below.
	CsvOptions *TableExternalDataConfigurationCsvOptions `pulumi:"csvOptions"`
	// Additional options if
	// `sourceFormat` is set to "GOOGLE_SHEETS". Structure is
	// documented below.
	GoogleSheetsOptions *TableExternalDataConfigurationGoogleSheetsOptions `pulumi:"googleSheetsOptions"`
	// When set, configures hive partitioning
	// support. Not all storage formats support hive partitioning -- requesting hive
	// partitioning on an unsupported format will lead to an error, as will providing
	// an invalid specification.
	HivePartitioningOptions *TableExternalDataConfigurationHivePartitioningOptions `pulumi:"hivePartitioningOptions"`
	// Indicates if BigQuery should
	// allow extra values that are not represented in the table schema.
	// If true, the extra values are ignored. If false, records with
	// extra columns are treated as bad records, and if there are too
	// many bad records, an invalid error is returned in the job result.
	// The default value is false.
	IgnoreUnknownValues *bool `pulumi:"ignoreUnknownValues"`
	// The maximum number of bad records that
	// BigQuery can ignore when reading data.
	MaxBadRecords *int `pulumi:"maxBadRecords"`
	// A JSON schema for the external table. Schema is required
	// for CSV and JSON formats if autodetect is not on. Schema is disallowed
	// for Google Cloud Bigtable, Cloud Datastore backups, Avro, ORC and Parquet formats.
	// ~>**NOTE:** Because this field expects a JSON string, any changes to the
	// string will create a diff, even if the JSON itself hasn't changed.
	// Furthermore drift for this field cannot not be detected because BigQuery
	// only uses this schema to compute the effective schema for the table, therefore
	// any changes on the configured value will force the table to be recreated.
	// This schema is effectively only applied when creating a table from an external
	// datasource, after creation the computed schema will be stored in
	// `google_bigquery_table.schema`
	Schema *string `pulumi:"schema"`
	// The data format. Supported values are:
	// "CSV", "GOOGLE_SHEETS", "NEWLINE_DELIMITED_JSON", "AVRO", "PARQUET",
	// and "DATSTORE_BACKUP". To use "GOOGLE_SHEETS"
	// the `scopes` must include
	// "https://www.googleapis.com/auth/drive.readonly".
	SourceFormat string `pulumi:"sourceFormat"`
	// A list of the fully-qualified URIs that point to
	// your data in Google Cloud.
	SourceUris []string `pulumi:"sourceUris"`
}

type TableExternalDataConfigurationArgs

type TableExternalDataConfigurationArgs struct {
	// - Let BigQuery try to autodetect the schema
	//   and format of the table.
	Autodetect pulumi.BoolInput `pulumi:"autodetect"`
	// The compression type of the data source.
	// Valid values are "NONE" or "GZIP".
	Compression pulumi.StringPtrInput `pulumi:"compression"`
	// Additional properties to set if
	// `sourceFormat` is set to "CSV". Structure is documented below.
	CsvOptions TableExternalDataConfigurationCsvOptionsPtrInput `pulumi:"csvOptions"`
	// Additional options if
	// `sourceFormat` is set to "GOOGLE_SHEETS". Structure is
	// documented below.
	GoogleSheetsOptions TableExternalDataConfigurationGoogleSheetsOptionsPtrInput `pulumi:"googleSheetsOptions"`
	// When set, configures hive partitioning
	// support. Not all storage formats support hive partitioning -- requesting hive
	// partitioning on an unsupported format will lead to an error, as will providing
	// an invalid specification.
	HivePartitioningOptions TableExternalDataConfigurationHivePartitioningOptionsPtrInput `pulumi:"hivePartitioningOptions"`
	// Indicates if BigQuery should
	// allow extra values that are not represented in the table schema.
	// If true, the extra values are ignored. If false, records with
	// extra columns are treated as bad records, and if there are too
	// many bad records, an invalid error is returned in the job result.
	// The default value is false.
	IgnoreUnknownValues pulumi.BoolPtrInput `pulumi:"ignoreUnknownValues"`
	// The maximum number of bad records that
	// BigQuery can ignore when reading data.
	MaxBadRecords pulumi.IntPtrInput `pulumi:"maxBadRecords"`
	// A JSON schema for the external table. Schema is required
	// for CSV and JSON formats if autodetect is not on. Schema is disallowed
	// for Google Cloud Bigtable, Cloud Datastore backups, Avro, ORC and Parquet formats.
	// ~>**NOTE:** Because this field expects a JSON string, any changes to the
	// string will create a diff, even if the JSON itself hasn't changed.
	// Furthermore drift for this field cannot not be detected because BigQuery
	// only uses this schema to compute the effective schema for the table, therefore
	// any changes on the configured value will force the table to be recreated.
	// This schema is effectively only applied when creating a table from an external
	// datasource, after creation the computed schema will be stored in
	// `google_bigquery_table.schema`
	Schema pulumi.StringPtrInput `pulumi:"schema"`
	// The data format. Supported values are:
	// "CSV", "GOOGLE_SHEETS", "NEWLINE_DELIMITED_JSON", "AVRO", "PARQUET",
	// and "DATSTORE_BACKUP". To use "GOOGLE_SHEETS"
	// the `scopes` must include
	// "https://www.googleapis.com/auth/drive.readonly".
	SourceFormat pulumi.StringInput `pulumi:"sourceFormat"`
	// A list of the fully-qualified URIs that point to
	// your data in Google Cloud.
	SourceUris pulumi.StringArrayInput `pulumi:"sourceUris"`
}

func (TableExternalDataConfigurationArgs) ElementType

func (TableExternalDataConfigurationArgs) ToTableExternalDataConfigurationOutput

func (i TableExternalDataConfigurationArgs) ToTableExternalDataConfigurationOutput() TableExternalDataConfigurationOutput

func (TableExternalDataConfigurationArgs) ToTableExternalDataConfigurationOutputWithContext

func (i TableExternalDataConfigurationArgs) ToTableExternalDataConfigurationOutputWithContext(ctx context.Context) TableExternalDataConfigurationOutput

func (TableExternalDataConfigurationArgs) ToTableExternalDataConfigurationPtrOutput

func (i TableExternalDataConfigurationArgs) ToTableExternalDataConfigurationPtrOutput() TableExternalDataConfigurationPtrOutput

func (TableExternalDataConfigurationArgs) ToTableExternalDataConfigurationPtrOutputWithContext

func (i TableExternalDataConfigurationArgs) ToTableExternalDataConfigurationPtrOutputWithContext(ctx context.Context) TableExternalDataConfigurationPtrOutput

type TableExternalDataConfigurationCsvOptions

type TableExternalDataConfigurationCsvOptions struct {
	// Indicates if BigQuery should accept rows
	// that are missing trailing optional columns.
	AllowJaggedRows *bool `pulumi:"allowJaggedRows"`
	// Indicates if BigQuery should allow
	// quoted data sections that contain newline characters in a CSV file.
	// The default value is false.
	AllowQuotedNewlines *bool `pulumi:"allowQuotedNewlines"`
	// The character encoding of the data. The supported
	// values are UTF-8 or ISO-8859-1.
	Encoding *string `pulumi:"encoding"`
	// The separator for fields in a CSV file.
	FieldDelimiter *string `pulumi:"fieldDelimiter"`
	// The value that is used to quote data sections in a
	// CSV file. If your data does not contain quoted sections, set the
	// property value to an empty string. If your data contains quoted newline
	// characters, you must also set the `allowQuotedNewlines` property to true.
	// The API-side default is `"`, specified in the provider escaped as `\"`. Due to
	// limitations with default values, this value is required to be
	// explicitly set.
	Quote string `pulumi:"quote"`
	// The number of rows at the top of the sheet
	// that BigQuery will skip when reading the data. At least one of `range` or
	// `skipLeadingRows` must be set.
	SkipLeadingRows *int `pulumi:"skipLeadingRows"`
}

type TableExternalDataConfigurationCsvOptionsArgs

type TableExternalDataConfigurationCsvOptionsArgs struct {
	// Indicates if BigQuery should accept rows
	// that are missing trailing optional columns.
	AllowJaggedRows pulumi.BoolPtrInput `pulumi:"allowJaggedRows"`
	// Indicates if BigQuery should allow
	// quoted data sections that contain newline characters in a CSV file.
	// The default value is false.
	AllowQuotedNewlines pulumi.BoolPtrInput `pulumi:"allowQuotedNewlines"`
	// The character encoding of the data. The supported
	// values are UTF-8 or ISO-8859-1.
	Encoding pulumi.StringPtrInput `pulumi:"encoding"`
	// The separator for fields in a CSV file.
	FieldDelimiter pulumi.StringPtrInput `pulumi:"fieldDelimiter"`
	// The value that is used to quote data sections in a
	// CSV file. If your data does not contain quoted sections, set the
	// property value to an empty string. If your data contains quoted newline
	// characters, you must also set the `allowQuotedNewlines` property to true.
	// The API-side default is `"`, specified in the provider escaped as `\"`. Due to
	// limitations with default values, this value is required to be
	// explicitly set.
	Quote pulumi.StringInput `pulumi:"quote"`
	// The number of rows at the top of the sheet
	// that BigQuery will skip when reading the data. At least one of `range` or
	// `skipLeadingRows` must be set.
	SkipLeadingRows pulumi.IntPtrInput `pulumi:"skipLeadingRows"`
}

func (TableExternalDataConfigurationCsvOptionsArgs) ElementType

func (TableExternalDataConfigurationCsvOptionsArgs) ToTableExternalDataConfigurationCsvOptionsOutput

func (i TableExternalDataConfigurationCsvOptionsArgs) ToTableExternalDataConfigurationCsvOptionsOutput() TableExternalDataConfigurationCsvOptionsOutput

func (TableExternalDataConfigurationCsvOptionsArgs) ToTableExternalDataConfigurationCsvOptionsOutputWithContext

func (i TableExternalDataConfigurationCsvOptionsArgs) ToTableExternalDataConfigurationCsvOptionsOutputWithContext(ctx context.Context) TableExternalDataConfigurationCsvOptionsOutput

func (TableExternalDataConfigurationCsvOptionsArgs) ToTableExternalDataConfigurationCsvOptionsPtrOutput

func (i TableExternalDataConfigurationCsvOptionsArgs) ToTableExternalDataConfigurationCsvOptionsPtrOutput() TableExternalDataConfigurationCsvOptionsPtrOutput

func (TableExternalDataConfigurationCsvOptionsArgs) ToTableExternalDataConfigurationCsvOptionsPtrOutputWithContext

func (i TableExternalDataConfigurationCsvOptionsArgs) ToTableExternalDataConfigurationCsvOptionsPtrOutputWithContext(ctx context.Context) TableExternalDataConfigurationCsvOptionsPtrOutput

type TableExternalDataConfigurationCsvOptionsInput

type TableExternalDataConfigurationCsvOptionsInput interface {
	pulumi.Input

	ToTableExternalDataConfigurationCsvOptionsOutput() TableExternalDataConfigurationCsvOptionsOutput
	ToTableExternalDataConfigurationCsvOptionsOutputWithContext(context.Context) TableExternalDataConfigurationCsvOptionsOutput
}

TableExternalDataConfigurationCsvOptionsInput is an input type that accepts TableExternalDataConfigurationCsvOptionsArgs and TableExternalDataConfigurationCsvOptionsOutput values. You can construct a concrete instance of `TableExternalDataConfigurationCsvOptionsInput` via:

TableExternalDataConfigurationCsvOptionsArgs{...}

type TableExternalDataConfigurationCsvOptionsOutput

type TableExternalDataConfigurationCsvOptionsOutput struct{ *pulumi.OutputState }

func (TableExternalDataConfigurationCsvOptionsOutput) AllowJaggedRows

Indicates if BigQuery should accept rows that are missing trailing optional columns.

func (TableExternalDataConfigurationCsvOptionsOutput) AllowQuotedNewlines

Indicates if BigQuery should allow quoted data sections that contain newline characters in a CSV file. The default value is false.

func (TableExternalDataConfigurationCsvOptionsOutput) ElementType

func (TableExternalDataConfigurationCsvOptionsOutput) Encoding

The character encoding of the data. The supported values are UTF-8 or ISO-8859-1.

func (TableExternalDataConfigurationCsvOptionsOutput) FieldDelimiter

The separator for fields in a CSV file.

func (TableExternalDataConfigurationCsvOptionsOutput) Quote

The value that is used to quote data sections in a CSV file. If your data does not contain quoted sections, set the property value to an empty string. If your data contains quoted newline characters, you must also set the `allowQuotedNewlines` property to true. The API-side default is `"`, specified in the provider escaped as `\"`. Due to limitations with default values, this value is required to be explicitly set.

func (TableExternalDataConfigurationCsvOptionsOutput) SkipLeadingRows

The number of rows at the top of the sheet that BigQuery will skip when reading the data. At least one of `range` or `skipLeadingRows` must be set.

func (TableExternalDataConfigurationCsvOptionsOutput) ToTableExternalDataConfigurationCsvOptionsOutput

func (o TableExternalDataConfigurationCsvOptionsOutput) ToTableExternalDataConfigurationCsvOptionsOutput() TableExternalDataConfigurationCsvOptionsOutput

func (TableExternalDataConfigurationCsvOptionsOutput) ToTableExternalDataConfigurationCsvOptionsOutputWithContext

func (o TableExternalDataConfigurationCsvOptionsOutput) ToTableExternalDataConfigurationCsvOptionsOutputWithContext(ctx context.Context) TableExternalDataConfigurationCsvOptionsOutput

func (TableExternalDataConfigurationCsvOptionsOutput) ToTableExternalDataConfigurationCsvOptionsPtrOutput

func (o TableExternalDataConfigurationCsvOptionsOutput) ToTableExternalDataConfigurationCsvOptionsPtrOutput() TableExternalDataConfigurationCsvOptionsPtrOutput

func (TableExternalDataConfigurationCsvOptionsOutput) ToTableExternalDataConfigurationCsvOptionsPtrOutputWithContext

func (o TableExternalDataConfigurationCsvOptionsOutput) ToTableExternalDataConfigurationCsvOptionsPtrOutputWithContext(ctx context.Context) TableExternalDataConfigurationCsvOptionsPtrOutput

type TableExternalDataConfigurationCsvOptionsPtrInput

type TableExternalDataConfigurationCsvOptionsPtrInput interface {
	pulumi.Input

	ToTableExternalDataConfigurationCsvOptionsPtrOutput() TableExternalDataConfigurationCsvOptionsPtrOutput
	ToTableExternalDataConfigurationCsvOptionsPtrOutputWithContext(context.Context) TableExternalDataConfigurationCsvOptionsPtrOutput
}

TableExternalDataConfigurationCsvOptionsPtrInput is an input type that accepts TableExternalDataConfigurationCsvOptionsArgs, TableExternalDataConfigurationCsvOptionsPtr and TableExternalDataConfigurationCsvOptionsPtrOutput values. You can construct a concrete instance of `TableExternalDataConfigurationCsvOptionsPtrInput` via:

        TableExternalDataConfigurationCsvOptionsArgs{...}

or:

        nil

type TableExternalDataConfigurationCsvOptionsPtrOutput

type TableExternalDataConfigurationCsvOptionsPtrOutput struct{ *pulumi.OutputState }

func (TableExternalDataConfigurationCsvOptionsPtrOutput) AllowJaggedRows

Indicates if BigQuery should accept rows that are missing trailing optional columns.

func (TableExternalDataConfigurationCsvOptionsPtrOutput) AllowQuotedNewlines

Indicates if BigQuery should allow quoted data sections that contain newline characters in a CSV file. The default value is false.

func (TableExternalDataConfigurationCsvOptionsPtrOutput) Elem

func (TableExternalDataConfigurationCsvOptionsPtrOutput) ElementType

func (TableExternalDataConfigurationCsvOptionsPtrOutput) Encoding

The character encoding of the data. The supported values are UTF-8 or ISO-8859-1.

func (TableExternalDataConfigurationCsvOptionsPtrOutput) FieldDelimiter

The separator for fields in a CSV file.

func (TableExternalDataConfigurationCsvOptionsPtrOutput) Quote

The value that is used to quote data sections in a CSV file. If your data does not contain quoted sections, set the property value to an empty string. If your data contains quoted newline characters, you must also set the `allowQuotedNewlines` property to true. The API-side default is `"`, specified in the provider escaped as `\"`. Due to limitations with default values, this value is required to be explicitly set.

func (TableExternalDataConfigurationCsvOptionsPtrOutput) SkipLeadingRows

The number of rows at the top of the sheet that BigQuery will skip when reading the data. At least one of `range` or `skipLeadingRows` must be set.

func (TableExternalDataConfigurationCsvOptionsPtrOutput) ToTableExternalDataConfigurationCsvOptionsPtrOutput

func (o TableExternalDataConfigurationCsvOptionsPtrOutput) ToTableExternalDataConfigurationCsvOptionsPtrOutput() TableExternalDataConfigurationCsvOptionsPtrOutput

func (TableExternalDataConfigurationCsvOptionsPtrOutput) ToTableExternalDataConfigurationCsvOptionsPtrOutputWithContext

func (o TableExternalDataConfigurationCsvOptionsPtrOutput) ToTableExternalDataConfigurationCsvOptionsPtrOutputWithContext(ctx context.Context) TableExternalDataConfigurationCsvOptionsPtrOutput

type TableExternalDataConfigurationGoogleSheetsOptions

type TableExternalDataConfigurationGoogleSheetsOptions struct {
	// Information required to partition based on ranges.
	// Structure is documented below.
	Range *string `pulumi:"range"`
	// The number of rows at the top of the sheet
	// that BigQuery will skip when reading the data. At least one of `range` or
	// `skipLeadingRows` must be set.
	SkipLeadingRows *int `pulumi:"skipLeadingRows"`
}

type TableExternalDataConfigurationGoogleSheetsOptionsArgs

type TableExternalDataConfigurationGoogleSheetsOptionsArgs struct {
	// Information required to partition based on ranges.
	// Structure is documented below.
	Range pulumi.StringPtrInput `pulumi:"range"`
	// The number of rows at the top of the sheet
	// that BigQuery will skip when reading the data. At least one of `range` or
	// `skipLeadingRows` must be set.
	SkipLeadingRows pulumi.IntPtrInput `pulumi:"skipLeadingRows"`
}

func (TableExternalDataConfigurationGoogleSheetsOptionsArgs) ElementType

func (TableExternalDataConfigurationGoogleSheetsOptionsArgs) ToTableExternalDataConfigurationGoogleSheetsOptionsOutput

func (i TableExternalDataConfigurationGoogleSheetsOptionsArgs) ToTableExternalDataConfigurationGoogleSheetsOptionsOutput() TableExternalDataConfigurationGoogleSheetsOptionsOutput

func (TableExternalDataConfigurationGoogleSheetsOptionsArgs) ToTableExternalDataConfigurationGoogleSheetsOptionsOutputWithContext

func (i TableExternalDataConfigurationGoogleSheetsOptionsArgs) ToTableExternalDataConfigurationGoogleSheetsOptionsOutputWithContext(ctx context.Context) TableExternalDataConfigurationGoogleSheetsOptionsOutput

func (TableExternalDataConfigurationGoogleSheetsOptionsArgs) ToTableExternalDataConfigurationGoogleSheetsOptionsPtrOutput

func (i TableExternalDataConfigurationGoogleSheetsOptionsArgs) ToTableExternalDataConfigurationGoogleSheetsOptionsPtrOutput() TableExternalDataConfigurationGoogleSheetsOptionsPtrOutput

func (TableExternalDataConfigurationGoogleSheetsOptionsArgs) ToTableExternalDataConfigurationGoogleSheetsOptionsPtrOutputWithContext

func (i TableExternalDataConfigurationGoogleSheetsOptionsArgs) ToTableExternalDataConfigurationGoogleSheetsOptionsPtrOutputWithContext(ctx context.Context) TableExternalDataConfigurationGoogleSheetsOptionsPtrOutput

type TableExternalDataConfigurationGoogleSheetsOptionsInput

type TableExternalDataConfigurationGoogleSheetsOptionsInput interface {
	pulumi.Input

	ToTableExternalDataConfigurationGoogleSheetsOptionsOutput() TableExternalDataConfigurationGoogleSheetsOptionsOutput
	ToTableExternalDataConfigurationGoogleSheetsOptionsOutputWithContext(context.Context) TableExternalDataConfigurationGoogleSheetsOptionsOutput
}

TableExternalDataConfigurationGoogleSheetsOptionsInput is an input type that accepts TableExternalDataConfigurationGoogleSheetsOptionsArgs and TableExternalDataConfigurationGoogleSheetsOptionsOutput values. You can construct a concrete instance of `TableExternalDataConfigurationGoogleSheetsOptionsInput` via:

TableExternalDataConfigurationGoogleSheetsOptionsArgs{...}

type TableExternalDataConfigurationGoogleSheetsOptionsOutput

type TableExternalDataConfigurationGoogleSheetsOptionsOutput struct{ *pulumi.OutputState }

func (TableExternalDataConfigurationGoogleSheetsOptionsOutput) ElementType

func (TableExternalDataConfigurationGoogleSheetsOptionsOutput) Range

Information required to partition based on ranges. Structure is documented below.

func (TableExternalDataConfigurationGoogleSheetsOptionsOutput) SkipLeadingRows

The number of rows at the top of the sheet that BigQuery will skip when reading the data. At least one of `range` or `skipLeadingRows` must be set.

func (TableExternalDataConfigurationGoogleSheetsOptionsOutput) ToTableExternalDataConfigurationGoogleSheetsOptionsOutput

func (TableExternalDataConfigurationGoogleSheetsOptionsOutput) ToTableExternalDataConfigurationGoogleSheetsOptionsOutputWithContext

func (o TableExternalDataConfigurationGoogleSheetsOptionsOutput) ToTableExternalDataConfigurationGoogleSheetsOptionsOutputWithContext(ctx context.Context) TableExternalDataConfigurationGoogleSheetsOptionsOutput

func (TableExternalDataConfigurationGoogleSheetsOptionsOutput) ToTableExternalDataConfigurationGoogleSheetsOptionsPtrOutput

func (TableExternalDataConfigurationGoogleSheetsOptionsOutput) ToTableExternalDataConfigurationGoogleSheetsOptionsPtrOutputWithContext

func (o TableExternalDataConfigurationGoogleSheetsOptionsOutput) ToTableExternalDataConfigurationGoogleSheetsOptionsPtrOutputWithContext(ctx context.Context) TableExternalDataConfigurationGoogleSheetsOptionsPtrOutput

type TableExternalDataConfigurationGoogleSheetsOptionsPtrInput

type TableExternalDataConfigurationGoogleSheetsOptionsPtrInput interface {
	pulumi.Input

	ToTableExternalDataConfigurationGoogleSheetsOptionsPtrOutput() TableExternalDataConfigurationGoogleSheetsOptionsPtrOutput
	ToTableExternalDataConfigurationGoogleSheetsOptionsPtrOutputWithContext(context.Context) TableExternalDataConfigurationGoogleSheetsOptionsPtrOutput
}

TableExternalDataConfigurationGoogleSheetsOptionsPtrInput is an input type that accepts TableExternalDataConfigurationGoogleSheetsOptionsArgs, TableExternalDataConfigurationGoogleSheetsOptionsPtr and TableExternalDataConfigurationGoogleSheetsOptionsPtrOutput values. You can construct a concrete instance of `TableExternalDataConfigurationGoogleSheetsOptionsPtrInput` via:

        TableExternalDataConfigurationGoogleSheetsOptionsArgs{...}

or:

        nil

type TableExternalDataConfigurationGoogleSheetsOptionsPtrOutput

type TableExternalDataConfigurationGoogleSheetsOptionsPtrOutput struct{ *pulumi.OutputState }

func (TableExternalDataConfigurationGoogleSheetsOptionsPtrOutput) Elem

func (TableExternalDataConfigurationGoogleSheetsOptionsPtrOutput) ElementType

func (TableExternalDataConfigurationGoogleSheetsOptionsPtrOutput) Range

Information required to partition based on ranges. Structure is documented below.

func (TableExternalDataConfigurationGoogleSheetsOptionsPtrOutput) SkipLeadingRows

The number of rows at the top of the sheet that BigQuery will skip when reading the data. At least one of `range` or `skipLeadingRows` must be set.

func (TableExternalDataConfigurationGoogleSheetsOptionsPtrOutput) ToTableExternalDataConfigurationGoogleSheetsOptionsPtrOutput

func (TableExternalDataConfigurationGoogleSheetsOptionsPtrOutput) ToTableExternalDataConfigurationGoogleSheetsOptionsPtrOutputWithContext

func (o TableExternalDataConfigurationGoogleSheetsOptionsPtrOutput) ToTableExternalDataConfigurationGoogleSheetsOptionsPtrOutputWithContext(ctx context.Context) TableExternalDataConfigurationGoogleSheetsOptionsPtrOutput

type TableExternalDataConfigurationHivePartitioningOptions

type TableExternalDataConfigurationHivePartitioningOptions struct {
	// When set, what mode of hive partitioning to use when
	// reading data. The following modes are supported.
	// * AUTO: automatically infer partition key name(s) and type(s).
	// * STRINGS: automatically infer partition key name(s). All types are
	//   Not all storage formats support hive partitioning. Requesting hive
	//   partitioning on an unsupported format will lead to an error.
	//   Currently supported formats are: JSON, CSV, ORC, Avro and Parquet.
	// * CUSTOM: when set to `CUSTOM`, you must encode the partition key schema within the `sourceUriPrefix` by setting `sourceUriPrefix` to `gs://bucket/path_to_table/{key1:TYPE1}/{key2:TYPE2}/{key3:TYPE3}`.
	Mode *string `pulumi:"mode"`
	// When hive partition detection is requested,
	// a common for all source uris must be required. The prefix must end immediately
	// before the partition key encoding begins. For example, consider files following
	// this data layout. `gs://bucket/path_to_table/dt=2019-06-01/country=USA/id=7/file.avro`
	// `gs://bucket/path_to_table/dt=2019-05-31/country=CA/id=3/file.avro` When hive
	// partitioning is requested with either AUTO or STRINGS detection, the common prefix
	// can be either of `gs://bucket/path_to_table` or `gs://bucket/path_to_table/`.
	// Note that when `mode` is set to `CUSTOM`, you must encode the partition key schema within the `sourceUriPrefix` by setting `sourceUriPrefix` to `gs://bucket/path_to_table/{key1:TYPE1}/{key2:TYPE2}/{key3:TYPE3}`.
	SourceUriPrefix *string `pulumi:"sourceUriPrefix"`
}

type TableExternalDataConfigurationHivePartitioningOptionsArgs

type TableExternalDataConfigurationHivePartitioningOptionsArgs struct {
	// When set, what mode of hive partitioning to use when
	// reading data. The following modes are supported.
	// * AUTO: automatically infer partition key name(s) and type(s).
	// * STRINGS: automatically infer partition key name(s). All types are
	//   Not all storage formats support hive partitioning. Requesting hive
	//   partitioning on an unsupported format will lead to an error.
	//   Currently supported formats are: JSON, CSV, ORC, Avro and Parquet.
	// * CUSTOM: when set to `CUSTOM`, you must encode the partition key schema within the `sourceUriPrefix` by setting `sourceUriPrefix` to `gs://bucket/path_to_table/{key1:TYPE1}/{key2:TYPE2}/{key3:TYPE3}`.
	Mode pulumi.StringPtrInput `pulumi:"mode"`
	// When hive partition detection is requested,
	// a common for all source uris must be required. The prefix must end immediately
	// before the partition key encoding begins. For example, consider files following
	// this data layout. `gs://bucket/path_to_table/dt=2019-06-01/country=USA/id=7/file.avro`
	// `gs://bucket/path_to_table/dt=2019-05-31/country=CA/id=3/file.avro` When hive
	// partitioning is requested with either AUTO or STRINGS detection, the common prefix
	// can be either of `gs://bucket/path_to_table` or `gs://bucket/path_to_table/`.
	// Note that when `mode` is set to `CUSTOM`, you must encode the partition key schema within the `sourceUriPrefix` by setting `sourceUriPrefix` to `gs://bucket/path_to_table/{key1:TYPE1}/{key2:TYPE2}/{key3:TYPE3}`.
	SourceUriPrefix pulumi.StringPtrInput `pulumi:"sourceUriPrefix"`
}

func (TableExternalDataConfigurationHivePartitioningOptionsArgs) ElementType

func (TableExternalDataConfigurationHivePartitioningOptionsArgs) ToTableExternalDataConfigurationHivePartitioningOptionsOutput

func (TableExternalDataConfigurationHivePartitioningOptionsArgs) ToTableExternalDataConfigurationHivePartitioningOptionsOutputWithContext

func (i TableExternalDataConfigurationHivePartitioningOptionsArgs) ToTableExternalDataConfigurationHivePartitioningOptionsOutputWithContext(ctx context.Context) TableExternalDataConfigurationHivePartitioningOptionsOutput

func (TableExternalDataConfigurationHivePartitioningOptionsArgs) ToTableExternalDataConfigurationHivePartitioningOptionsPtrOutput

func (TableExternalDataConfigurationHivePartitioningOptionsArgs) ToTableExternalDataConfigurationHivePartitioningOptionsPtrOutputWithContext

func (i TableExternalDataConfigurationHivePartitioningOptionsArgs) ToTableExternalDataConfigurationHivePartitioningOptionsPtrOutputWithContext(ctx context.Context) TableExternalDataConfigurationHivePartitioningOptionsPtrOutput

type TableExternalDataConfigurationHivePartitioningOptionsInput

type TableExternalDataConfigurationHivePartitioningOptionsInput interface {
	pulumi.Input

	ToTableExternalDataConfigurationHivePartitioningOptionsOutput() TableExternalDataConfigurationHivePartitioningOptionsOutput
	ToTableExternalDataConfigurationHivePartitioningOptionsOutputWithContext(context.Context) TableExternalDataConfigurationHivePartitioningOptionsOutput
}

TableExternalDataConfigurationHivePartitioningOptionsInput is an input type that accepts TableExternalDataConfigurationHivePartitioningOptionsArgs and TableExternalDataConfigurationHivePartitioningOptionsOutput values. You can construct a concrete instance of `TableExternalDataConfigurationHivePartitioningOptionsInput` via:

TableExternalDataConfigurationHivePartitioningOptionsArgs{...}

type TableExternalDataConfigurationHivePartitioningOptionsOutput

type TableExternalDataConfigurationHivePartitioningOptionsOutput struct{ *pulumi.OutputState }

func (TableExternalDataConfigurationHivePartitioningOptionsOutput) ElementType

func (TableExternalDataConfigurationHivePartitioningOptionsOutput) Mode

When set, what mode of hive partitioning to use when reading data. The following modes are supported.

  • AUTO: automatically infer partition key name(s) and type(s).
  • STRINGS: automatically infer partition key name(s). All types are Not all storage formats support hive partitioning. Requesting hive partitioning on an unsupported format will lead to an error. Currently supported formats are: JSON, CSV, ORC, Avro and Parquet.
  • CUSTOM: when set to `CUSTOM`, you must encode the partition key schema within the `sourceUriPrefix` by setting `sourceUriPrefix` to `gs://bucket/path_to_table/{key1:TYPE1}/{key2:TYPE2}/{key3:TYPE3}`.

func (TableExternalDataConfigurationHivePartitioningOptionsOutput) SourceUriPrefix

When hive partition detection is requested, a common for all source uris must be required. The prefix must end immediately before the partition key encoding begins. For example, consider files following this data layout. `gs://bucket/path_to_table/dt=2019-06-01/country=USA/id=7/file.avro` `gs://bucket/path_to_table/dt=2019-05-31/country=CA/id=3/file.avro` When hive partitioning is requested with either AUTO or STRINGS detection, the common prefix can be either of `gs://bucket/path_to_table` or `gs://bucket/path_to_table/`. Note that when `mode` is set to `CUSTOM`, you must encode the partition key schema within the `sourceUriPrefix` by setting `sourceUriPrefix` to `gs://bucket/path_to_table/{key1:TYPE1}/{key2:TYPE2}/{key3:TYPE3}`.

func (TableExternalDataConfigurationHivePartitioningOptionsOutput) ToTableExternalDataConfigurationHivePartitioningOptionsOutput

func (TableExternalDataConfigurationHivePartitioningOptionsOutput) ToTableExternalDataConfigurationHivePartitioningOptionsOutputWithContext

func (o TableExternalDataConfigurationHivePartitioningOptionsOutput) ToTableExternalDataConfigurationHivePartitioningOptionsOutputWithContext(ctx context.Context) TableExternalDataConfigurationHivePartitioningOptionsOutput

func (TableExternalDataConfigurationHivePartitioningOptionsOutput) ToTableExternalDataConfigurationHivePartitioningOptionsPtrOutput

func (TableExternalDataConfigurationHivePartitioningOptionsOutput) ToTableExternalDataConfigurationHivePartitioningOptionsPtrOutputWithContext

func (o TableExternalDataConfigurationHivePartitioningOptionsOutput) ToTableExternalDataConfigurationHivePartitioningOptionsPtrOutputWithContext(ctx context.Context) TableExternalDataConfigurationHivePartitioningOptionsPtrOutput

type TableExternalDataConfigurationHivePartitioningOptionsPtrInput

type TableExternalDataConfigurationHivePartitioningOptionsPtrInput interface {
	pulumi.Input

	ToTableExternalDataConfigurationHivePartitioningOptionsPtrOutput() TableExternalDataConfigurationHivePartitioningOptionsPtrOutput
	ToTableExternalDataConfigurationHivePartitioningOptionsPtrOutputWithContext(context.Context) TableExternalDataConfigurationHivePartitioningOptionsPtrOutput
}

TableExternalDataConfigurationHivePartitioningOptionsPtrInput is an input type that accepts TableExternalDataConfigurationHivePartitioningOptionsArgs, TableExternalDataConfigurationHivePartitioningOptionsPtr and TableExternalDataConfigurationHivePartitioningOptionsPtrOutput values. You can construct a concrete instance of `TableExternalDataConfigurationHivePartitioningOptionsPtrInput` via:

        TableExternalDataConfigurationHivePartitioningOptionsArgs{...}

or:

        nil

type TableExternalDataConfigurationHivePartitioningOptionsPtrOutput

type TableExternalDataConfigurationHivePartitioningOptionsPtrOutput struct{ *pulumi.OutputState }

func (TableExternalDataConfigurationHivePartitioningOptionsPtrOutput) Elem

func (TableExternalDataConfigurationHivePartitioningOptionsPtrOutput) ElementType

func (TableExternalDataConfigurationHivePartitioningOptionsPtrOutput) Mode

When set, what mode of hive partitioning to use when reading data. The following modes are supported.

  • AUTO: automatically infer partition key name(s) and type(s).
  • STRINGS: automatically infer partition key name(s). All types are Not all storage formats support hive partitioning. Requesting hive partitioning on an unsupported format will lead to an error. Currently supported formats are: JSON, CSV, ORC, Avro and Parquet.
  • CUSTOM: when set to `CUSTOM`, you must encode the partition key schema within the `sourceUriPrefix` by setting `sourceUriPrefix` to `gs://bucket/path_to_table/{key1:TYPE1}/{key2:TYPE2}/{key3:TYPE3}`.

func (TableExternalDataConfigurationHivePartitioningOptionsPtrOutput) SourceUriPrefix

When hive partition detection is requested, a common for all source uris must be required. The prefix must end immediately before the partition key encoding begins. For example, consider files following this data layout. `gs://bucket/path_to_table/dt=2019-06-01/country=USA/id=7/file.avro` `gs://bucket/path_to_table/dt=2019-05-31/country=CA/id=3/file.avro` When hive partitioning is requested with either AUTO or STRINGS detection, the common prefix can be either of `gs://bucket/path_to_table` or `gs://bucket/path_to_table/`. Note that when `mode` is set to `CUSTOM`, you must encode the partition key schema within the `sourceUriPrefix` by setting `sourceUriPrefix` to `gs://bucket/path_to_table/{key1:TYPE1}/{key2:TYPE2}/{key3:TYPE3}`.

func (TableExternalDataConfigurationHivePartitioningOptionsPtrOutput) ToTableExternalDataConfigurationHivePartitioningOptionsPtrOutput

func (TableExternalDataConfigurationHivePartitioningOptionsPtrOutput) ToTableExternalDataConfigurationHivePartitioningOptionsPtrOutputWithContext

func (o TableExternalDataConfigurationHivePartitioningOptionsPtrOutput) ToTableExternalDataConfigurationHivePartitioningOptionsPtrOutputWithContext(ctx context.Context) TableExternalDataConfigurationHivePartitioningOptionsPtrOutput

type TableExternalDataConfigurationInput

type TableExternalDataConfigurationInput interface {
	pulumi.Input

	ToTableExternalDataConfigurationOutput() TableExternalDataConfigurationOutput
	ToTableExternalDataConfigurationOutputWithContext(context.Context) TableExternalDataConfigurationOutput
}

TableExternalDataConfigurationInput is an input type that accepts TableExternalDataConfigurationArgs and TableExternalDataConfigurationOutput values. You can construct a concrete instance of `TableExternalDataConfigurationInput` via:

TableExternalDataConfigurationArgs{...}

type TableExternalDataConfigurationOutput

type TableExternalDataConfigurationOutput struct{ *pulumi.OutputState }

func (TableExternalDataConfigurationOutput) Autodetect

  • Let BigQuery try to autodetect the schema and format of the table.

func (TableExternalDataConfigurationOutput) Compression

The compression type of the data source. Valid values are "NONE" or "GZIP".

func (TableExternalDataConfigurationOutput) CsvOptions

Additional properties to set if `sourceFormat` is set to "CSV". Structure is documented below.

func (TableExternalDataConfigurationOutput) ElementType

func (TableExternalDataConfigurationOutput) GoogleSheetsOptions

Additional options if `sourceFormat` is set to "GOOGLE_SHEETS". Structure is documented below.

func (TableExternalDataConfigurationOutput) HivePartitioningOptions

When set, configures hive partitioning support. Not all storage formats support hive partitioning -- requesting hive partitioning on an unsupported format will lead to an error, as will providing an invalid specification.

func (TableExternalDataConfigurationOutput) IgnoreUnknownValues

Indicates if BigQuery should allow extra values that are not represented in the table schema. If true, the extra values are ignored. If false, records with extra columns are treated as bad records, and if there are too many bad records, an invalid error is returned in the job result. The default value is false.

func (TableExternalDataConfigurationOutput) MaxBadRecords

The maximum number of bad records that BigQuery can ignore when reading data.

func (TableExternalDataConfigurationOutput) Schema

A JSON schema for the external table. Schema is required for CSV and JSON formats if autodetect is not on. Schema is disallowed for Google Cloud Bigtable, Cloud Datastore backups, Avro, ORC and Parquet formats. ~>**NOTE:** Because this field expects a JSON string, any changes to the string will create a diff, even if the JSON itself hasn't changed. Furthermore drift for this field cannot not be detected because BigQuery only uses this schema to compute the effective schema for the table, therefore any changes on the configured value will force the table to be recreated. This schema is effectively only applied when creating a table from an external datasource, after creation the computed schema will be stored in `google_bigquery_table.schema`

func (TableExternalDataConfigurationOutput) SourceFormat

The data format. Supported values are: "CSV", "GOOGLE_SHEETS", "NEWLINE_DELIMITED_JSON", "AVRO", "PARQUET", and "DATSTORE_BACKUP". To use "GOOGLE_SHEETS" the `scopes` must include "https://www.googleapis.com/auth/drive.readonly".

func (TableExternalDataConfigurationOutput) SourceUris

A list of the fully-qualified URIs that point to your data in Google Cloud.

func (TableExternalDataConfigurationOutput) ToTableExternalDataConfigurationOutput

func (o TableExternalDataConfigurationOutput) ToTableExternalDataConfigurationOutput() TableExternalDataConfigurationOutput

func (TableExternalDataConfigurationOutput) ToTableExternalDataConfigurationOutputWithContext

func (o TableExternalDataConfigurationOutput) ToTableExternalDataConfigurationOutputWithContext(ctx context.Context) TableExternalDataConfigurationOutput

func (TableExternalDataConfigurationOutput) ToTableExternalDataConfigurationPtrOutput

func (o TableExternalDataConfigurationOutput) ToTableExternalDataConfigurationPtrOutput() TableExternalDataConfigurationPtrOutput

func (TableExternalDataConfigurationOutput) ToTableExternalDataConfigurationPtrOutputWithContext

func (o TableExternalDataConfigurationOutput) ToTableExternalDataConfigurationPtrOutputWithContext(ctx context.Context) TableExternalDataConfigurationPtrOutput

type TableExternalDataConfigurationPtrInput

type TableExternalDataConfigurationPtrInput interface {
	pulumi.Input

	ToTableExternalDataConfigurationPtrOutput() TableExternalDataConfigurationPtrOutput
	ToTableExternalDataConfigurationPtrOutputWithContext(context.Context) TableExternalDataConfigurationPtrOutput
}

TableExternalDataConfigurationPtrInput is an input type that accepts TableExternalDataConfigurationArgs, TableExternalDataConfigurationPtr and TableExternalDataConfigurationPtrOutput values. You can construct a concrete instance of `TableExternalDataConfigurationPtrInput` via:

        TableExternalDataConfigurationArgs{...}

or:

        nil

type TableExternalDataConfigurationPtrOutput

type TableExternalDataConfigurationPtrOutput struct{ *pulumi.OutputState }

func (TableExternalDataConfigurationPtrOutput) Autodetect

  • Let BigQuery try to autodetect the schema and format of the table.

func (TableExternalDataConfigurationPtrOutput) Compression

The compression type of the data source. Valid values are "NONE" or "GZIP".

func (TableExternalDataConfigurationPtrOutput) CsvOptions

Additional properties to set if `sourceFormat` is set to "CSV". Structure is documented below.

func (TableExternalDataConfigurationPtrOutput) Elem

func (TableExternalDataConfigurationPtrOutput) ElementType

func (TableExternalDataConfigurationPtrOutput) GoogleSheetsOptions

Additional options if `sourceFormat` is set to "GOOGLE_SHEETS". Structure is documented below.

func (TableExternalDataConfigurationPtrOutput) HivePartitioningOptions

When set, configures hive partitioning support. Not all storage formats support hive partitioning -- requesting hive partitioning on an unsupported format will lead to an error, as will providing an invalid specification.

func (TableExternalDataConfigurationPtrOutput) IgnoreUnknownValues

Indicates if BigQuery should allow extra values that are not represented in the table schema. If true, the extra values are ignored. If false, records with extra columns are treated as bad records, and if there are too many bad records, an invalid error is returned in the job result. The default value is false.

func (TableExternalDataConfigurationPtrOutput) MaxBadRecords

The maximum number of bad records that BigQuery can ignore when reading data.

func (TableExternalDataConfigurationPtrOutput) Schema

A JSON schema for the external table. Schema is required for CSV and JSON formats if autodetect is not on. Schema is disallowed for Google Cloud Bigtable, Cloud Datastore backups, Avro, ORC and Parquet formats. ~>**NOTE:** Because this field expects a JSON string, any changes to the string will create a diff, even if the JSON itself hasn't changed. Furthermore drift for this field cannot not be detected because BigQuery only uses this schema to compute the effective schema for the table, therefore any changes on the configured value will force the table to be recreated. This schema is effectively only applied when creating a table from an external datasource, after creation the computed schema will be stored in `google_bigquery_table.schema`

func (TableExternalDataConfigurationPtrOutput) SourceFormat

The data format. Supported values are: "CSV", "GOOGLE_SHEETS", "NEWLINE_DELIMITED_JSON", "AVRO", "PARQUET", and "DATSTORE_BACKUP". To use "GOOGLE_SHEETS" the `scopes` must include "https://www.googleapis.com/auth/drive.readonly".

func (TableExternalDataConfigurationPtrOutput) SourceUris

A list of the fully-qualified URIs that point to your data in Google Cloud.

func (TableExternalDataConfigurationPtrOutput) ToTableExternalDataConfigurationPtrOutput

func (o TableExternalDataConfigurationPtrOutput) ToTableExternalDataConfigurationPtrOutput() TableExternalDataConfigurationPtrOutput

func (TableExternalDataConfigurationPtrOutput) ToTableExternalDataConfigurationPtrOutputWithContext

func (o TableExternalDataConfigurationPtrOutput) ToTableExternalDataConfigurationPtrOutputWithContext(ctx context.Context) TableExternalDataConfigurationPtrOutput

type TableMaterializedView

type TableMaterializedView struct {
	// Specifies whether to use BigQuery's automatic refresh for this materialized view when the base table is updated.
	// The default value is true.
	EnableRefresh *bool `pulumi:"enableRefresh"`
	// A query whose result is persisted.
	Query string `pulumi:"query"`
	// The maximum frequency at which this materialized view will be refreshed.
	// The default value is 1800000
	RefreshIntervalMs *int `pulumi:"refreshIntervalMs"`
}

type TableMaterializedViewArgs

type TableMaterializedViewArgs struct {
	// Specifies whether to use BigQuery's automatic refresh for this materialized view when the base table is updated.
	// The default value is true.
	EnableRefresh pulumi.BoolPtrInput `pulumi:"enableRefresh"`
	// A query whose result is persisted.
	Query pulumi.StringInput `pulumi:"query"`
	// The maximum frequency at which this materialized view will be refreshed.
	// The default value is 1800000
	RefreshIntervalMs pulumi.IntPtrInput `pulumi:"refreshIntervalMs"`
}

func (TableMaterializedViewArgs) ElementType

func (TableMaterializedViewArgs) ElementType() reflect.Type

func (TableMaterializedViewArgs) ToTableMaterializedViewOutput

func (i TableMaterializedViewArgs) ToTableMaterializedViewOutput() TableMaterializedViewOutput

func (TableMaterializedViewArgs) ToTableMaterializedViewOutputWithContext

func (i TableMaterializedViewArgs) ToTableMaterializedViewOutputWithContext(ctx context.Context) TableMaterializedViewOutput

func (TableMaterializedViewArgs) ToTableMaterializedViewPtrOutput

func (i TableMaterializedViewArgs) ToTableMaterializedViewPtrOutput() TableMaterializedViewPtrOutput

func (TableMaterializedViewArgs) ToTableMaterializedViewPtrOutputWithContext

func (i TableMaterializedViewArgs) ToTableMaterializedViewPtrOutputWithContext(ctx context.Context) TableMaterializedViewPtrOutput

type TableMaterializedViewInput

type TableMaterializedViewInput interface {
	pulumi.Input

	ToTableMaterializedViewOutput() TableMaterializedViewOutput
	ToTableMaterializedViewOutputWithContext(context.Context) TableMaterializedViewOutput
}

TableMaterializedViewInput is an input type that accepts TableMaterializedViewArgs and TableMaterializedViewOutput values. You can construct a concrete instance of `TableMaterializedViewInput` via:

TableMaterializedViewArgs{...}

type TableMaterializedViewOutput

type TableMaterializedViewOutput struct{ *pulumi.OutputState }

func (TableMaterializedViewOutput) ElementType

func (TableMaterializedViewOutput) EnableRefresh

Specifies whether to use BigQuery's automatic refresh for this materialized view when the base table is updated. The default value is true.

func (TableMaterializedViewOutput) Query

A query whose result is persisted.

func (TableMaterializedViewOutput) RefreshIntervalMs

func (o TableMaterializedViewOutput) RefreshIntervalMs() pulumi.IntPtrOutput

The maximum frequency at which this materialized view will be refreshed. The default value is 1800000

func (TableMaterializedViewOutput) ToTableMaterializedViewOutput

func (o TableMaterializedViewOutput) ToTableMaterializedViewOutput() TableMaterializedViewOutput

func (TableMaterializedViewOutput) ToTableMaterializedViewOutputWithContext

func (o TableMaterializedViewOutput) ToTableMaterializedViewOutputWithContext(ctx context.Context) TableMaterializedViewOutput

func (TableMaterializedViewOutput) ToTableMaterializedViewPtrOutput

func (o TableMaterializedViewOutput) ToTableMaterializedViewPtrOutput() TableMaterializedViewPtrOutput

func (TableMaterializedViewOutput) ToTableMaterializedViewPtrOutputWithContext

func (o TableMaterializedViewOutput) ToTableMaterializedViewPtrOutputWithContext(ctx context.Context) TableMaterializedViewPtrOutput

type TableMaterializedViewPtrInput

type TableMaterializedViewPtrInput interface {
	pulumi.Input

	ToTableMaterializedViewPtrOutput() TableMaterializedViewPtrOutput
	ToTableMaterializedViewPtrOutputWithContext(context.Context) TableMaterializedViewPtrOutput
}

TableMaterializedViewPtrInput is an input type that accepts TableMaterializedViewArgs, TableMaterializedViewPtr and TableMaterializedViewPtrOutput values. You can construct a concrete instance of `TableMaterializedViewPtrInput` via:

        TableMaterializedViewArgs{...}

or:

        nil

type TableMaterializedViewPtrOutput

type TableMaterializedViewPtrOutput struct{ *pulumi.OutputState }

func (TableMaterializedViewPtrOutput) Elem

func (TableMaterializedViewPtrOutput) ElementType

func (TableMaterializedViewPtrOutput) EnableRefresh

Specifies whether to use BigQuery's automatic refresh for this materialized view when the base table is updated. The default value is true.

func (TableMaterializedViewPtrOutput) Query

A query whose result is persisted.

func (TableMaterializedViewPtrOutput) RefreshIntervalMs

func (o TableMaterializedViewPtrOutput) RefreshIntervalMs() pulumi.IntPtrOutput

The maximum frequency at which this materialized view will be refreshed. The default value is 1800000

func (TableMaterializedViewPtrOutput) ToTableMaterializedViewPtrOutput

func (o TableMaterializedViewPtrOutput) ToTableMaterializedViewPtrOutput() TableMaterializedViewPtrOutput

func (TableMaterializedViewPtrOutput) ToTableMaterializedViewPtrOutputWithContext

func (o TableMaterializedViewPtrOutput) ToTableMaterializedViewPtrOutputWithContext(ctx context.Context) TableMaterializedViewPtrOutput

type TableRangePartitioning

type TableRangePartitioning struct {
	// The field used to determine how to create a range-based
	// partition.
	Field string `pulumi:"field"`
	// Information required to partition based on ranges.
	// Structure is documented below.
	Range TableRangePartitioningRange `pulumi:"range"`
}

type TableRangePartitioningArgs

type TableRangePartitioningArgs struct {
	// The field used to determine how to create a range-based
	// partition.
	Field pulumi.StringInput `pulumi:"field"`
	// Information required to partition based on ranges.
	// Structure is documented below.
	Range TableRangePartitioningRangeInput `pulumi:"range"`
}

func (TableRangePartitioningArgs) ElementType

func (TableRangePartitioningArgs) ElementType() reflect.Type

func (TableRangePartitioningArgs) ToTableRangePartitioningOutput

func (i TableRangePartitioningArgs) ToTableRangePartitioningOutput() TableRangePartitioningOutput

func (TableRangePartitioningArgs) ToTableRangePartitioningOutputWithContext

func (i TableRangePartitioningArgs) ToTableRangePartitioningOutputWithContext(ctx context.Context) TableRangePartitioningOutput

func (TableRangePartitioningArgs) ToTableRangePartitioningPtrOutput

func (i TableRangePartitioningArgs) ToTableRangePartitioningPtrOutput() TableRangePartitioningPtrOutput

func (TableRangePartitioningArgs) ToTableRangePartitioningPtrOutputWithContext

func (i TableRangePartitioningArgs) ToTableRangePartitioningPtrOutputWithContext(ctx context.Context) TableRangePartitioningPtrOutput

type TableRangePartitioningInput

type TableRangePartitioningInput interface {
	pulumi.Input

	ToTableRangePartitioningOutput() TableRangePartitioningOutput
	ToTableRangePartitioningOutputWithContext(context.Context) TableRangePartitioningOutput
}

TableRangePartitioningInput is an input type that accepts TableRangePartitioningArgs and TableRangePartitioningOutput values. You can construct a concrete instance of `TableRangePartitioningInput` via:

TableRangePartitioningArgs{...}

type TableRangePartitioningOutput

type TableRangePartitioningOutput struct{ *pulumi.OutputState }

func (TableRangePartitioningOutput) ElementType

func (TableRangePartitioningOutput) Field

The field used to determine how to create a range-based partition.

func (TableRangePartitioningOutput) Range

Information required to partition based on ranges. Structure is documented below.

func (TableRangePartitioningOutput) ToTableRangePartitioningOutput

func (o TableRangePartitioningOutput) ToTableRangePartitioningOutput() TableRangePartitioningOutput

func (TableRangePartitioningOutput) ToTableRangePartitioningOutputWithContext

func (o TableRangePartitioningOutput) ToTableRangePartitioningOutputWithContext(ctx context.Context) TableRangePartitioningOutput

func (TableRangePartitioningOutput) ToTableRangePartitioningPtrOutput

func (o TableRangePartitioningOutput) ToTableRangePartitioningPtrOutput() TableRangePartitioningPtrOutput

func (TableRangePartitioningOutput) ToTableRangePartitioningPtrOutputWithContext

func (o TableRangePartitioningOutput) ToTableRangePartitioningPtrOutputWithContext(ctx context.Context) TableRangePartitioningPtrOutput

type TableRangePartitioningPtrInput

type TableRangePartitioningPtrInput interface {
	pulumi.Input

	ToTableRangePartitioningPtrOutput() TableRangePartitioningPtrOutput
	ToTableRangePartitioningPtrOutputWithContext(context.Context) TableRangePartitioningPtrOutput
}

TableRangePartitioningPtrInput is an input type that accepts TableRangePartitioningArgs, TableRangePartitioningPtr and TableRangePartitioningPtrOutput values. You can construct a concrete instance of `TableRangePartitioningPtrInput` via:

        TableRangePartitioningArgs{...}

or:

        nil

type TableRangePartitioningPtrOutput

type TableRangePartitioningPtrOutput struct{ *pulumi.OutputState }

func (TableRangePartitioningPtrOutput) Elem

func (TableRangePartitioningPtrOutput) ElementType

func (TableRangePartitioningPtrOutput) Field

The field used to determine how to create a range-based partition.

func (TableRangePartitioningPtrOutput) Range

Information required to partition based on ranges. Structure is documented below.

func (TableRangePartitioningPtrOutput) ToTableRangePartitioningPtrOutput

func (o TableRangePartitioningPtrOutput) ToTableRangePartitioningPtrOutput() TableRangePartitioningPtrOutput

func (TableRangePartitioningPtrOutput) ToTableRangePartitioningPtrOutputWithContext

func (o TableRangePartitioningPtrOutput) ToTableRangePartitioningPtrOutputWithContext(ctx context.Context) TableRangePartitioningPtrOutput

type TableRangePartitioningRange

type TableRangePartitioningRange struct {
	// End of the range partitioning, exclusive.
	End int `pulumi:"end"`
	// The width of each range within the partition.
	Interval int `pulumi:"interval"`
	// Start of the range partitioning, inclusive.
	Start int `pulumi:"start"`
}

type TableRangePartitioningRangeArgs

type TableRangePartitioningRangeArgs struct {
	// End of the range partitioning, exclusive.
	End pulumi.IntInput `pulumi:"end"`
	// The width of each range within the partition.
	Interval pulumi.IntInput `pulumi:"interval"`
	// Start of the range partitioning, inclusive.
	Start pulumi.IntInput `pulumi:"start"`
}

func (TableRangePartitioningRangeArgs) ElementType

func (TableRangePartitioningRangeArgs) ToTableRangePartitioningRangeOutput

func (i TableRangePartitioningRangeArgs) ToTableRangePartitioningRangeOutput() TableRangePartitioningRangeOutput

func (TableRangePartitioningRangeArgs) ToTableRangePartitioningRangeOutputWithContext

func (i TableRangePartitioningRangeArgs) ToTableRangePartitioningRangeOutputWithContext(ctx context.Context) TableRangePartitioningRangeOutput

func (TableRangePartitioningRangeArgs) ToTableRangePartitioningRangePtrOutput

func (i TableRangePartitioningRangeArgs) ToTableRangePartitioningRangePtrOutput() TableRangePartitioningRangePtrOutput

func (TableRangePartitioningRangeArgs) ToTableRangePartitioningRangePtrOutputWithContext

func (i TableRangePartitioningRangeArgs) ToTableRangePartitioningRangePtrOutputWithContext(ctx context.Context) TableRangePartitioningRangePtrOutput

type TableRangePartitioningRangeInput

type TableRangePartitioningRangeInput interface {
	pulumi.Input

	ToTableRangePartitioningRangeOutput() TableRangePartitioningRangeOutput
	ToTableRangePartitioningRangeOutputWithContext(context.Context) TableRangePartitioningRangeOutput
}

TableRangePartitioningRangeInput is an input type that accepts TableRangePartitioningRangeArgs and TableRangePartitioningRangeOutput values. You can construct a concrete instance of `TableRangePartitioningRangeInput` via:

TableRangePartitioningRangeArgs{...}

type TableRangePartitioningRangeOutput

type TableRangePartitioningRangeOutput struct{ *pulumi.OutputState }

func (TableRangePartitioningRangeOutput) ElementType

func (TableRangePartitioningRangeOutput) End

End of the range partitioning, exclusive.

func (TableRangePartitioningRangeOutput) Interval

The width of each range within the partition.

func (TableRangePartitioningRangeOutput) Start

Start of the range partitioning, inclusive.

func (TableRangePartitioningRangeOutput) ToTableRangePartitioningRangeOutput

func (o TableRangePartitioningRangeOutput) ToTableRangePartitioningRangeOutput() TableRangePartitioningRangeOutput

func (TableRangePartitioningRangeOutput) ToTableRangePartitioningRangeOutputWithContext

func (o TableRangePartitioningRangeOutput) ToTableRangePartitioningRangeOutputWithContext(ctx context.Context) TableRangePartitioningRangeOutput

func (TableRangePartitioningRangeOutput) ToTableRangePartitioningRangePtrOutput

func (o TableRangePartitioningRangeOutput) ToTableRangePartitioningRangePtrOutput() TableRangePartitioningRangePtrOutput

func (TableRangePartitioningRangeOutput) ToTableRangePartitioningRangePtrOutputWithContext

func (o TableRangePartitioningRangeOutput) ToTableRangePartitioningRangePtrOutputWithContext(ctx context.Context) TableRangePartitioningRangePtrOutput

type TableRangePartitioningRangePtrInput

type TableRangePartitioningRangePtrInput interface {
	pulumi.Input

	ToTableRangePartitioningRangePtrOutput() TableRangePartitioningRangePtrOutput
	ToTableRangePartitioningRangePtrOutputWithContext(context.Context) TableRangePartitioningRangePtrOutput
}

TableRangePartitioningRangePtrInput is an input type that accepts TableRangePartitioningRangeArgs, TableRangePartitioningRangePtr and TableRangePartitioningRangePtrOutput values. You can construct a concrete instance of `TableRangePartitioningRangePtrInput` via:

        TableRangePartitioningRangeArgs{...}

or:

        nil

type TableRangePartitioningRangePtrOutput

type TableRangePartitioningRangePtrOutput struct{ *pulumi.OutputState }

func (TableRangePartitioningRangePtrOutput) Elem

func (TableRangePartitioningRangePtrOutput) ElementType

func (TableRangePartitioningRangePtrOutput) End

End of the range partitioning, exclusive.

func (TableRangePartitioningRangePtrOutput) Interval

The width of each range within the partition.

func (TableRangePartitioningRangePtrOutput) Start

Start of the range partitioning, inclusive.

func (TableRangePartitioningRangePtrOutput) ToTableRangePartitioningRangePtrOutput

func (o TableRangePartitioningRangePtrOutput) ToTableRangePartitioningRangePtrOutput() TableRangePartitioningRangePtrOutput

func (TableRangePartitioningRangePtrOutput) ToTableRangePartitioningRangePtrOutputWithContext

func (o TableRangePartitioningRangePtrOutput) ToTableRangePartitioningRangePtrOutputWithContext(ctx context.Context) TableRangePartitioningRangePtrOutput

type TableState

type TableState struct {
	// Specifies column names to use for data clustering.
	// Up to four top-level columns are allowed, and should be specified in
	// descending priority order.
	Clusterings pulumi.StringArrayInput
	// The time when this table was created, in milliseconds since the epoch.
	CreationTime pulumi.IntPtrInput
	// The dataset ID to create the table in.
	// Changing this forces a new resource to be created.
	DatasetId pulumi.StringPtrInput
	// The field description.
	Description pulumi.StringPtrInput
	// Specifies how the table should be encrypted.
	// If left blank, the table will be encrypted with a Google-managed key; that process
	// is transparent to the user.  Structure is documented below.
	EncryptionConfiguration TableEncryptionConfigurationPtrInput
	// A hash of the resource.
	Etag pulumi.StringPtrInput
	// The time when this table expires, in
	// milliseconds since the epoch. If not present, the table will persist
	// indefinitely. Expired tables will be deleted and their storage
	// reclaimed.
	ExpirationTime pulumi.IntPtrInput
	// Describes the data format,
	// location, and other properties of a table stored outside of BigQuery.
	// By defining these properties, the data source can then be queried as
	// if it were a standard BigQuery table. Structure is documented below.
	ExternalDataConfiguration TableExternalDataConfigurationPtrInput
	// A descriptive name for the table.
	FriendlyName pulumi.StringPtrInput
	// A mapping of labels to assign to the resource.
	Labels pulumi.StringMapInput
	// The time when this table was last modified, in milliseconds since the epoch.
	LastModifiedTime pulumi.IntPtrInput
	// The geographic location where the table resides. This value is inherited from the dataset.
	Location pulumi.StringPtrInput
	// If specified, configures this table as a materialized view.
	// Structure is documented below.
	MaterializedView TableMaterializedViewPtrInput
	// The size of this table in bytes, excluding any data in the streaming buffer.
	NumBytes pulumi.IntPtrInput
	// The number of bytes in the table that are considered "long-term storage".
	NumLongTermBytes pulumi.IntPtrInput
	// The number of rows of data in this table, excluding any data in the streaming buffer.
	NumRows pulumi.IntPtrInput
	// The ID of the project in which the resource belongs. If it
	// is not provided, the provider project is used.
	Project pulumi.StringPtrInput
	// If specified, configures range-based
	// partitioning for this table. Structure is documented below.
	RangePartitioning TableRangePartitioningPtrInput
	// A JSON schema for the external table. Schema is required
	// for CSV and JSON formats if autodetect is not on. Schema is disallowed
	// for Google Cloud Bigtable, Cloud Datastore backups, Avro, ORC and Parquet formats.
	// ~>**NOTE:** Because this field expects a JSON string, any changes to the
	// string will create a diff, even if the JSON itself hasn't changed.
	// Furthermore drift for this field cannot not be detected because BigQuery
	// only uses this schema to compute the effective schema for the table, therefore
	// any changes on the configured value will force the table to be recreated.
	// This schema is effectively only applied when creating a table from an external
	// datasource, after creation the computed schema will be stored in
	// `google_bigquery_table.schema`
	Schema pulumi.StringPtrInput
	// The URI of the created resource.
	SelfLink pulumi.StringPtrInput
	// A unique ID for the resource.
	// Changing this forces a new resource to be created.
	TableId pulumi.StringPtrInput
	// If specified, configures time-based
	// partitioning for this table. Structure is documented below.
	TimePartitioning TableTimePartitioningPtrInput
	// The supported types are DAY, HOUR, MONTH, and YEAR,
	// which will generate one partition per day, hour, month, and year, respectively.
	Type pulumi.StringPtrInput
	// If specified, configures this table as a view.
	// Structure is documented below.
	View TableViewPtrInput
}

func (TableState) ElementType

func (TableState) ElementType() reflect.Type

type TableTimePartitioning

type TableTimePartitioning struct {
	// Number of milliseconds for which to keep the
	// storage for a partition.
	ExpirationMs *int `pulumi:"expirationMs"`
	// The field used to determine how to create a range-based
	// partition.
	Field *string `pulumi:"field"`
	// If set to true, queries over this table
	// require a partition filter that can be used for partition elimination to be
	// specified.
	RequirePartitionFilter *bool `pulumi:"requirePartitionFilter"`
	// The supported types are DAY, HOUR, MONTH, and YEAR,
	// which will generate one partition per day, hour, month, and year, respectively.
	Type string `pulumi:"type"`
}

type TableTimePartitioningArgs

type TableTimePartitioningArgs struct {
	// Number of milliseconds for which to keep the
	// storage for a partition.
	ExpirationMs pulumi.IntPtrInput `pulumi:"expirationMs"`
	// The field used to determine how to create a range-based
	// partition.
	Field pulumi.StringPtrInput `pulumi:"field"`
	// If set to true, queries over this table
	// require a partition filter that can be used for partition elimination to be
	// specified.
	RequirePartitionFilter pulumi.BoolPtrInput `pulumi:"requirePartitionFilter"`
	// The supported types are DAY, HOUR, MONTH, and YEAR,
	// which will generate one partition per day, hour, month, and year, respectively.
	Type pulumi.StringInput `pulumi:"type"`
}

func (TableTimePartitioningArgs) ElementType

func (TableTimePartitioningArgs) ElementType() reflect.Type

func (TableTimePartitioningArgs) ToTableTimePartitioningOutput

func (i TableTimePartitioningArgs) ToTableTimePartitioningOutput() TableTimePartitioningOutput

func (TableTimePartitioningArgs) ToTableTimePartitioningOutputWithContext

func (i TableTimePartitioningArgs) ToTableTimePartitioningOutputWithContext(ctx context.Context) TableTimePartitioningOutput

func (TableTimePartitioningArgs) ToTableTimePartitioningPtrOutput

func (i TableTimePartitioningArgs) ToTableTimePartitioningPtrOutput() TableTimePartitioningPtrOutput

func (TableTimePartitioningArgs) ToTableTimePartitioningPtrOutputWithContext

func (i TableTimePartitioningArgs) ToTableTimePartitioningPtrOutputWithContext(ctx context.Context) TableTimePartitioningPtrOutput

type TableTimePartitioningInput

type TableTimePartitioningInput interface {
	pulumi.Input

	ToTableTimePartitioningOutput() TableTimePartitioningOutput
	ToTableTimePartitioningOutputWithContext(context.Context) TableTimePartitioningOutput
}

TableTimePartitioningInput is an input type that accepts TableTimePartitioningArgs and TableTimePartitioningOutput values. You can construct a concrete instance of `TableTimePartitioningInput` via:

TableTimePartitioningArgs{...}

type TableTimePartitioningOutput

type TableTimePartitioningOutput struct{ *pulumi.OutputState }

func (TableTimePartitioningOutput) ElementType

func (TableTimePartitioningOutput) ExpirationMs

Number of milliseconds for which to keep the storage for a partition.

func (TableTimePartitioningOutput) Field

The field used to determine how to create a range-based partition.

func (TableTimePartitioningOutput) RequirePartitionFilter

func (o TableTimePartitioningOutput) RequirePartitionFilter() pulumi.BoolPtrOutput

If set to true, queries over this table require a partition filter that can be used for partition elimination to be specified.

func (TableTimePartitioningOutput) ToTableTimePartitioningOutput

func (o TableTimePartitioningOutput) ToTableTimePartitioningOutput() TableTimePartitioningOutput

func (TableTimePartitioningOutput) ToTableTimePartitioningOutputWithContext

func (o TableTimePartitioningOutput) ToTableTimePartitioningOutputWithContext(ctx context.Context) TableTimePartitioningOutput

func (TableTimePartitioningOutput) ToTableTimePartitioningPtrOutput

func (o TableTimePartitioningOutput) ToTableTimePartitioningPtrOutput() TableTimePartitioningPtrOutput

func (TableTimePartitioningOutput) ToTableTimePartitioningPtrOutputWithContext

func (o TableTimePartitioningOutput) ToTableTimePartitioningPtrOutputWithContext(ctx context.Context) TableTimePartitioningPtrOutput

func (TableTimePartitioningOutput) Type

The supported types are DAY, HOUR, MONTH, and YEAR, which will generate one partition per day, hour, month, and year, respectively.

type TableTimePartitioningPtrInput

type TableTimePartitioningPtrInput interface {
	pulumi.Input

	ToTableTimePartitioningPtrOutput() TableTimePartitioningPtrOutput
	ToTableTimePartitioningPtrOutputWithContext(context.Context) TableTimePartitioningPtrOutput
}

TableTimePartitioningPtrInput is an input type that accepts TableTimePartitioningArgs, TableTimePartitioningPtr and TableTimePartitioningPtrOutput values. You can construct a concrete instance of `TableTimePartitioningPtrInput` via:

        TableTimePartitioningArgs{...}

or:

        nil

type TableTimePartitioningPtrOutput

type TableTimePartitioningPtrOutput struct{ *pulumi.OutputState }

func (TableTimePartitioningPtrOutput) Elem

func (TableTimePartitioningPtrOutput) ElementType

func (TableTimePartitioningPtrOutput) ExpirationMs

Number of milliseconds for which to keep the storage for a partition.

func (TableTimePartitioningPtrOutput) Field

The field used to determine how to create a range-based partition.

func (TableTimePartitioningPtrOutput) RequirePartitionFilter

func (o TableTimePartitioningPtrOutput) RequirePartitionFilter() pulumi.BoolPtrOutput

If set to true, queries over this table require a partition filter that can be used for partition elimination to be specified.

func (TableTimePartitioningPtrOutput) ToTableTimePartitioningPtrOutput

func (o TableTimePartitioningPtrOutput) ToTableTimePartitioningPtrOutput() TableTimePartitioningPtrOutput

func (TableTimePartitioningPtrOutput) ToTableTimePartitioningPtrOutputWithContext

func (o TableTimePartitioningPtrOutput) ToTableTimePartitioningPtrOutputWithContext(ctx context.Context) TableTimePartitioningPtrOutput

func (TableTimePartitioningPtrOutput) Type

The supported types are DAY, HOUR, MONTH, and YEAR, which will generate one partition per day, hour, month, and year, respectively.

type TableView

type TableView struct {
	// A query whose result is persisted.
	Query string `pulumi:"query"`
	// Specifies whether to use BigQuery's legacy SQL for this view.
	// The default value is true. If set to false, the view will use BigQuery's standard SQL.
	UseLegacySql *bool `pulumi:"useLegacySql"`
}

type TableViewArgs

type TableViewArgs struct {
	// A query whose result is persisted.
	Query pulumi.StringInput `pulumi:"query"`
	// Specifies whether to use BigQuery's legacy SQL for this view.
	// The default value is true. If set to false, the view will use BigQuery's standard SQL.
	UseLegacySql pulumi.BoolPtrInput `pulumi:"useLegacySql"`
}

func (TableViewArgs) ElementType

func (TableViewArgs) ElementType() reflect.Type

func (TableViewArgs) ToTableViewOutput

func (i TableViewArgs) ToTableViewOutput() TableViewOutput

func (TableViewArgs) ToTableViewOutputWithContext

func (i TableViewArgs) ToTableViewOutputWithContext(ctx context.Context) TableViewOutput

func (TableViewArgs) ToTableViewPtrOutput

func (i TableViewArgs) ToTableViewPtrOutput() TableViewPtrOutput

func (TableViewArgs) ToTableViewPtrOutputWithContext

func (i TableViewArgs) ToTableViewPtrOutputWithContext(ctx context.Context) TableViewPtrOutput

type TableViewInput

type TableViewInput interface {
	pulumi.Input

	ToTableViewOutput() TableViewOutput
	ToTableViewOutputWithContext(context.Context) TableViewOutput
}

TableViewInput is an input type that accepts TableViewArgs and TableViewOutput values. You can construct a concrete instance of `TableViewInput` via:

TableViewArgs{...}

type TableViewOutput

type TableViewOutput struct{ *pulumi.OutputState }

func (TableViewOutput) ElementType

func (TableViewOutput) ElementType() reflect.Type

func (TableViewOutput) Query

A query whose result is persisted.

func (TableViewOutput) ToTableViewOutput

func (o TableViewOutput) ToTableViewOutput() TableViewOutput

func (TableViewOutput) ToTableViewOutputWithContext

func (o TableViewOutput) ToTableViewOutputWithContext(ctx context.Context) TableViewOutput

func (TableViewOutput) ToTableViewPtrOutput

func (o TableViewOutput) ToTableViewPtrOutput() TableViewPtrOutput

func (TableViewOutput) ToTableViewPtrOutputWithContext

func (o TableViewOutput) ToTableViewPtrOutputWithContext(ctx context.Context) TableViewPtrOutput

func (TableViewOutput) UseLegacySql

func (o TableViewOutput) UseLegacySql() pulumi.BoolPtrOutput

Specifies whether to use BigQuery's legacy SQL for this view. The default value is true. If set to false, the view will use BigQuery's standard SQL.

type TableViewPtrInput

type TableViewPtrInput interface {
	pulumi.Input

	ToTableViewPtrOutput() TableViewPtrOutput
	ToTableViewPtrOutputWithContext(context.Context) TableViewPtrOutput
}

TableViewPtrInput is an input type that accepts TableViewArgs, TableViewPtr and TableViewPtrOutput values. You can construct a concrete instance of `TableViewPtrInput` via:

        TableViewArgs{...}

or:

        nil

func TableViewPtr

func TableViewPtr(v *TableViewArgs) TableViewPtrInput

type TableViewPtrOutput

type TableViewPtrOutput struct{ *pulumi.OutputState }

func (TableViewPtrOutput) Elem

func (TableViewPtrOutput) ElementType

func (TableViewPtrOutput) ElementType() reflect.Type

func (TableViewPtrOutput) Query

A query whose result is persisted.

func (TableViewPtrOutput) ToTableViewPtrOutput

func (o TableViewPtrOutput) ToTableViewPtrOutput() TableViewPtrOutput

func (TableViewPtrOutput) ToTableViewPtrOutputWithContext

func (o TableViewPtrOutput) ToTableViewPtrOutputWithContext(ctx context.Context) TableViewPtrOutput

func (TableViewPtrOutput) UseLegacySql

func (o TableViewPtrOutput) UseLegacySql() pulumi.BoolPtrOutput

Specifies whether to use BigQuery's legacy SQL for this view. The default value is true. If set to false, the view will use BigQuery's standard SQL.

Jump to

Keyboard shortcuts

? : This menu
/ : Search site
f or F : Jump to
y or Y : Canonical URL