Documentation ¶
Index ¶
- type AccessDeniedException
- type AllowedStatistics
- type AnalyticsMode
- type ColumnSelector
- type ColumnStatisticsConfiguration
- type CompressionFormat
- type ConditionExpression
- type ConflictException
- type CsvOptions
- type CsvOutputOptions
- type DataCatalogInputDefinition
- type DataCatalogOutput
- type DatabaseInputDefinition
- type DatabaseOutput
- type DatabaseOutputMode
- type DatabaseTableOutputOptions
- type Dataset
- type DatasetParameter
- type DatetimeOptions
- type EncryptionMode
- type EntityDetectorConfiguration
- type ExcelOptions
- type FilesLimit
- type FilterExpression
- type FormatOptions
- type Input
- type InputFormat
- type InternalServerException
- type Job
- type JobRun
- type JobRunState
- type JobSample
- type JobType
- type JsonOptions
- type LogSubscription
- type Metadata
- type Order
- type OrderedBy
- type Output
- type OutputFormat
- type OutputFormatOptions
- type ParameterType
- type PathOptions
- type ProfileConfiguration
- type Project
- type Recipe
- type RecipeAction
- type RecipeReference
- type RecipeStep
- type RecipeVersionErrorDetail
- type ResourceNotFoundException
- type Rule
- type RulesetItem
- type S3Location
- type S3TableOutputOptions
- type Sample
- type SampleMode
- type SampleType
- type Schedule
- type ServiceQuotaExceededException
- type SessionStatus
- type Source
- type StatisticOverride
- type StatisticsConfiguration
- type Threshold
- type ThresholdType
- type ThresholdUnit
- type ValidationConfiguration
- type ValidationException
- type ValidationMode
- type ViewFrame
Constants ¶
This section is empty.
Variables ¶
This section is empty.
Functions ¶
This section is empty.
Types ¶
type AccessDeniedException ¶
type AccessDeniedException struct { Message *string ErrorCodeOverride *string // contains filtered or unexported fields }
Access to the specified resource was denied.
func (*AccessDeniedException) Error ¶
func (e *AccessDeniedException) Error() string
func (*AccessDeniedException) ErrorCode ¶
func (e *AccessDeniedException) ErrorCode() string
func (*AccessDeniedException) ErrorFault ¶
func (e *AccessDeniedException) ErrorFault() smithy.ErrorFault
func (*AccessDeniedException) ErrorMessage ¶
func (e *AccessDeniedException) ErrorMessage() string
type AllowedStatistics ¶
type AllowedStatistics struct { // One or more column statistics to allow for columns that contain detected // entities. // // This member is required. Statistics []string // contains filtered or unexported fields }
Configuration of statistics that are allowed to be run on columns that contain detected entities. When undefined, no statistics will be computed on columns that contain detected entities.
type AnalyticsMode ¶
type AnalyticsMode string
const ( AnalyticsModeEnable AnalyticsMode = "ENABLE" AnalyticsModeDisable AnalyticsMode = "DISABLE" )
Enum values for AnalyticsMode
func (AnalyticsMode) Values ¶
func (AnalyticsMode) Values() []AnalyticsMode
Values returns all known values for AnalyticsMode. Note that this can be expanded in the future, and so it is only as up to date as the client.
The ordering of this slice is not guaranteed to be stable across updates.
type ColumnSelector ¶
type ColumnSelector struct { // The name of a column from a dataset. Name *string // A regular expression for selecting a column from a dataset. Regex *string // contains filtered or unexported fields }
Selector of a column from a dataset for profile job configuration. One selector includes either a column name or a regular expression.
type ColumnStatisticsConfiguration ¶
type ColumnStatisticsConfiguration struct { // Configuration for evaluations. Statistics can be used to select evaluations and // override parameters of evaluations. // // This member is required. Statistics *StatisticsConfiguration // List of column selectors. Selectors can be used to select columns from the // dataset. When selectors are undefined, configuration will be applied to all // supported columns. Selectors []ColumnSelector // contains filtered or unexported fields }
Configuration for column evaluations for a profile job. ColumnStatisticsConfiguration can be used to select evaluations and override parameters of evaluations for particular columns.
type CompressionFormat ¶
type CompressionFormat string
const ( CompressionFormatGzip CompressionFormat = "GZIP" CompressionFormatLz4 CompressionFormat = "LZ4" CompressionFormatSnappy CompressionFormat = "SNAPPY" CompressionFormatBzip2 CompressionFormat = "BZIP2" CompressionFormatDeflate CompressionFormat = "DEFLATE" CompressionFormatLzo CompressionFormat = "LZO" CompressionFormatBrotli CompressionFormat = "BROTLI" CompressionFormatZstd CompressionFormat = "ZSTD" CompressionFormatZlib CompressionFormat = "ZLIB" )
Enum values for CompressionFormat
func (CompressionFormat) Values ¶
func (CompressionFormat) Values() []CompressionFormat
Values returns all known values for CompressionFormat. Note that this can be expanded in the future, and so it is only as up to date as the client.
The ordering of this slice is not guaranteed to be stable across updates.
type ConditionExpression ¶
type ConditionExpression struct { // A specific condition to apply to a recipe action. For more information, see [Recipe structure] in // the Glue DataBrew Developer Guide. // // [Recipe structure]: https://docs.aws.amazon.com/databrew/latest/dg/recipes.html#recipes.structure // // This member is required. Condition *string // A column to apply this condition to. // // This member is required. TargetColumn *string // A value that the condition must evaluate to for the condition to succeed. Value *string // contains filtered or unexported fields }
Represents an individual condition that evaluates to true or false.
Conditions are used with recipe actions. The action is only performed for column values where the condition evaluates to true.
If a recipe requires more than one condition, then the recipe must specify multiple ConditionExpression elements. Each condition is applied to the rows in a dataset first, before the recipe action is performed.
type ConflictException ¶
type ConflictException struct { Message *string ErrorCodeOverride *string // contains filtered or unexported fields }
Updating or deleting a resource can cause an inconsistent state.
func (*ConflictException) Error ¶
func (e *ConflictException) Error() string
func (*ConflictException) ErrorCode ¶
func (e *ConflictException) ErrorCode() string
func (*ConflictException) ErrorFault ¶
func (e *ConflictException) ErrorFault() smithy.ErrorFault
func (*ConflictException) ErrorMessage ¶
func (e *ConflictException) ErrorMessage() string
type CsvOptions ¶
type CsvOptions struct { // A single character that specifies the delimiter being used in the CSV file. Delimiter *string // A variable that specifies whether the first row in the file is parsed as the // header. If this value is false, column names are auto-generated. HeaderRow *bool // contains filtered or unexported fields }
Represents a set of options that define how DataBrew will read a comma-separated value (CSV) file when creating a dataset from that file.
type CsvOutputOptions ¶
type CsvOutputOptions struct { // A single character that specifies the delimiter used to create CSV job output. Delimiter *string // contains filtered or unexported fields }
Represents a set of options that define how DataBrew will write a comma-separated value (CSV) file.
type DataCatalogInputDefinition ¶
type DataCatalogInputDefinition struct { // The name of a database in the Data Catalog. // // This member is required. DatabaseName *string // The name of a database table in the Data Catalog. This table corresponds to a // DataBrew dataset. // // This member is required. TableName *string // The unique identifier of the Amazon Web Services account that holds the Data // Catalog that stores the data. CatalogId *string // Represents an Amazon location where DataBrew can store intermediate results. TempDirectory *S3Location // contains filtered or unexported fields }
Represents how metadata stored in the Glue Data Catalog is defined in a DataBrew dataset.
type DataCatalogOutput ¶
type DataCatalogOutput struct { // The name of a database in the Data Catalog. // // This member is required. DatabaseName *string // The name of a table in the Data Catalog. // // This member is required. TableName *string // The unique identifier of the Amazon Web Services account that holds the Data // Catalog that stores the data. CatalogId *string // Represents options that specify how and where DataBrew writes the database // output generated by recipe jobs. DatabaseOptions *DatabaseTableOutputOptions // A value that, if true, means that any data in the location specified for output // is overwritten with new output. Not supported with DatabaseOptions. Overwrite bool // Represents options that specify how and where DataBrew writes the Amazon S3 // output generated by recipe jobs. S3Options *S3TableOutputOptions // contains filtered or unexported fields }
Represents options that specify how and where in the Glue Data Catalog DataBrew writes the output generated by recipe jobs.
type DatabaseInputDefinition ¶
type DatabaseInputDefinition struct { // The Glue Connection that stores the connection information for the target // database. // // This member is required. GlueConnectionName *string // The table within the target database. DatabaseTableName *string // Custom SQL to run against the provided Glue connection. This SQL will be used // as the input for DataBrew projects and jobs. QueryString *string // Represents an Amazon S3 location (bucket name, bucket owner, and object key) // where DataBrew can read input data, or write output from a job. TempDirectory *S3Location // contains filtered or unexported fields }
Connection information for dataset input files stored in a database.
type DatabaseOutput ¶
type DatabaseOutput struct { // Represents options that specify how and where DataBrew writes the database // output generated by recipe jobs. // // This member is required. DatabaseOptions *DatabaseTableOutputOptions // The Glue connection that stores the connection information for the target // database. // // This member is required. GlueConnectionName *string // The output mode to write into the database. Currently supported option: // NEW_TABLE. DatabaseOutputMode DatabaseOutputMode // contains filtered or unexported fields }
Represents a JDBC database output object which defines the output destination for a DataBrew recipe job to write into.
type DatabaseOutputMode ¶
type DatabaseOutputMode string
const (
DatabaseOutputModeNewTable DatabaseOutputMode = "NEW_TABLE"
)
Enum values for DatabaseOutputMode
func (DatabaseOutputMode) Values ¶
func (DatabaseOutputMode) Values() []DatabaseOutputMode
Values returns all known values for DatabaseOutputMode. Note that this can be expanded in the future, and so it is only as up to date as the client.
The ordering of this slice is not guaranteed to be stable across updates.
type DatabaseTableOutputOptions ¶
type DatabaseTableOutputOptions struct { // A prefix for the name of a table DataBrew will create in the database. // // This member is required. TableName *string // Represents an Amazon S3 location (bucket name and object key) where DataBrew // can store intermediate results. TempDirectory *S3Location // contains filtered or unexported fields }
Represents options that specify how and where DataBrew writes the database output generated by recipe jobs.
type Dataset ¶
type Dataset struct { // Information on how DataBrew can find the dataset, in either the Glue Data // Catalog or Amazon S3. // // This member is required. Input *Input // The unique name of the dataset. // // This member is required. Name *string // The ID of the Amazon Web Services account that owns the dataset. AccountId *string // The date and time that the dataset was created. CreateDate *time.Time // The Amazon Resource Name (ARN) of the user who created the dataset. CreatedBy *string // The file format of a dataset that is created from an Amazon S3 file or folder. Format InputFormat // A set of options that define how DataBrew interprets the data in the dataset. FormatOptions *FormatOptions // The Amazon Resource Name (ARN) of the user who last modified the dataset. LastModifiedBy *string // The last modification date and time of the dataset. LastModifiedDate *time.Time // A set of options that defines how DataBrew interprets an Amazon S3 path of the // dataset. PathOptions *PathOptions // The unique Amazon Resource Name (ARN) for the dataset. ResourceArn *string // The location of the data for the dataset, either Amazon S3 or the Glue Data // Catalog. Source Source // Metadata tags that have been applied to the dataset. Tags map[string]string // contains filtered or unexported fields }
Represents a dataset that can be processed by DataBrew.
type DatasetParameter ¶
type DatasetParameter struct { // The name of the parameter that is used in the dataset's Amazon S3 path. // // This member is required. Name *string // The type of the dataset parameter, can be one of a 'String', 'Number' or // 'Datetime'. // // This member is required. Type ParameterType // Optional boolean value that defines whether the captured value of this // parameter should be used to create a new column in a dataset. CreateColumn bool // Additional parameter options such as a format and a timezone. Required for // datetime parameters. DatetimeOptions *DatetimeOptions // The optional filter expression structure to apply additional matching criteria // to the parameter. Filter *FilterExpression // contains filtered or unexported fields }
Represents a dataset parameter that defines type and conditions for a parameter in the Amazon S3 path of the dataset.
type DatetimeOptions ¶
type DatetimeOptions struct { // Required option, that defines the datetime format used for a date parameter in // the Amazon S3 path. Should use only supported datetime specifiers and separation // characters, all literal a-z or A-Z characters should be escaped with single // quotes. E.g. "MM.dd.yyyy-'at'-HH:mm". // // This member is required. Format *string // Optional value for a non-US locale code, needed for correct interpretation of // some date formats. LocaleCode *string // Optional value for a timezone offset of the datetime parameter value in the // Amazon S3 path. Shouldn't be used if Format for this parameter includes timezone // fields. If no offset specified, UTC is assumed. TimezoneOffset *string // contains filtered or unexported fields }
Represents additional options for correct interpretation of datetime parameters used in the Amazon S3 path of a dataset.
type EncryptionMode ¶
type EncryptionMode string
const ( EncryptionModeSsekms EncryptionMode = "SSE-KMS" EncryptionModeSses3 EncryptionMode = "SSE-S3" )
Enum values for EncryptionMode
func (EncryptionMode) Values ¶
func (EncryptionMode) Values() []EncryptionMode
Values returns all known values for EncryptionMode. Note that this can be expanded in the future, and so it is only as up to date as the client.
The ordering of this slice is not guaranteed to be stable across updates.
type EntityDetectorConfiguration ¶
type EntityDetectorConfiguration struct { // Entity types to detect. Can be any of the following: // // - USA_SSN // // - EMAIL // // - USA_ITIN // // - USA_PASSPORT_NUMBER // // - PHONE_NUMBER // // - USA_DRIVING_LICENSE // // - BANK_ACCOUNT // // - CREDIT_CARD // // - IP_ADDRESS // // - MAC_ADDRESS // // - USA_DEA_NUMBER // // - USA_HCPCS_CODE // // - USA_NATIONAL_PROVIDER_IDENTIFIER // // - USA_NATIONAL_DRUG_CODE // // - USA_HEALTH_INSURANCE_CLAIM_NUMBER // // - USA_MEDICARE_BENEFICIARY_IDENTIFIER // // - USA_CPT_CODE // // - PERSON_NAME // // - DATE // // The Entity type group USA_ALL is also supported, and includes all of the above // entity types except PERSON_NAME and DATE. // // This member is required. EntityTypes []string // Configuration of statistics that are allowed to be run on columns that contain // detected entities. When undefined, no statistics will be computed on columns // that contain detected entities. AllowedStatistics []AllowedStatistics // contains filtered or unexported fields }
Configuration of entity detection for a profile job. When undefined, entity detection is disabled.
type ExcelOptions ¶
type ExcelOptions struct { // A variable that specifies whether the first row in the file is parsed as the // header. If this value is false, column names are auto-generated. HeaderRow *bool // One or more sheet numbers in the Excel file that will be included in the // dataset. SheetIndexes []int32 // One or more named sheets in the Excel file that will be included in the dataset. SheetNames []string // contains filtered or unexported fields }
Represents a set of options that define how DataBrew will interpret a Microsoft Excel file when creating a dataset from that file.
type FilesLimit ¶
type FilesLimit struct { // The number of Amazon S3 files to select. // // This member is required. MaxFiles *int32 // A criteria to use for Amazon S3 files sorting before their selection. By // default uses DESCENDING order, i.e. most recent files are selected first. // Another possible value is ASCENDING. Order Order // A criteria to use for Amazon S3 files sorting before their selection. By // default uses LAST_MODIFIED_DATE as a sorting criteria. Currently it's the only // allowed value. OrderedBy OrderedBy // contains filtered or unexported fields }
Represents a limit imposed on number of Amazon S3 files that should be selected for a dataset from a connected Amazon S3 path.
type FilterExpression ¶
type FilterExpression struct { // The expression which includes condition names followed by substitution // variables, possibly grouped and combined with other conditions. For example, // "(starts_with :prefix1 or starts_with :prefix2) and (ends_with :suffix1 or // ends_with :suffix2)". Substitution variables should start with ':' symbol. // // This member is required. Expression *string // The map of substitution variable names to their values used in this filter // expression. // // This member is required. ValuesMap map[string]string // contains filtered or unexported fields }
Represents a structure for defining parameter conditions. Supported conditions are described here: Supported conditions for dynamic datasetsin the Glue DataBrew Developer Guide.
type FormatOptions ¶
type FormatOptions struct { // Options that define how CSV input is to be interpreted by DataBrew. Csv *CsvOptions // Options that define how Excel input is to be interpreted by DataBrew. Excel *ExcelOptions // Options that define how JSON input is to be interpreted by DataBrew. Json *JsonOptions // contains filtered or unexported fields }
Represents a set of options that define the structure of either comma-separated value (CSV), Excel, or JSON input.
type Input ¶
type Input struct { // The Glue Data Catalog parameters for the data. DataCatalogInputDefinition *DataCatalogInputDefinition // Connection information for dataset input files stored in a database. DatabaseInputDefinition *DatabaseInputDefinition // Contains additional resource information needed for specific datasets. Metadata *Metadata // The Amazon S3 location where the data is stored. S3InputDefinition *S3Location // contains filtered or unexported fields }
Represents information on how DataBrew can find data, in either the Glue Data Catalog or Amazon S3.
type InputFormat ¶
type InputFormat string
const ( InputFormatCsv InputFormat = "CSV" InputFormatJson InputFormat = "JSON" InputFormatParquet InputFormat = "PARQUET" InputFormatExcel InputFormat = "EXCEL" InputFormatOrc InputFormat = "ORC" )
Enum values for InputFormat
func (InputFormat) Values ¶
func (InputFormat) Values() []InputFormat
Values returns all known values for InputFormat. Note that this can be expanded in the future, and so it is only as up to date as the client.
The ordering of this slice is not guaranteed to be stable across updates.
type InternalServerException ¶
type InternalServerException struct { Message *string ErrorCodeOverride *string // contains filtered or unexported fields }
An internal service failure occurred.
func (*InternalServerException) Error ¶
func (e *InternalServerException) Error() string
func (*InternalServerException) ErrorCode ¶
func (e *InternalServerException) ErrorCode() string
func (*InternalServerException) ErrorFault ¶
func (e *InternalServerException) ErrorFault() smithy.ErrorFault
func (*InternalServerException) ErrorMessage ¶
func (e *InternalServerException) ErrorMessage() string
type Job ¶
type Job struct { // The unique name of the job. // // This member is required. Name *string // The ID of the Amazon Web Services account that owns the job. AccountId *string // The date and time that the job was created. CreateDate *time.Time // The Amazon Resource Name (ARN) of the user who created the job. CreatedBy *string // One or more artifacts that represent the Glue Data Catalog output from running // the job. DataCatalogOutputs []DataCatalogOutput // Represents a list of JDBC database output objects which defines the output // destination for a DataBrew recipe job to write into. DatabaseOutputs []DatabaseOutput // A dataset that the job is to process. DatasetName *string // The Amazon Resource Name (ARN) of an encryption key that is used to protect the // job output. For more information, see [Encrypting data written by DataBrew jobs] // // [Encrypting data written by DataBrew jobs]: https://docs.aws.amazon.com/databrew/latest/dg/encryption-security-configuration.html EncryptionKeyArn *string // The encryption mode for the job, which can be one of the following: // // - SSE-KMS - Server-side encryption with keys managed by KMS. // // - SSE-S3 - Server-side encryption with keys managed by Amazon S3. EncryptionMode EncryptionMode // A sample configuration for profile jobs only, which determines the number of // rows on which the profile job is run. If a JobSample value isn't provided, the // default value is used. The default value is CUSTOM_ROWS for the mode parameter // and 20,000 for the size parameter. JobSample *JobSample // The Amazon Resource Name (ARN) of the user who last modified the job. LastModifiedBy *string // The modification date and time of the job. LastModifiedDate *time.Time // The current status of Amazon CloudWatch logging for the job. LogSubscription LogSubscription // The maximum number of nodes that can be consumed when the job processes data. MaxCapacity int32 // The maximum number of times to retry the job after a job run fails. MaxRetries int32 // One or more artifacts that represent output from running the job. Outputs []Output // The name of the project that the job is associated with. ProjectName *string // A set of steps that the job runs. RecipeReference *RecipeReference // The unique Amazon Resource Name (ARN) for the job. ResourceArn *string // The Amazon Resource Name (ARN) of the role to be assumed for this job. RoleArn *string // Metadata tags that have been applied to the job. Tags map[string]string // The job's timeout in minutes. A job that attempts to run longer than this // timeout period ends with a status of TIMEOUT . Timeout int32 // The job type of the job, which must be one of the following: // // - PROFILE - A job to analyze a dataset, to determine its size, data types, // data distribution, and more. // // - RECIPE - A job to apply one or more transformations to a dataset. Type JobType // List of validation configurations that are applied to the profile job. ValidationConfigurations []ValidationConfiguration // contains filtered or unexported fields }
Represents all of the attributes of a DataBrew job.
type JobRun ¶
type JobRun struct { // The number of times that DataBrew has attempted to run the job. Attempt int32 // The date and time when the job completed processing. CompletedOn *time.Time // One or more artifacts that represent the Glue Data Catalog output from running // the job. DataCatalogOutputs []DataCatalogOutput // Represents a list of JDBC database output objects which defines the output // destination for a DataBrew recipe job to write into. DatabaseOutputs []DatabaseOutput // The name of the dataset for the job to process. DatasetName *string // A message indicating an error (if any) that was encountered when the job ran. ErrorMessage *string // The amount of time, in seconds, during which a job run consumed resources. ExecutionTime int32 // The name of the job being processed during this run. JobName *string // A sample configuration for profile jobs only, which determines the number of // rows on which the profile job is run. If a JobSample value isn't provided, the // default is used. The default value is CUSTOM_ROWS for the mode parameter and // 20,000 for the size parameter. JobSample *JobSample // The name of an Amazon CloudWatch log group, where the job writes diagnostic // messages when it runs. LogGroupName *string // The current status of Amazon CloudWatch logging for the job run. LogSubscription LogSubscription // One or more output artifacts from a job run. Outputs []Output // The set of steps processed by the job. RecipeReference *RecipeReference // The unique identifier of the job run. RunId *string // The Amazon Resource Name (ARN) of the user who initiated the job run. StartedBy *string // The date and time when the job run began. StartedOn *time.Time // The current state of the job run entity itself. State JobRunState // List of validation configurations that are applied to the profile job run. ValidationConfigurations []ValidationConfiguration // contains filtered or unexported fields }
Represents one run of a DataBrew job.
type JobRunState ¶
type JobRunState string
const ( JobRunStateStarting JobRunState = "STARTING" JobRunStateRunning JobRunState = "RUNNING" JobRunStateStopping JobRunState = "STOPPING" JobRunStateStopped JobRunState = "STOPPED" JobRunStateSucceeded JobRunState = "SUCCEEDED" JobRunStateFailed JobRunState = "FAILED" JobRunStateTimeout JobRunState = "TIMEOUT" )
Enum values for JobRunState
func (JobRunState) Values ¶
func (JobRunState) Values() []JobRunState
Values returns all known values for JobRunState. Note that this can be expanded in the future, and so it is only as up to date as the client.
The ordering of this slice is not guaranteed to be stable across updates.
type JobSample ¶
type JobSample struct { // A value that determines whether the profile job is run on the entire dataset or // a specified number of rows. This value must be one of the following: // // - FULL_DATASET - The profile job is run on the entire dataset. // // - CUSTOM_ROWS - The profile job is run on the number of rows specified in the // Size parameter. Mode SampleMode // The Size parameter is only required when the mode is CUSTOM_ROWS. The profile // job is run on the specified number of rows. The maximum value for size is // Long.MAX_VALUE. // // Long.MAX_VALUE = 9223372036854775807 Size *int64 // contains filtered or unexported fields }
A sample configuration for profile jobs only, which determines the number of rows on which the profile job is run. If a JobSample value isn't provided, the default is used. The default value is CUSTOM_ROWS for the mode parameter and 20,000 for the size parameter.
type JsonOptions ¶
type JsonOptions struct { // A value that specifies whether JSON input contains embedded new line characters. MultiLine bool // contains filtered or unexported fields }
Represents the JSON-specific options that define how input is to be interpreted by Glue DataBrew.
type LogSubscription ¶
type LogSubscription string
const ( LogSubscriptionEnable LogSubscription = "ENABLE" LogSubscriptionDisable LogSubscription = "DISABLE" )
Enum values for LogSubscription
func (LogSubscription) Values ¶
func (LogSubscription) Values() []LogSubscription
Values returns all known values for LogSubscription. Note that this can be expanded in the future, and so it is only as up to date as the client.
The ordering of this slice is not guaranteed to be stable across updates.
type Metadata ¶
type Metadata struct { // The Amazon Resource Name (ARN) associated with the dataset. Currently, DataBrew // only supports ARNs from Amazon AppFlow. SourceArn *string // contains filtered or unexported fields }
Contains additional resource information needed for specific datasets.
type OrderedBy ¶
type OrderedBy string
const (
OrderedByLastModifiedDate OrderedBy = "LAST_MODIFIED_DATE"
)
Enum values for OrderedBy
type Output ¶
type Output struct { // The location in Amazon S3 where the job writes its output. // // This member is required. Location *S3Location // The compression algorithm used to compress the output text of the job. CompressionFormat CompressionFormat // The data format of the output of the job. Format OutputFormat // Represents options that define how DataBrew formats job output files. FormatOptions *OutputFormatOptions // Maximum number of files to be generated by the job and written to the output // folder. For output partitioned by column(s), the MaxOutputFiles value is the // maximum number of files per partition. MaxOutputFiles *int32 // A value that, if true, means that any data in the location specified for output // is overwritten with new output. Overwrite bool // The names of one or more partition columns for the output of the job. PartitionColumns []string // contains filtered or unexported fields }
Represents options that specify how and where in Amazon S3 DataBrew writes the output generated by recipe jobs or profile jobs.
type OutputFormat ¶
type OutputFormat string
const ( OutputFormatCsv OutputFormat = "CSV" OutputFormatJson OutputFormat = "JSON" OutputFormatParquet OutputFormat = "PARQUET" OutputFormatGlueparquet OutputFormat = "GLUEPARQUET" OutputFormatAvro OutputFormat = "AVRO" OutputFormatOrc OutputFormat = "ORC" OutputFormatXml OutputFormat = "XML" OutputFormatTableauhyper OutputFormat = "TABLEAUHYPER" )
Enum values for OutputFormat
func (OutputFormat) Values ¶
func (OutputFormat) Values() []OutputFormat
Values returns all known values for OutputFormat. Note that this can be expanded in the future, and so it is only as up to date as the client.
The ordering of this slice is not guaranteed to be stable across updates.
type OutputFormatOptions ¶
type OutputFormatOptions struct { // Represents a set of options that define the structure of comma-separated value // (CSV) job output. Csv *CsvOutputOptions // contains filtered or unexported fields }
Represents a set of options that define the structure of comma-separated (CSV) job output.
type ParameterType ¶
type ParameterType string
const ( ParameterTypeDatetime ParameterType = "Datetime" ParameterTypeNumber ParameterType = "Number" ParameterTypeString ParameterType = "String" )
Enum values for ParameterType
func (ParameterType) Values ¶
func (ParameterType) Values() []ParameterType
Values returns all known values for ParameterType. Note that this can be expanded in the future, and so it is only as up to date as the client.
The ordering of this slice is not guaranteed to be stable across updates.
type PathOptions ¶
type PathOptions struct { // If provided, this structure imposes a limit on a number of files that should be // selected. FilesLimit *FilesLimit // If provided, this structure defines a date range for matching Amazon S3 objects // based on their LastModifiedDate attribute in Amazon S3. LastModifiedDateCondition *FilterExpression // A structure that maps names of parameters used in the Amazon S3 path of a // dataset to their definitions. Parameters map[string]DatasetParameter // contains filtered or unexported fields }
Represents a set of options that define how DataBrew selects files for a given Amazon S3 path in a dataset.
type ProfileConfiguration ¶
type ProfileConfiguration struct { // List of configurations for column evaluations. ColumnStatisticsConfigurations // are used to select evaluations and override parameters of evaluations for // particular columns. When ColumnStatisticsConfigurations is undefined, the // profile job will profile all supported columns and run all supported // evaluations. ColumnStatisticsConfigurations []ColumnStatisticsConfiguration // Configuration for inter-column evaluations. Configuration can be used to select // evaluations and override parameters of evaluations. When configuration is // undefined, the profile job will run all supported inter-column evaluations. DatasetStatisticsConfiguration *StatisticsConfiguration // Configuration of entity detection for a profile job. When undefined, entity // detection is disabled. EntityDetectorConfiguration *EntityDetectorConfiguration // List of column selectors. ProfileColumns can be used to select columns from the // dataset. When ProfileColumns is undefined, the profile job will profile all // supported columns. ProfileColumns []ColumnSelector // contains filtered or unexported fields }
Configuration for profile jobs. Configuration can be used to select columns, do evaluations, and override default parameters of evaluations. When configuration is undefined, the profile job will apply default settings to all supported columns.
type Project ¶
type Project struct { // The unique name of a project. // // This member is required. Name *string // The name of a recipe that will be developed during a project session. // // This member is required. RecipeName *string // The ID of the Amazon Web Services account that owns the project. AccountId *string // The date and time that the project was created. CreateDate *time.Time // The Amazon Resource Name (ARN) of the user who crated the project. CreatedBy *string // The dataset that the project is to act upon. DatasetName *string // The Amazon Resource Name (ARN) of the user who last modified the project. LastModifiedBy *string // The last modification date and time for the project. LastModifiedDate *time.Time // The date and time when the project was opened. OpenDate *time.Time // The Amazon Resource Name (ARN) of the user that opened the project for use. OpenedBy *string // The Amazon Resource Name (ARN) for the project. ResourceArn *string // The Amazon Resource Name (ARN) of the role that will be assumed for this // project. RoleArn *string // The sample size and sampling type to apply to the data. If this parameter isn't // specified, then the sample consists of the first 500 rows from the dataset. Sample *Sample // Metadata tags that have been applied to the project. Tags map[string]string // contains filtered or unexported fields }
Represents all of the attributes of a DataBrew project.
type Recipe ¶
type Recipe struct { // The unique name for the recipe. // // This member is required. Name *string // The date and time that the recipe was created. CreateDate *time.Time // The Amazon Resource Name (ARN) of the user who created the recipe. CreatedBy *string // The description of the recipe. Description *string // The Amazon Resource Name (ARN) of the user who last modified the recipe. LastModifiedBy *string // The last modification date and time of the recipe. LastModifiedDate *time.Time // The name of the project that the recipe is associated with. ProjectName *string // The Amazon Resource Name (ARN) of the user who published the recipe. PublishedBy *string // The date and time when the recipe was published. PublishedDate *time.Time // The identifier for the version for the recipe. Must be one of the following: // // - Numeric version ( X.Y ) - X and Y stand for major and minor version numbers. // The maximum length of each is 6 digits, and neither can be negative values. Both // X and Y are required, and "0.0" isn't a valid version. // // - LATEST_WORKING - the most recent valid version being developed in a DataBrew // project. // // - LATEST_PUBLISHED - the most recent published version. RecipeVersion *string // The Amazon Resource Name (ARN) for the recipe. ResourceArn *string // A list of steps that are defined by the recipe. Steps []RecipeStep // Metadata tags that have been applied to the recipe. Tags map[string]string // contains filtered or unexported fields }
Represents one or more actions to be performed on a DataBrew dataset.
type RecipeAction ¶
type RecipeAction struct { // The name of a valid DataBrew transformation to be performed on the data. // // This member is required. Operation *string // Contextual parameters for the transformation. Parameters map[string]string // contains filtered or unexported fields }
Represents a transformation and associated parameters that are used to apply a change to a DataBrew dataset. For more information, see Recipe actions reference.
type RecipeReference ¶
type RecipeReference struct { // The name of the recipe. // // This member is required. Name *string // The identifier for the version for the recipe. RecipeVersion *string // contains filtered or unexported fields }
Represents the name and version of a DataBrew recipe.
type RecipeStep ¶
type RecipeStep struct { // The particular action to be performed in the recipe step. // // This member is required. Action *RecipeAction // One or more conditions that must be met for the recipe step to succeed. // // All of the conditions in the array must be met. In other words, all of the // conditions must be combined using a logical AND operation. ConditionExpressions []ConditionExpression // contains filtered or unexported fields }
Represents a single step from a DataBrew recipe to be performed.
type RecipeVersionErrorDetail ¶
type RecipeVersionErrorDetail struct { // The HTTP status code for the error. ErrorCode *string // The text of the error message. ErrorMessage *string // The identifier for the recipe version associated with this error. RecipeVersion *string // contains filtered or unexported fields }
Represents any errors encountered when attempting to delete multiple recipe versions.
type ResourceNotFoundException ¶
type ResourceNotFoundException struct { Message *string ErrorCodeOverride *string // contains filtered or unexported fields }
One or more resources can't be found.
func (*ResourceNotFoundException) Error ¶
func (e *ResourceNotFoundException) Error() string
func (*ResourceNotFoundException) ErrorCode ¶
func (e *ResourceNotFoundException) ErrorCode() string
func (*ResourceNotFoundException) ErrorFault ¶
func (e *ResourceNotFoundException) ErrorFault() smithy.ErrorFault
func (*ResourceNotFoundException) ErrorMessage ¶
func (e *ResourceNotFoundException) ErrorMessage() string
type Rule ¶
type Rule struct { // The expression which includes column references, condition names followed by // variable references, possibly grouped and combined with other conditions. For // example, (:col1 starts_with :prefix1 or :col1 starts_with :prefix2) and (:col1 // ends_with :suffix1 or :col1 ends_with :suffix2) . Column and value references // are substitution variables that should start with the ':' symbol. Depending on // the context, substitution variables' values can be either an actual value or a // column name. These values are defined in the SubstitutionMap. If a // CheckExpression starts with a column reference, then ColumnSelectors in the rule // should be null. If ColumnSelectors has been defined, then there should be no // column reference in the left side of a condition, for example, is_between :val1 // and :val2 . // // For more information, see [Available checks] // // [Available checks]: https://docs.aws.amazon.com/databrew/latest/dg/profile.data-quality-available-checks.html // // This member is required. CheckExpression *string // The name of the rule. // // This member is required. Name *string // List of column selectors. Selectors can be used to select columns using a name // or regular expression from the dataset. Rule will be applied to selected // columns. ColumnSelectors []ColumnSelector // A value that specifies whether the rule is disabled. Once a rule is disabled, a // profile job will not validate it during a job run. Default value is false. Disabled bool // The map of substitution variable names to their values used in a check // expression. Variable names should start with a ':' (colon). Variable values can // either be actual values or column names. To differentiate between the two, // column names should be enclosed in backticks, for example, ":col1": "`Column // A`". SubstitutionMap map[string]string // The threshold used with a non-aggregate check expression. Non-aggregate check // expressions will be applied to each row in a specific column, and the threshold // will be used to determine whether the validation succeeds. Threshold *Threshold // contains filtered or unexported fields }
Represents a single data quality requirement that should be validated in the scope of this dataset.
type RulesetItem ¶
type RulesetItem struct { // The name of the ruleset. // // This member is required. Name *string // The Amazon Resource Name (ARN) of a resource (dataset) that the ruleset is // associated with. // // This member is required. TargetArn *string // The ID of the Amazon Web Services account that owns the ruleset. AccountId *string // The date and time that the ruleset was created. CreateDate *time.Time // The Amazon Resource Name (ARN) of the user who created the ruleset. CreatedBy *string // The description of the ruleset. Description *string // The Amazon Resource Name (ARN) of the user who last modified the ruleset. LastModifiedBy *string // The modification date and time of the ruleset. LastModifiedDate *time.Time // The Amazon Resource Name (ARN) for the ruleset. ResourceArn *string // The number of rules that are defined in the ruleset. RuleCount int32 // Metadata tags that have been applied to the ruleset. Tags map[string]string // contains filtered or unexported fields }
Contains metadata about the ruleset.
type S3Location ¶
type S3Location struct { // The Amazon S3 bucket name. // // This member is required. Bucket *string // The Amazon Web Services account ID of the bucket owner. BucketOwner *string // The unique name of the object in the bucket. Key *string // contains filtered or unexported fields }
Represents an Amazon S3 location (bucket name, bucket owner, and object key) where DataBrew can read input data, or write output from a job.
type S3TableOutputOptions ¶
type S3TableOutputOptions struct { // Represents an Amazon S3 location (bucket name and object key) where DataBrew // can write output from a job. // // This member is required. Location *S3Location // contains filtered or unexported fields }
Represents options that specify how and where DataBrew writes the Amazon S3 output generated by recipe jobs.
type Sample ¶
type Sample struct { // The way in which DataBrew obtains rows from a dataset. // // This member is required. Type SampleType // The number of rows in the sample. Size *int32 // contains filtered or unexported fields }
Represents the sample size and sampling type for DataBrew to use for interactive data analysis.
type SampleMode ¶
type SampleMode string
const ( SampleModeFullDataset SampleMode = "FULL_DATASET" SampleModeCustomRows SampleMode = "CUSTOM_ROWS" )
Enum values for SampleMode
func (SampleMode) Values ¶
func (SampleMode) Values() []SampleMode
Values returns all known values for SampleMode. Note that this can be expanded in the future, and so it is only as up to date as the client.
The ordering of this slice is not guaranteed to be stable across updates.
type SampleType ¶
type SampleType string
const ( SampleTypeFirstN SampleType = "FIRST_N" SampleTypeLastN SampleType = "LAST_N" SampleTypeRandom SampleType = "RANDOM" )
Enum values for SampleType
func (SampleType) Values ¶
func (SampleType) Values() []SampleType
Values returns all known values for SampleType. Note that this can be expanded in the future, and so it is only as up to date as the client.
The ordering of this slice is not guaranteed to be stable across updates.
type Schedule ¶
type Schedule struct { // The name of the schedule. // // This member is required. Name *string // The ID of the Amazon Web Services account that owns the schedule. AccountId *string // The date and time that the schedule was created. CreateDate *time.Time // The Amazon Resource Name (ARN) of the user who created the schedule. CreatedBy *string // The dates and times when the job is to run. For more information, see [Cron expressions] in the // Glue DataBrew Developer Guide. // // [Cron expressions]: https://docs.aws.amazon.com/databrew/latest/dg/jobs.cron.html CronExpression *string // A list of jobs to be run, according to the schedule. JobNames []string // The Amazon Resource Name (ARN) of the user who last modified the schedule. LastModifiedBy *string // The date and time when the schedule was last modified. LastModifiedDate *time.Time // The Amazon Resource Name (ARN) of the schedule. ResourceArn *string // Metadata tags that have been applied to the schedule. Tags map[string]string // contains filtered or unexported fields }
Represents one or more dates and times when a job is to run.
type ServiceQuotaExceededException ¶
type ServiceQuotaExceededException struct { Message *string ErrorCodeOverride *string // contains filtered or unexported fields }
A service quota is exceeded.
func (*ServiceQuotaExceededException) Error ¶
func (e *ServiceQuotaExceededException) Error() string
func (*ServiceQuotaExceededException) ErrorCode ¶
func (e *ServiceQuotaExceededException) ErrorCode() string
func (*ServiceQuotaExceededException) ErrorFault ¶
func (e *ServiceQuotaExceededException) ErrorFault() smithy.ErrorFault
func (*ServiceQuotaExceededException) ErrorMessage ¶
func (e *ServiceQuotaExceededException) ErrorMessage() string
type SessionStatus ¶
type SessionStatus string
const ( SessionStatusAssigned SessionStatus = "ASSIGNED" SessionStatusFailed SessionStatus = "FAILED" SessionStatusInitializing SessionStatus = "INITIALIZING" SessionStatusProvisioning SessionStatus = "PROVISIONING" SessionStatusReady SessionStatus = "READY" SessionStatusRecycling SessionStatus = "RECYCLING" SessionStatusRotating SessionStatus = "ROTATING" SessionStatusTerminated SessionStatus = "TERMINATED" SessionStatusTerminating SessionStatus = "TERMINATING" SessionStatusUpdating SessionStatus = "UPDATING" )
Enum values for SessionStatus
func (SessionStatus) Values ¶
func (SessionStatus) Values() []SessionStatus
Values returns all known values for SessionStatus. Note that this can be expanded in the future, and so it is only as up to date as the client.
The ordering of this slice is not guaranteed to be stable across updates.
type Source ¶
type Source string
type StatisticOverride ¶
type StatisticOverride struct { // A map that includes overrides of an evaluation’s parameters. // // This member is required. Parameters map[string]string // The name of an evaluation // // This member is required. Statistic *string // contains filtered or unexported fields }
Override of a particular evaluation for a profile job.
type StatisticsConfiguration ¶
type StatisticsConfiguration struct { // List of included evaluations. When the list is undefined, all supported // evaluations will be included. IncludedStatistics []string // List of overrides for evaluations. Overrides []StatisticOverride // contains filtered or unexported fields }
Configuration of evaluations for a profile job. This configuration can be used to select evaluations and override the parameters of selected evaluations.
type Threshold ¶
type Threshold struct { // The value of a threshold. // // This member is required. Value float64 // The type of a threshold. Used for comparison of an actual count of rows that // satisfy the rule to the threshold value. Type ThresholdType // Unit of threshold value. Can be either a COUNT or PERCENTAGE of the full sample // size used for validation. Unit ThresholdUnit // contains filtered or unexported fields }
The threshold used with a non-aggregate check expression. The non-aggregate check expression will be applied to each row in a specific column. Then the threshold will be used to determine whether the validation succeeds.
type ThresholdType ¶
type ThresholdType string
const ( ThresholdTypeGreaterThanOrEqual ThresholdType = "GREATER_THAN_OR_EQUAL" ThresholdTypeLessThanOrEqual ThresholdType = "LESS_THAN_OR_EQUAL" ThresholdTypeGreaterThan ThresholdType = "GREATER_THAN" ThresholdTypeLessThan ThresholdType = "LESS_THAN" )
Enum values for ThresholdType
func (ThresholdType) Values ¶
func (ThresholdType) Values() []ThresholdType
Values returns all known values for ThresholdType. Note that this can be expanded in the future, and so it is only as up to date as the client.
The ordering of this slice is not guaranteed to be stable across updates.
type ThresholdUnit ¶
type ThresholdUnit string
const ( ThresholdUnitCount ThresholdUnit = "COUNT" ThresholdUnitPercentage ThresholdUnit = "PERCENTAGE" )
Enum values for ThresholdUnit
func (ThresholdUnit) Values ¶
func (ThresholdUnit) Values() []ThresholdUnit
Values returns all known values for ThresholdUnit. Note that this can be expanded in the future, and so it is only as up to date as the client.
The ordering of this slice is not guaranteed to be stable across updates.
type ValidationConfiguration ¶
type ValidationConfiguration struct { // The Amazon Resource Name (ARN) for the ruleset to be validated in the profile // job. The TargetArn of the selected ruleset should be the same as the Amazon // Resource Name (ARN) of the dataset that is associated with the profile job. // // This member is required. RulesetArn *string // Mode of data quality validation. Default mode is “CHECK_ALL” which verifies all // rules defined in the selected ruleset. ValidationMode ValidationMode // contains filtered or unexported fields }
Configuration for data quality validation. Used to select the Rulesets and Validation Mode to be used in the profile job. When ValidationConfiguration is null, the profile job will run without data quality validation.
type ValidationException ¶
type ValidationException struct { Message *string ErrorCodeOverride *string // contains filtered or unexported fields }
The input parameters for this request failed validation.
func (*ValidationException) Error ¶
func (e *ValidationException) Error() string
func (*ValidationException) ErrorCode ¶
func (e *ValidationException) ErrorCode() string
func (*ValidationException) ErrorFault ¶
func (e *ValidationException) ErrorFault() smithy.ErrorFault
func (*ValidationException) ErrorMessage ¶
func (e *ValidationException) ErrorMessage() string
type ValidationMode ¶
type ValidationMode string
const (
ValidationModeCheckAll ValidationMode = "CHECK_ALL"
)
Enum values for ValidationMode
func (ValidationMode) Values ¶
func (ValidationMode) Values() []ValidationMode
Values returns all known values for ValidationMode. Note that this can be expanded in the future, and so it is only as up to date as the client.
The ordering of this slice is not guaranteed to be stable across updates.
type ViewFrame ¶
type ViewFrame struct { // The starting index for the range of columns to return in the view frame. // // This member is required. StartColumnIndex *int32 // Controls if analytics computation is enabled or disabled. Enabled by default. Analytics AnalyticsMode // The number of columns to include in the view frame, beginning with the // StartColumnIndex value and ignoring any columns in the HiddenColumns list. ColumnRange *int32 // A list of columns to hide in the view frame. HiddenColumns []string // The number of rows to include in the view frame, beginning with the // StartRowIndex value. RowRange *int32 // The starting index for the range of rows to return in the view frame. StartRowIndex *int32 // contains filtered or unexported fields }
Represents the data being transformed during an action.