Documentation ¶
Overview ¶
Package batch provides the API client, operations, and parameter types for AWS Batch.
Batch ¶
Using Batch, you can run batch computing workloads on the Amazon Web Services Cloud. Batch computing is a common means for developers, scientists, and engineers to access large amounts of compute resources. Batch uses the advantages of the batch computing to remove the undifferentiated heavy lifting of configuring and managing required infrastructure. At the same time, it also adopts a familiar batch computing software approach. You can use Batch to efficiently provision resources, and work toward eliminating capacity constraints, reducing your overall compute costs, and delivering results more quickly.
As a fully managed service, Batch can run batch computing workloads of any scale. Batch automatically provisions compute resources and optimizes workload distribution based on the quantity and scale of your specific workloads. With Batch, there's no need to install or manage batch computing software. This means that you can focus on analyzing results and solving your specific problems instead.
Index ¶
- Constants
- func NewDefaultEndpointResolver() *internalendpoints.Resolver
- func WithAPIOptions(optFns ...func(*middleware.Stack) error) func(*Options)
- func WithEndpointResolver(v EndpointResolver) func(*Options)deprecated
- func WithEndpointResolverV2(v EndpointResolverV2) func(*Options)
- func WithSigV4SigningName(name string) func(*Options)
- func WithSigV4SigningRegion(region string) func(*Options)
- type AuthResolverParameters
- type AuthSchemeResolver
- type CancelJobInput
- type CancelJobOutput
- type Client
- func (c *Client) CancelJob(ctx context.Context, params *CancelJobInput, optFns ...func(*Options)) (*CancelJobOutput, error)
- func (c *Client) CreateComputeEnvironment(ctx context.Context, params *CreateComputeEnvironmentInput, ...) (*CreateComputeEnvironmentOutput, error)
- func (c *Client) CreateJobQueue(ctx context.Context, params *CreateJobQueueInput, optFns ...func(*Options)) (*CreateJobQueueOutput, error)
- func (c *Client) CreateSchedulingPolicy(ctx context.Context, params *CreateSchedulingPolicyInput, ...) (*CreateSchedulingPolicyOutput, error)
- func (c *Client) DeleteComputeEnvironment(ctx context.Context, params *DeleteComputeEnvironmentInput, ...) (*DeleteComputeEnvironmentOutput, error)
- func (c *Client) DeleteJobQueue(ctx context.Context, params *DeleteJobQueueInput, optFns ...func(*Options)) (*DeleteJobQueueOutput, error)
- func (c *Client) DeleteSchedulingPolicy(ctx context.Context, params *DeleteSchedulingPolicyInput, ...) (*DeleteSchedulingPolicyOutput, error)
- func (c *Client) DeregisterJobDefinition(ctx context.Context, params *DeregisterJobDefinitionInput, ...) (*DeregisterJobDefinitionOutput, error)
- func (c *Client) DescribeComputeEnvironments(ctx context.Context, params *DescribeComputeEnvironmentsInput, ...) (*DescribeComputeEnvironmentsOutput, error)
- func (c *Client) DescribeJobDefinitions(ctx context.Context, params *DescribeJobDefinitionsInput, ...) (*DescribeJobDefinitionsOutput, error)
- func (c *Client) DescribeJobQueues(ctx context.Context, params *DescribeJobQueuesInput, optFns ...func(*Options)) (*DescribeJobQueuesOutput, error)
- func (c *Client) DescribeJobs(ctx context.Context, params *DescribeJobsInput, optFns ...func(*Options)) (*DescribeJobsOutput, error)
- func (c *Client) DescribeSchedulingPolicies(ctx context.Context, params *DescribeSchedulingPoliciesInput, ...) (*DescribeSchedulingPoliciesOutput, error)
- func (c *Client) ListJobs(ctx context.Context, params *ListJobsInput, optFns ...func(*Options)) (*ListJobsOutput, error)
- func (c *Client) ListSchedulingPolicies(ctx context.Context, params *ListSchedulingPoliciesInput, ...) (*ListSchedulingPoliciesOutput, error)
- func (c *Client) ListTagsForResource(ctx context.Context, params *ListTagsForResourceInput, ...) (*ListTagsForResourceOutput, error)
- func (c *Client) Options() Options
- func (c *Client) RegisterJobDefinition(ctx context.Context, params *RegisterJobDefinitionInput, ...) (*RegisterJobDefinitionOutput, error)
- func (c *Client) SubmitJob(ctx context.Context, params *SubmitJobInput, optFns ...func(*Options)) (*SubmitJobOutput, error)
- func (c *Client) TagResource(ctx context.Context, params *TagResourceInput, optFns ...func(*Options)) (*TagResourceOutput, error)
- func (c *Client) TerminateJob(ctx context.Context, params *TerminateJobInput, optFns ...func(*Options)) (*TerminateJobOutput, error)
- func (c *Client) UntagResource(ctx context.Context, params *UntagResourceInput, optFns ...func(*Options)) (*UntagResourceOutput, error)
- func (c *Client) UpdateComputeEnvironment(ctx context.Context, params *UpdateComputeEnvironmentInput, ...) (*UpdateComputeEnvironmentOutput, error)
- func (c *Client) UpdateJobQueue(ctx context.Context, params *UpdateJobQueueInput, optFns ...func(*Options)) (*UpdateJobQueueOutput, error)
- func (c *Client) UpdateSchedulingPolicy(ctx context.Context, params *UpdateSchedulingPolicyInput, ...) (*UpdateSchedulingPolicyOutput, error)
- type CreateComputeEnvironmentInput
- type CreateComputeEnvironmentOutput
- type CreateJobQueueInput
- type CreateJobQueueOutput
- type CreateSchedulingPolicyInput
- type CreateSchedulingPolicyOutput
- type DeleteComputeEnvironmentInput
- type DeleteComputeEnvironmentOutput
- type DeleteJobQueueInput
- type DeleteJobQueueOutput
- type DeleteSchedulingPolicyInput
- type DeleteSchedulingPolicyOutput
- type DeregisterJobDefinitionInput
- type DeregisterJobDefinitionOutput
- type DescribeComputeEnvironmentsAPIClient
- type DescribeComputeEnvironmentsInput
- type DescribeComputeEnvironmentsOutput
- type DescribeComputeEnvironmentsPaginator
- type DescribeComputeEnvironmentsPaginatorOptions
- type DescribeJobDefinitionsAPIClient
- type DescribeJobDefinitionsInput
- type DescribeJobDefinitionsOutput
- type DescribeJobDefinitionsPaginator
- type DescribeJobDefinitionsPaginatorOptions
- type DescribeJobQueuesAPIClient
- type DescribeJobQueuesInput
- type DescribeJobQueuesOutput
- type DescribeJobQueuesPaginator
- type DescribeJobQueuesPaginatorOptions
- type DescribeJobsInput
- type DescribeJobsOutput
- type DescribeSchedulingPoliciesInput
- type DescribeSchedulingPoliciesOutput
- type EndpointParameters
- type EndpointResolver
- type EndpointResolverFunc
- type EndpointResolverOptions
- type EndpointResolverV2
- type HTTPClient
- type HTTPSignerV4
- type ListJobsAPIClient
- type ListJobsInput
- type ListJobsOutput
- type ListJobsPaginator
- type ListJobsPaginatorOptions
- type ListSchedulingPoliciesAPIClient
- type ListSchedulingPoliciesInput
- type ListSchedulingPoliciesOutput
- type ListSchedulingPoliciesPaginator
- type ListSchedulingPoliciesPaginatorOptions
- type ListTagsForResourceInput
- type ListTagsForResourceOutput
- type Options
- type RegisterJobDefinitionInput
- type RegisterJobDefinitionOutput
- type ResolveEndpoint
- type SubmitJobInput
- type SubmitJobOutput
- type TagResourceInput
- type TagResourceOutput
- type TerminateJobInput
- type TerminateJobOutput
- type UntagResourceInput
- type UntagResourceOutput
- type UpdateComputeEnvironmentInput
- type UpdateComputeEnvironmentOutput
- type UpdateJobQueueInput
- type UpdateJobQueueOutput
- type UpdateSchedulingPolicyInput
- type UpdateSchedulingPolicyOutput
Constants ¶
const ServiceAPIVersion = "2016-08-10"
const ServiceID = "Batch"
Variables ¶
This section is empty.
Functions ¶
func NewDefaultEndpointResolver ¶
func NewDefaultEndpointResolver() *internalendpoints.Resolver
NewDefaultEndpointResolver constructs a new service endpoint resolver
func WithAPIOptions ¶
func WithAPIOptions(optFns ...func(*middleware.Stack) error) func(*Options)
WithAPIOptions returns a functional option for setting the Client's APIOptions option.
func WithEndpointResolver
deprecated
func WithEndpointResolver(v EndpointResolver) func(*Options)
Deprecated: EndpointResolver and WithEndpointResolver. Providing a value for this field will likely prevent you from using any endpoint-related service features released after the introduction of EndpointResolverV2 and BaseEndpoint.
To migrate an EndpointResolver implementation that uses a custom endpoint, set the client option BaseEndpoint instead.
func WithEndpointResolverV2 ¶
func WithEndpointResolverV2(v EndpointResolverV2) func(*Options)
WithEndpointResolverV2 returns a functional option for setting the Client's EndpointResolverV2 option.
func WithSigV4SigningName ¶
WithSigV4SigningName applies an override to the authentication workflow to use the given signing name for SigV4-authenticated operations.
This is an advanced setting. The value here is FINAL, taking precedence over the resolved signing name from both auth scheme resolution and endpoint resolution.
func WithSigV4SigningRegion ¶
WithSigV4SigningRegion applies an override to the authentication workflow to use the given signing region for SigV4-authenticated operations.
This is an advanced setting. The value here is FINAL, taking precedence over the resolved signing region from both auth scheme resolution and endpoint resolution.
Types ¶
type AuthResolverParameters ¶
type AuthResolverParameters struct { // The name of the operation being invoked. Operation string // The region in which the operation is being invoked. Region string }
AuthResolverParameters contains the set of inputs necessary for auth scheme resolution.
type AuthSchemeResolver ¶
type AuthSchemeResolver interface {
ResolveAuthSchemes(context.Context, *AuthResolverParameters) ([]*smithyauth.Option, error)
}
AuthSchemeResolver returns a set of possible authentication options for an operation.
type CancelJobInput ¶
type CancelJobInput struct { // The Batch job ID of the job to cancel. // // This member is required. JobId *string // A message to attach to the job that explains the reason for canceling it. This // message is returned by future DescribeJobsoperations on the job. This message is also // recorded in the Batch activity logs. // // This member is required. Reason *string // contains filtered or unexported fields }
Contains the parameters for CancelJob .
type CancelJobOutput ¶
type CancelJobOutput struct { // Metadata pertaining to the operation's result. ResultMetadata middleware.Metadata // contains filtered or unexported fields }
type Client ¶
type Client struct {
// contains filtered or unexported fields
}
Client provides the API client to make operations call for AWS Batch.
func New ¶
New returns an initialized Client based on the functional options. Provide additional functional options to further configure the behavior of the client, such as changing the client's endpoint or adding custom middleware behavior.
func NewFromConfig ¶
NewFromConfig returns a new client from the provided config.
func (*Client) CancelJob ¶
func (c *Client) CancelJob(ctx context.Context, params *CancelJobInput, optFns ...func(*Options)) (*CancelJobOutput, error)
Cancels a job in an Batch job queue. Jobs that are in the SUBMITTED or PENDING are canceled. A job in RUNNABLE remains in RUNNABLE until it reaches the head of the job queue. Then the job status is updated to FAILED .
A PENDING job is canceled after all dependency jobs are completed. Therefore, it may take longer than expected to cancel a job in PENDING status.
When you try to cancel an array parent job in PENDING , Batch attempts to cancel all child jobs. The array parent job is canceled when all child jobs are completed.
Jobs that progressed to the STARTING or RUNNING state aren't canceled. However, the API operation still succeeds, even if no job is canceled. These jobs must be terminated with the TerminateJoboperation.
func (*Client) CreateComputeEnvironment ¶
func (c *Client) CreateComputeEnvironment(ctx context.Context, params *CreateComputeEnvironmentInput, optFns ...func(*Options)) (*CreateComputeEnvironmentOutput, error)
Creates an Batch compute environment. You can create MANAGED or UNMANAGED compute environments. MANAGED compute environments can use Amazon EC2 or Fargate resources. UNMANAGED compute environments can only use EC2 resources.
In a managed compute environment, Batch manages the capacity and instance types of the compute resources within the environment. This is based on the compute resource specification that you define or the launch templatethat you specify when you create the compute environment. Either, you can choose to use EC2 On-Demand Instances and EC2 Spot Instances. Or, you can use Fargate and Fargate Spot capacity in your managed compute environment. You can optionally set a maximum price so that Spot Instances only launch when the Spot Instance price is less than a specified percentage of the On-Demand price.
Multi-node parallel jobs aren't supported on Spot Instances.
In an unmanaged compute environment, you can manage your own EC2 compute resources and have flexibility with how you configure your compute resources. For example, you can use custom AMIs. However, you must verify that each of your AMIs meet the Amazon ECS container instance AMI specification. For more information, see container instance AMIsin the Amazon Elastic Container Service Developer Guide. After you created your unmanaged compute environment, you can use the DescribeComputeEnvironmentsoperation to find the Amazon ECS cluster that's associated with it. Then, launch your container instances into that Amazon ECS cluster. For more information, see Launching an Amazon ECS container instancein the Amazon Elastic Container Service Developer Guide.
To create a compute environment that uses EKS resources, the caller must have permissions to call eks:DescribeCluster .
Batch doesn't automatically upgrade the AMIs in a compute environment after it's created. For example, it also doesn't update the AMIs in your compute environment when a newer version of the Amazon ECS optimized AMI is available. You're responsible for the management of the guest operating system. This includes any updates and security patches. You're also responsible for any additional application software or utilities that you install on the compute resources. There are two ways to use a new AMI for your Batch jobs. The original method is to complete these steps:
Create a new compute environment with the new AMI.
Add the compute environment to an existing job queue.
Remove the earlier compute environment from your job queue.
Delete the earlier compute environment.
In April 2022, Batch added enhanced support for updating compute environments. For more information, see Updating compute environments. To use the enhanced updating of compute environments to update AMIs, follow these rules:
Either don't set the service role ( serviceRole ) parameter or set it to the AWSBatchServiceRole service-linked role.
Set the allocation strategy ( allocationStrategy ) parameter to BEST_FIT_PROGRESSIVE , SPOT_CAPACITY_OPTIMIZED , or SPOT_PRICE_CAPACITY_OPTIMIZED .
Set the update to latest image version ( updateToLatestImageVersion ) parameter to true . The updateToLatestImageVersion parameter is used when you update a compute environment. This parameter is ignored when you create a compute environment.
Don't specify an AMI ID in imageId , imageIdOverride (in ec2Configurationec2Configuration ), or in the launch template ( launchTemplate ). In that case, Batch selects the latest Amazon ECS optimized AMI that's supported by Batch at the time the infrastructure update is initiated. Alternatively, you can specify the AMI ID in the imageId or imageIdOverride parameters, or the launch template identified by the LaunchTemplate properties. Changing any of these properties starts an infrastructure update. If the AMI ID is specified in the launch template, it can't be replaced by specifying an AMI ID in either the imageId or imageIdOverride parameters. It can only be replaced by specifying a different launch template, or if the launch template version is set to $Default or $Latest , by setting either a new default version for the launch template (if $Default ) or by adding a new version to the launch template (if $Latest ).
If these rules are followed, any update that starts an infrastructure update causes the AMI ID to be re-selected. If the version setting in the launch template ( launchTemplate ) is set to $Latest or $Default , the latest or default version of the launch template is evaluated up at the time of the infrastructure update, even if the launchTemplate wasn't updated.
func (*Client) CreateJobQueue ¶
func (c *Client) CreateJobQueue(ctx context.Context, params *CreateJobQueueInput, optFns ...func(*Options)) (*CreateJobQueueOutput, error)
Creates an Batch job queue. When you create a job queue, you associate one or more compute environments to the queue and assign an order of preference for the compute environments.
You also set a priority to the job queue that determines the order that the Batch scheduler places jobs onto its associated compute environments. For example, if a compute environment is associated with more than one job queue, the job queue with a higher priority is given preference for scheduling jobs to that compute environment.
func (*Client) CreateSchedulingPolicy ¶
func (c *Client) CreateSchedulingPolicy(ctx context.Context, params *CreateSchedulingPolicyInput, optFns ...func(*Options)) (*CreateSchedulingPolicyOutput, error)
Creates an Batch scheduling policy.
func (*Client) DeleteComputeEnvironment ¶
func (c *Client) DeleteComputeEnvironment(ctx context.Context, params *DeleteComputeEnvironmentInput, optFns ...func(*Options)) (*DeleteComputeEnvironmentOutput, error)
Deletes an Batch compute environment.
Before you can delete a compute environment, you must set its state to DISABLED with the UpdateComputeEnvironmentAPI operation and disassociate it from any job queues with the UpdateJobQueue API operation. Compute environments that use Fargate resources must terminate all active jobs on that compute environment before deleting the compute environment. If this isn't done, the compute environment enters an invalid state.
func (*Client) DeleteJobQueue ¶
func (c *Client) DeleteJobQueue(ctx context.Context, params *DeleteJobQueueInput, optFns ...func(*Options)) (*DeleteJobQueueOutput, error)
Deletes the specified job queue. You must first disable submissions for a queue with the UpdateJobQueueoperation. All jobs in the queue are eventually terminated when you delete a job queue. The jobs are terminated at a rate of about 16 jobs each second.
It's not necessary to disassociate compute environments from a queue before submitting a DeleteJobQueue request.
func (*Client) DeleteSchedulingPolicy ¶
func (c *Client) DeleteSchedulingPolicy(ctx context.Context, params *DeleteSchedulingPolicyInput, optFns ...func(*Options)) (*DeleteSchedulingPolicyOutput, error)
Deletes the specified scheduling policy.
You can't delete a scheduling policy that's used in any job queues.
func (*Client) DeregisterJobDefinition ¶
func (c *Client) DeregisterJobDefinition(ctx context.Context, params *DeregisterJobDefinitionInput, optFns ...func(*Options)) (*DeregisterJobDefinitionOutput, error)
Deregisters an Batch job definition. Job definitions are permanently deleted after 180 days.
func (*Client) DescribeComputeEnvironments ¶
func (c *Client) DescribeComputeEnvironments(ctx context.Context, params *DescribeComputeEnvironmentsInput, optFns ...func(*Options)) (*DescribeComputeEnvironmentsOutput, error)
Describes one or more of your compute environments.
If you're using an unmanaged compute environment, you can use the DescribeComputeEnvironment operation to determine the ecsClusterArn that you launch your Amazon ECS container instances into.
func (*Client) DescribeJobDefinitions ¶
func (c *Client) DescribeJobDefinitions(ctx context.Context, params *DescribeJobDefinitionsInput, optFns ...func(*Options)) (*DescribeJobDefinitionsOutput, error)
Describes a list of job definitions. You can specify a status (such as ACTIVE ) to only return job definitions that match that status.
func (*Client) DescribeJobQueues ¶
func (c *Client) DescribeJobQueues(ctx context.Context, params *DescribeJobQueuesInput, optFns ...func(*Options)) (*DescribeJobQueuesOutput, error)
Describes one or more of your job queues.
func (*Client) DescribeJobs ¶
func (c *Client) DescribeJobs(ctx context.Context, params *DescribeJobsInput, optFns ...func(*Options)) (*DescribeJobsOutput, error)
Describes a list of Batch jobs.
func (*Client) DescribeSchedulingPolicies ¶
func (c *Client) DescribeSchedulingPolicies(ctx context.Context, params *DescribeSchedulingPoliciesInput, optFns ...func(*Options)) (*DescribeSchedulingPoliciesOutput, error)
Describes one or more of your scheduling policies.
func (*Client) ListJobs ¶
func (c *Client) ListJobs(ctx context.Context, params *ListJobsInput, optFns ...func(*Options)) (*ListJobsOutput, error)
Returns a list of Batch jobs.
You must specify only one of the following items:
A job queue ID to return a list of jobs in that job queue
A multi-node parallel job ID to return a list of nodes for that job
An array job ID to return a list of the children for that job
You can filter the results by job status with the jobStatus parameter. If you don't specify a status, only RUNNING jobs are returned.
func (*Client) ListSchedulingPolicies ¶
func (c *Client) ListSchedulingPolicies(ctx context.Context, params *ListSchedulingPoliciesInput, optFns ...func(*Options)) (*ListSchedulingPoliciesOutput, error)
Returns a list of Batch scheduling policies.
func (*Client) ListTagsForResource ¶
func (c *Client) ListTagsForResource(ctx context.Context, params *ListTagsForResourceInput, optFns ...func(*Options)) (*ListTagsForResourceOutput, error)
Lists the tags for an Batch resource. Batch resources that support tags are compute environments, jobs, job definitions, job queues, and scheduling policies. ARNs for child jobs of array and multi-node parallel (MNP) jobs aren't supported.
func (*Client) Options ¶
Options returns a copy of the client configuration.
Callers SHOULD NOT perform mutations on any inner structures within client config. Config overrides should instead be made on a per-operation basis through functional options.
func (*Client) RegisterJobDefinition ¶
func (c *Client) RegisterJobDefinition(ctx context.Context, params *RegisterJobDefinitionInput, optFns ...func(*Options)) (*RegisterJobDefinitionOutput, error)
Registers an Batch job definition.
func (*Client) SubmitJob ¶
func (c *Client) SubmitJob(ctx context.Context, params *SubmitJobInput, optFns ...func(*Options)) (*SubmitJobOutput, error)
Submits an Batch job from a job definition. Parameters that are specified during SubmitJoboverride parameters defined in the job definition. vCPU and memory requirements that are specified in the resourceRequirements objects in the job definition are the exception. They can't be overridden this way using the memory and vcpus parameters. Rather, you must specify updates to job definition parameters in a resourceRequirements object that's included in the containerOverrides parameter.
Job queues with a scheduling policy are limited to 500 active fair share identifiers at a time.
Jobs that run on Fargate resources can't be guaranteed to run for more than 14 days. This is because, after 14 days, Fargate resources might become unavailable and job might be terminated.
func (*Client) TagResource ¶
func (c *Client) TagResource(ctx context.Context, params *TagResourceInput, optFns ...func(*Options)) (*TagResourceOutput, error)
Associates the specified tags to a resource with the specified resourceArn . If existing tags on a resource aren't specified in the request parameters, they aren't changed. When a resource is deleted, the tags that are associated with that resource are deleted as well. Batch resources that support tags are compute environments, jobs, job definitions, job queues, and scheduling policies. ARNs for child jobs of array and multi-node parallel (MNP) jobs aren't supported.
func (*Client) TerminateJob ¶
func (c *Client) TerminateJob(ctx context.Context, params *TerminateJobInput, optFns ...func(*Options)) (*TerminateJobOutput, error)
Terminates a job in a job queue. Jobs that are in the STARTING or RUNNING state are terminated, which causes them to transition to FAILED . Jobs that have not progressed to the STARTING state are cancelled.
func (*Client) UntagResource ¶
func (c *Client) UntagResource(ctx context.Context, params *UntagResourceInput, optFns ...func(*Options)) (*UntagResourceOutput, error)
Deletes specified tags from an Batch resource.
func (*Client) UpdateComputeEnvironment ¶
func (c *Client) UpdateComputeEnvironment(ctx context.Context, params *UpdateComputeEnvironmentInput, optFns ...func(*Options)) (*UpdateComputeEnvironmentOutput, error)
Updates an Batch compute environment.
func (*Client) UpdateJobQueue ¶
func (c *Client) UpdateJobQueue(ctx context.Context, params *UpdateJobQueueInput, optFns ...func(*Options)) (*UpdateJobQueueOutput, error)
Updates a job queue.
func (*Client) UpdateSchedulingPolicy ¶
func (c *Client) UpdateSchedulingPolicy(ctx context.Context, params *UpdateSchedulingPolicyInput, optFns ...func(*Options)) (*UpdateSchedulingPolicyOutput, error)
Updates a scheduling policy.
type CreateComputeEnvironmentInput ¶
type CreateComputeEnvironmentInput struct { // The name for your compute environment. It can be up to 128 characters long. It // can contain uppercase and lowercase letters, numbers, hyphens (-), and // underscores (_). // // This member is required. ComputeEnvironmentName *string // The type of the compute environment: MANAGED or UNMANAGED . For more // information, see [Compute Environments]in the Batch User Guide. // // [Compute Environments]: https://docs.aws.amazon.com/batch/latest/userguide/compute_environments.html // // This member is required. Type types.CEType // Details about the compute resources managed by the compute environment. This // parameter is required for managed compute environments. For more information, // see [Compute Environments]in the Batch User Guide. // // [Compute Environments]: https://docs.aws.amazon.com/batch/latest/userguide/compute_environments.html ComputeResources *types.ComputeResource // The details for the Amazon EKS cluster that supports the compute environment. EksConfiguration *types.EksConfiguration // The full Amazon Resource Name (ARN) of the IAM role that allows Batch to make // calls to other Amazon Web Services services on your behalf. For more // information, see [Batch service IAM role]in the Batch User Guide. // // If your account already created the Batch service-linked role, that role is // used by default for your compute environment unless you specify a different role // here. If the Batch service-linked role doesn't exist in your account, and no // role is specified here, the service attempts to create the Batch service-linked // role in your account. // // If your specified role has a path other than / , then you must specify either // the full role ARN (recommended) or prefix the role name with the path. For // example, if a role with the name bar has a path of /foo/ , specify /foo/bar as // the role name. For more information, see [Friendly names and paths]in the IAM User Guide. // // Depending on how you created your Batch service role, its ARN might contain the // service-role path prefix. When you only specify the name of the service role, // Batch assumes that your ARN doesn't use the service-role path prefix. Because // of this, we recommend that you specify the full ARN of your service role when // you create compute environments. // // [Friendly names and paths]: https://docs.aws.amazon.com/IAM/latest/UserGuide/reference_identifiers.html#identifiers-friendly-names // [Batch service IAM role]: https://docs.aws.amazon.com/batch/latest/userguide/service_IAM_role.html ServiceRole *string // The state of the compute environment. If the state is ENABLED , then the compute // environment accepts jobs from a queue and can scale out automatically based on // queues. // // If the state is ENABLED , then the Batch scheduler can attempt to place jobs // from an associated job queue on the compute resources within the environment. If // the compute environment is managed, then it can scale its instances out or in // automatically, based on the job queue demand. // // If the state is DISABLED , then the Batch scheduler doesn't attempt to place // jobs within the environment. Jobs in a STARTING or RUNNING state continue to // progress normally. Managed compute environments in the DISABLED state don't // scale out. // // Compute environments in a DISABLED state may continue to incur billing charges. // To prevent additional charges, turn off and then delete the compute environment. // For more information, see [State]in the Batch User Guide. // // When an instance is idle, the instance scales down to the minvCpus value. // However, the instance size doesn't change. For example, consider a c5.8xlarge // instance with a minvCpus value of 4 and a desiredvCpus value of 36 . This // instance doesn't scale down to a c5.large instance. // // [State]: https://docs.aws.amazon.com/batch/latest/userguide/compute_environment_parameters.html#compute_environment_state State types.CEState // The tags that you apply to the compute environment to help you categorize and // organize your resources. Each tag consists of a key and an optional value. For // more information, see [Tagging Amazon Web Services Resources]in Amazon Web Services General Reference. // // These tags can be updated or removed using the [TagResource] and [UntagResource] API operations. These tags // don't propagate to the underlying compute resources. // // [Tagging Amazon Web Services Resources]: https://docs.aws.amazon.com/general/latest/gr/aws_tagging.html // [TagResource]: https://docs.aws.amazon.com/batch/latest/APIReference/API_TagResource.html // [UntagResource]: https://docs.aws.amazon.com/batch/latest/APIReference/API_UntagResource.html Tags map[string]string // The maximum number of vCPUs for an unmanaged compute environment. This // parameter is only used for fair share scheduling to reserve vCPU capacity for // new share identifiers. If this parameter isn't provided for a fair share job // queue, no vCPU capacity is reserved. // // This parameter is only supported when the type parameter is set to UNMANAGED . UnmanagedvCpus *int32 // contains filtered or unexported fields }
Contains the parameters for CreateComputeEnvironment .
type CreateComputeEnvironmentOutput ¶
type CreateComputeEnvironmentOutput struct { // The Amazon Resource Name (ARN) of the compute environment. ComputeEnvironmentArn *string // The name of the compute environment. It can be up to 128 characters long. It // can contain uppercase and lowercase letters, numbers, hyphens (-), and // underscores (_). ComputeEnvironmentName *string // Metadata pertaining to the operation's result. ResultMetadata middleware.Metadata // contains filtered or unexported fields }
type CreateJobQueueInput ¶
type CreateJobQueueInput struct { // The set of compute environments mapped to a job queue and their order relative // to each other. The job scheduler uses this parameter to determine which compute // environment runs a specific job. Compute environments must be in the VALID // state before you can associate them with a job queue. You can associate up to // three compute environments with a job queue. All of the compute environments // must be either EC2 ( EC2 or SPOT ) or Fargate ( FARGATE or FARGATE_SPOT ); EC2 // and Fargate compute environments can't be mixed. // // All compute environments that are associated with a job queue must share the // same architecture. Batch doesn't support mixing compute environment architecture // types in a single job queue. // // This member is required. ComputeEnvironmentOrder []types.ComputeEnvironmentOrder // The name of the job queue. It can be up to 128 letters long. It can contain // uppercase and lowercase letters, numbers, hyphens (-), and underscores (_). // // This member is required. JobQueueName *string // The priority of the job queue. Job queues with a higher priority (or a higher // integer value for the priority parameter) are evaluated first when associated // with the same compute environment. Priority is determined in descending order. // For example, a job queue with a priority value of 10 is given scheduling // preference over a job queue with a priority value of 1 . All of the compute // environments must be either EC2 ( EC2 or SPOT ) or Fargate ( FARGATE or // FARGATE_SPOT ); EC2 and Fargate compute environments can't be mixed. // // This member is required. Priority *int32 // The set of actions that Batch performs on jobs that remain at the head of the // job queue in the specified state longer than specified times. Batch will perform // each action after maxTimeSeconds has passed. JobStateTimeLimitActions []types.JobStateTimeLimitAction // The Amazon Resource Name (ARN) of the fair share scheduling policy. If this // parameter is specified, the job queue uses a fair share scheduling policy. If // this parameter isn't specified, the job queue uses a first in, first out (FIFO) // scheduling policy. After a job queue is created, you can replace but can't // remove the fair share scheduling policy. The format is // aws:Partition:batch:Region:Account:scheduling-policy/Name . An example is // aws:aws:batch:us-west-2:123456789012:scheduling-policy/MySchedulingPolicy . SchedulingPolicyArn *string // The state of the job queue. If the job queue state is ENABLED , it is able to // accept jobs. If the job queue state is DISABLED , new jobs can't be added to the // queue, but jobs already in the queue can finish. State types.JQState // The tags that you apply to the job queue to help you categorize and organize // your resources. Each tag consists of a key and an optional value. For more // information, see [Tagging your Batch resources]in Batch User Guide. // // [Tagging your Batch resources]: https://docs.aws.amazon.com/batch/latest/userguide/using-tags.html Tags map[string]string // contains filtered or unexported fields }
Contains the parameters for CreateJobQueue .
type CreateJobQueueOutput ¶
type CreateJobQueueOutput struct { // The Amazon Resource Name (ARN) of the job queue. // // This member is required. JobQueueArn *string // The name of the job queue. // // This member is required. JobQueueName *string // Metadata pertaining to the operation's result. ResultMetadata middleware.Metadata // contains filtered or unexported fields }
type CreateSchedulingPolicyInput ¶
type CreateSchedulingPolicyInput struct { // The name of the scheduling policy. It can be up to 128 letters long. It can // contain uppercase and lowercase letters, numbers, hyphens (-), and underscores // (_). // // This member is required. Name *string FairsharePolicy *types.FairsharePolicy // The tags that you apply to the scheduling policy to help you categorize and // organize your resources. Each tag consists of a key and an optional value. For // more information, see [Tagging Amazon Web Services Resources]in Amazon Web Services General Reference. // // These tags can be updated or removed using the [TagResource] and [UntagResource] API operations. // // [Tagging Amazon Web Services Resources]: https://docs.aws.amazon.com/general/latest/gr/aws_tagging.html // [TagResource]: https://docs.aws.amazon.com/batch/latest/APIReference/API_TagResource.html // [UntagResource]: https://docs.aws.amazon.com/batch/latest/APIReference/API_UntagResource.html Tags map[string]string // contains filtered or unexported fields }
Contains the parameters for CreateSchedulingPolicy .
type CreateSchedulingPolicyOutput ¶
type CreateSchedulingPolicyOutput struct { // The Amazon Resource Name (ARN) of the scheduling policy. The format is // aws:Partition:batch:Region:Account:scheduling-policy/Name . For example, // aws:aws:batch:us-west-2:123456789012:scheduling-policy/MySchedulingPolicy . // // This member is required. Arn *string // The name of the scheduling policy. // // This member is required. Name *string // Metadata pertaining to the operation's result. ResultMetadata middleware.Metadata // contains filtered or unexported fields }
type DeleteComputeEnvironmentInput ¶
type DeleteComputeEnvironmentInput struct { // The name or Amazon Resource Name (ARN) of the compute environment to delete. // // This member is required. ComputeEnvironment *string // contains filtered or unexported fields }
Contains the parameters for DeleteComputeEnvironment .
type DeleteComputeEnvironmentOutput ¶
type DeleteComputeEnvironmentOutput struct { // Metadata pertaining to the operation's result. ResultMetadata middleware.Metadata // contains filtered or unexported fields }
type DeleteJobQueueInput ¶
type DeleteJobQueueInput struct { // The short name or full Amazon Resource Name (ARN) of the queue to delete. // // This member is required. JobQueue *string // contains filtered or unexported fields }
Contains the parameters for DeleteJobQueue .
type DeleteJobQueueOutput ¶
type DeleteJobQueueOutput struct { // Metadata pertaining to the operation's result. ResultMetadata middleware.Metadata // contains filtered or unexported fields }
type DeleteSchedulingPolicyInput ¶
type DeleteSchedulingPolicyInput struct { // The Amazon Resource Name (ARN) of the scheduling policy to delete. // // This member is required. Arn *string // contains filtered or unexported fields }
Contains the parameters for DeleteSchedulingPolicy .
type DeleteSchedulingPolicyOutput ¶
type DeleteSchedulingPolicyOutput struct { // Metadata pertaining to the operation's result. ResultMetadata middleware.Metadata // contains filtered or unexported fields }
type DeregisterJobDefinitionInput ¶
type DeregisterJobDefinitionInput struct { // The name and revision ( name:revision ) or full Amazon Resource Name (ARN) of // the job definition to deregister. // // This member is required. JobDefinition *string // contains filtered or unexported fields }
type DeregisterJobDefinitionOutput ¶
type DeregisterJobDefinitionOutput struct { // Metadata pertaining to the operation's result. ResultMetadata middleware.Metadata // contains filtered or unexported fields }
type DescribeComputeEnvironmentsAPIClient ¶
type DescribeComputeEnvironmentsAPIClient interface {
DescribeComputeEnvironments(context.Context, *DescribeComputeEnvironmentsInput, ...func(*Options)) (*DescribeComputeEnvironmentsOutput, error)
}
DescribeComputeEnvironmentsAPIClient is a client that implements the DescribeComputeEnvironments operation.
type DescribeComputeEnvironmentsInput ¶
type DescribeComputeEnvironmentsInput struct { // A list of up to 100 compute environment names or full Amazon Resource Name // (ARN) entries. ComputeEnvironments []string // The maximum number of cluster results returned by DescribeComputeEnvironments // in paginated output. When this parameter is used, DescribeComputeEnvironments // only returns maxResults results in a single page along with a nextToken // response element. The remaining results of the initial request can be seen by // sending another DescribeComputeEnvironments request with the returned nextToken // value. This value can be between 1 and 100. If this parameter isn't used, then // DescribeComputeEnvironments returns up to 100 results and a nextToken value if // applicable. MaxResults *int32 // The nextToken value returned from a previous paginated // DescribeComputeEnvironments request where maxResults was used and the results // exceeded the value of that parameter. Pagination continues from the end of the // previous results that returned the nextToken value. This value is null when // there are no more results to return. // // Treat this token as an opaque identifier that's only used to retrieve the next // items in a list and not for other programmatic purposes. NextToken *string // contains filtered or unexported fields }
Contains the parameters for DescribeComputeEnvironments .
type DescribeComputeEnvironmentsOutput ¶
type DescribeComputeEnvironmentsOutput struct { // The list of compute environments. ComputeEnvironments []types.ComputeEnvironmentDetail // The nextToken value to include in a future DescribeComputeEnvironments request. // When the results of a DescribeComputeEnvironments request exceed maxResults , // this value can be used to retrieve the next page of results. This value is null // when there are no more results to return. NextToken *string // Metadata pertaining to the operation's result. ResultMetadata middleware.Metadata // contains filtered or unexported fields }
type DescribeComputeEnvironmentsPaginator ¶
type DescribeComputeEnvironmentsPaginator struct {
// contains filtered or unexported fields
}
DescribeComputeEnvironmentsPaginator is a paginator for DescribeComputeEnvironments
func NewDescribeComputeEnvironmentsPaginator ¶
func NewDescribeComputeEnvironmentsPaginator(client DescribeComputeEnvironmentsAPIClient, params *DescribeComputeEnvironmentsInput, optFns ...func(*DescribeComputeEnvironmentsPaginatorOptions)) *DescribeComputeEnvironmentsPaginator
NewDescribeComputeEnvironmentsPaginator returns a new DescribeComputeEnvironmentsPaginator
func (*DescribeComputeEnvironmentsPaginator) HasMorePages ¶
func (p *DescribeComputeEnvironmentsPaginator) HasMorePages() bool
HasMorePages returns a boolean indicating whether more pages are available
func (*DescribeComputeEnvironmentsPaginator) NextPage ¶
func (p *DescribeComputeEnvironmentsPaginator) NextPage(ctx context.Context, optFns ...func(*Options)) (*DescribeComputeEnvironmentsOutput, error)
NextPage retrieves the next DescribeComputeEnvironments page.
type DescribeComputeEnvironmentsPaginatorOptions ¶
type DescribeComputeEnvironmentsPaginatorOptions struct { // The maximum number of cluster results returned by DescribeComputeEnvironments // in paginated output. When this parameter is used, DescribeComputeEnvironments // only returns maxResults results in a single page along with a nextToken // response element. The remaining results of the initial request can be seen by // sending another DescribeComputeEnvironments request with the returned nextToken // value. This value can be between 1 and 100. If this parameter isn't used, then // DescribeComputeEnvironments returns up to 100 results and a nextToken value if // applicable. Limit int32 // Set to true if pagination should stop if the service returns a pagination token // that matches the most recent token provided to the service. StopOnDuplicateToken bool }
DescribeComputeEnvironmentsPaginatorOptions is the paginator options for DescribeComputeEnvironments
type DescribeJobDefinitionsAPIClient ¶
type DescribeJobDefinitionsAPIClient interface {
DescribeJobDefinitions(context.Context, *DescribeJobDefinitionsInput, ...func(*Options)) (*DescribeJobDefinitionsOutput, error)
}
DescribeJobDefinitionsAPIClient is a client that implements the DescribeJobDefinitions operation.
type DescribeJobDefinitionsInput ¶
type DescribeJobDefinitionsInput struct { // The name of the job definition to describe. JobDefinitionName *string // A list of up to 100 job definitions. Each entry in the list can either be an // ARN in the format // arn:aws:batch:${Region}:${Account}:job-definition/${JobDefinitionName}:${Revision} // or a short version using the form ${JobDefinitionName}:${Revision} . This // parameter can't be used with other parameters. JobDefinitions []string // The maximum number of results returned by DescribeJobDefinitions in paginated // output. When this parameter is used, DescribeJobDefinitions only returns // maxResults results in a single page and a nextToken response element. The // remaining results of the initial request can be seen by sending another // DescribeJobDefinitions request with the returned nextToken value. This value // can be between 1 and 100. If this parameter isn't used, then // DescribeJobDefinitions returns up to 100 results and a nextToken value if // applicable. MaxResults *int32 // The nextToken value returned from a previous paginated DescribeJobDefinitions // request where maxResults was used and the results exceeded the value of that // parameter. Pagination continues from the end of the previous results that // returned the nextToken value. This value is null when there are no more results // to return. // // Treat this token as an opaque identifier that's only used to retrieve the next // items in a list and not for other programmatic purposes. NextToken *string // The status used to filter job definitions. Status *string // contains filtered or unexported fields }
Contains the parameters for DescribeJobDefinitions .
type DescribeJobDefinitionsOutput ¶
type DescribeJobDefinitionsOutput struct { // The list of job definitions. JobDefinitions []types.JobDefinition // The nextToken value to include in a future DescribeJobDefinitions request. When // the results of a DescribeJobDefinitions request exceed maxResults , this value // can be used to retrieve the next page of results. This value is null when there // are no more results to return. NextToken *string // Metadata pertaining to the operation's result. ResultMetadata middleware.Metadata // contains filtered or unexported fields }
type DescribeJobDefinitionsPaginator ¶
type DescribeJobDefinitionsPaginator struct {
// contains filtered or unexported fields
}
DescribeJobDefinitionsPaginator is a paginator for DescribeJobDefinitions
func NewDescribeJobDefinitionsPaginator ¶
func NewDescribeJobDefinitionsPaginator(client DescribeJobDefinitionsAPIClient, params *DescribeJobDefinitionsInput, optFns ...func(*DescribeJobDefinitionsPaginatorOptions)) *DescribeJobDefinitionsPaginator
NewDescribeJobDefinitionsPaginator returns a new DescribeJobDefinitionsPaginator
func (*DescribeJobDefinitionsPaginator) HasMorePages ¶
func (p *DescribeJobDefinitionsPaginator) HasMorePages() bool
HasMorePages returns a boolean indicating whether more pages are available
func (*DescribeJobDefinitionsPaginator) NextPage ¶
func (p *DescribeJobDefinitionsPaginator) NextPage(ctx context.Context, optFns ...func(*Options)) (*DescribeJobDefinitionsOutput, error)
NextPage retrieves the next DescribeJobDefinitions page.
type DescribeJobDefinitionsPaginatorOptions ¶
type DescribeJobDefinitionsPaginatorOptions struct { // The maximum number of results returned by DescribeJobDefinitions in paginated // output. When this parameter is used, DescribeJobDefinitions only returns // maxResults results in a single page and a nextToken response element. The // remaining results of the initial request can be seen by sending another // DescribeJobDefinitions request with the returned nextToken value. This value // can be between 1 and 100. If this parameter isn't used, then // DescribeJobDefinitions returns up to 100 results and a nextToken value if // applicable. Limit int32 // Set to true if pagination should stop if the service returns a pagination token // that matches the most recent token provided to the service. StopOnDuplicateToken bool }
DescribeJobDefinitionsPaginatorOptions is the paginator options for DescribeJobDefinitions
type DescribeJobQueuesAPIClient ¶
type DescribeJobQueuesAPIClient interface {
DescribeJobQueues(context.Context, *DescribeJobQueuesInput, ...func(*Options)) (*DescribeJobQueuesOutput, error)
}
DescribeJobQueuesAPIClient is a client that implements the DescribeJobQueues operation.
type DescribeJobQueuesInput ¶
type DescribeJobQueuesInput struct { // A list of up to 100 queue names or full queue Amazon Resource Name (ARN) // entries. JobQueues []string // The maximum number of results returned by DescribeJobQueues in paginated // output. When this parameter is used, DescribeJobQueues only returns maxResults // results in a single page and a nextToken response element. The remaining // results of the initial request can be seen by sending another DescribeJobQueues // request with the returned nextToken value. This value can be between 1 and 100. // If this parameter isn't used, then DescribeJobQueues returns up to 100 results // and a nextToken value if applicable. MaxResults *int32 // The nextToken value returned from a previous paginated DescribeJobQueues // request where maxResults was used and the results exceeded the value of that // parameter. Pagination continues from the end of the previous results that // returned the nextToken value. This value is null when there are no more results // to return. // // Treat this token as an opaque identifier that's only used to retrieve the next // items in a list and not for other programmatic purposes. NextToken *string // contains filtered or unexported fields }
Contains the parameters for DescribeJobQueues .
type DescribeJobQueuesOutput ¶
type DescribeJobQueuesOutput struct { // The list of job queues. JobQueues []types.JobQueueDetail // The nextToken value to include in a future DescribeJobQueues request. When the // results of a DescribeJobQueues request exceed maxResults , this value can be // used to retrieve the next page of results. This value is null when there are no // more results to return. NextToken *string // Metadata pertaining to the operation's result. ResultMetadata middleware.Metadata // contains filtered or unexported fields }
type DescribeJobQueuesPaginator ¶
type DescribeJobQueuesPaginator struct {
// contains filtered or unexported fields
}
DescribeJobQueuesPaginator is a paginator for DescribeJobQueues
func NewDescribeJobQueuesPaginator ¶
func NewDescribeJobQueuesPaginator(client DescribeJobQueuesAPIClient, params *DescribeJobQueuesInput, optFns ...func(*DescribeJobQueuesPaginatorOptions)) *DescribeJobQueuesPaginator
NewDescribeJobQueuesPaginator returns a new DescribeJobQueuesPaginator
func (*DescribeJobQueuesPaginator) HasMorePages ¶
func (p *DescribeJobQueuesPaginator) HasMorePages() bool
HasMorePages returns a boolean indicating whether more pages are available
func (*DescribeJobQueuesPaginator) NextPage ¶
func (p *DescribeJobQueuesPaginator) NextPage(ctx context.Context, optFns ...func(*Options)) (*DescribeJobQueuesOutput, error)
NextPage retrieves the next DescribeJobQueues page.
type DescribeJobQueuesPaginatorOptions ¶
type DescribeJobQueuesPaginatorOptions struct { // The maximum number of results returned by DescribeJobQueues in paginated // output. When this parameter is used, DescribeJobQueues only returns maxResults // results in a single page and a nextToken response element. The remaining // results of the initial request can be seen by sending another DescribeJobQueues // request with the returned nextToken value. This value can be between 1 and 100. // If this parameter isn't used, then DescribeJobQueues returns up to 100 results // and a nextToken value if applicable. Limit int32 // Set to true if pagination should stop if the service returns a pagination token // that matches the most recent token provided to the service. StopOnDuplicateToken bool }
DescribeJobQueuesPaginatorOptions is the paginator options for DescribeJobQueues
type DescribeJobsInput ¶
type DescribeJobsInput struct { // A list of up to 100 job IDs. // // This member is required. Jobs []string // contains filtered or unexported fields }
Contains the parameters for DescribeJobs .
type DescribeJobsOutput ¶
type DescribeJobsOutput struct { // The list of jobs. Jobs []types.JobDetail // Metadata pertaining to the operation's result. ResultMetadata middleware.Metadata // contains filtered or unexported fields }
type DescribeSchedulingPoliciesInput ¶
type DescribeSchedulingPoliciesInput struct { // A list of up to 100 scheduling policy Amazon Resource Name (ARN) entries. // // This member is required. Arns []string // contains filtered or unexported fields }
Contains the parameters for DescribeSchedulingPolicies .
type DescribeSchedulingPoliciesOutput ¶
type DescribeSchedulingPoliciesOutput struct { // The list of scheduling policies. SchedulingPolicies []types.SchedulingPolicyDetail // Metadata pertaining to the operation's result. ResultMetadata middleware.Metadata // contains filtered or unexported fields }
type EndpointParameters ¶
type EndpointParameters struct { // The AWS region used to dispatch the request. // // Parameter is // required. // // AWS::Region Region *string // When true, use the dual-stack endpoint. If the configured endpoint does not // support dual-stack, dispatching the request MAY return an error. // // Defaults to // false if no value is provided. // // AWS::UseDualStack UseDualStack *bool // When true, send this request to the FIPS-compliant regional endpoint. If the // configured endpoint does not have a FIPS compliant endpoint, dispatching the // request will return an error. // // Defaults to false if no value is // provided. // // AWS::UseFIPS UseFIPS *bool // Override the endpoint used to send this request // // Parameter is // required. // // SDK::Endpoint Endpoint *string }
EndpointParameters provides the parameters that influence how endpoints are resolved.
func (EndpointParameters) ValidateRequired ¶
func (p EndpointParameters) ValidateRequired() error
ValidateRequired validates required parameters are set.
func (EndpointParameters) WithDefaults ¶
func (p EndpointParameters) WithDefaults() EndpointParameters
WithDefaults returns a shallow copy of EndpointParameterswith default values applied to members where applicable.
type EndpointResolver ¶
type EndpointResolver interface {
ResolveEndpoint(region string, options EndpointResolverOptions) (aws.Endpoint, error)
}
EndpointResolver interface for resolving service endpoints.
func EndpointResolverFromURL ¶
func EndpointResolverFromURL(url string, optFns ...func(*aws.Endpoint)) EndpointResolver
EndpointResolverFromURL returns an EndpointResolver configured using the provided endpoint url. By default, the resolved endpoint resolver uses the client region as signing region, and the endpoint source is set to EndpointSourceCustom.You can provide functional options to configure endpoint values for the resolved endpoint.
type EndpointResolverFunc ¶
type EndpointResolverFunc func(region string, options EndpointResolverOptions) (aws.Endpoint, error)
EndpointResolverFunc is a helper utility that wraps a function so it satisfies the EndpointResolver interface. This is useful when you want to add additional endpoint resolving logic, or stub out specific endpoints with custom values.
func (EndpointResolverFunc) ResolveEndpoint ¶
func (fn EndpointResolverFunc) ResolveEndpoint(region string, options EndpointResolverOptions) (endpoint aws.Endpoint, err error)
type EndpointResolverOptions ¶
type EndpointResolverOptions = internalendpoints.Options
EndpointResolverOptions is the service endpoint resolver options
type EndpointResolverV2 ¶
type EndpointResolverV2 interface { // ResolveEndpoint attempts to resolve the endpoint with the provided options, // returning the endpoint if found. Otherwise an error is returned. ResolveEndpoint(ctx context.Context, params EndpointParameters) ( smithyendpoints.Endpoint, error, ) }
EndpointResolverV2 provides the interface for resolving service endpoints.
func NewDefaultEndpointResolverV2 ¶
func NewDefaultEndpointResolverV2() EndpointResolverV2
type HTTPSignerV4 ¶
type ListJobsAPIClient ¶
type ListJobsAPIClient interface {
ListJobs(context.Context, *ListJobsInput, ...func(*Options)) (*ListJobsOutput, error)
}
ListJobsAPIClient is a client that implements the ListJobs operation.
type ListJobsInput ¶
type ListJobsInput struct { // The job ID for an array job. Specifying an array job ID with this parameter // lists all child jobs from within the specified array. ArrayJobId *string // The filter to apply to the query. Only one filter can be used at a time. When // the filter is used, jobStatus is ignored. The filter doesn't apply to child // jobs in an array or multi-node parallel (MNP) jobs. The results are sorted by // the createdAt field, with the most recent jobs being first. // // JOB_NAME The value of the filter is a case-insensitive match for the job name. // If the value ends with an asterisk (*), the filter matches any job name that // begins with the string before the '*'. This corresponds to the jobName value. // For example, test1 matches both Test1 and test1 , and test1* matches both test1 // and Test10 . When the JOB_NAME filter is used, the results are grouped by the // job name and version. // // JOB_DEFINITION The value for the filter is the name or Amazon Resource Name // (ARN) of the job definition. This corresponds to the jobDefinition value. The // value is case sensitive. When the value for the filter is the job definition // name, the results include all the jobs that used any revision of that job // definition name. If the value ends with an asterisk (*), the filter matches any // job definition name that begins with the string before the '*'. For example, jd1 // matches only jd1 , and jd1* matches both jd1 and jd1A . The version of the job // definition that's used doesn't affect the sort order. When the JOB_DEFINITION // filter is used and the ARN is used (which is in the form // arn:${Partition}:batch:${Region}:${Account}:job-definition/${JobDefinitionName}:${Revision} // ), the results include jobs that used the specified revision of the job // definition. Asterisk (*) isn't supported when the ARN is used. // // BEFORE_CREATED_AT The value for the filter is the time that's before the job // was created. This corresponds to the createdAt value. The value is a string // representation of the number of milliseconds since 00:00:00 UTC (midnight) on // January 1, 1970. // // AFTER_CREATED_AT The value for the filter is the time that's after the job was // created. This corresponds to the createdAt value. The value is a string // representation of the number of milliseconds since 00:00:00 UTC (midnight) on // January 1, 1970. Filters []types.KeyValuesPair // The name or full Amazon Resource Name (ARN) of the job queue used to list jobs. JobQueue *string // The job status used to filter jobs in the specified queue. If the filters // parameter is specified, the jobStatus parameter is ignored and jobs with any // status are returned. If you don't specify a status, only RUNNING jobs are // returned. JobStatus types.JobStatus // The maximum number of results returned by ListJobs in paginated output. When // this parameter is used, ListJobs only returns maxResults results in a single // page and a nextToken response element. The remaining results of the initial // request can be seen by sending another ListJobs request with the returned // nextToken value. This value can be between 1 and 100. If this parameter isn't // used, then ListJobs returns up to 100 results and a nextToken value if // applicable. MaxResults *int32 // The job ID for a multi-node parallel job. Specifying a multi-node parallel job // ID with this parameter lists all nodes that are associated with the specified // job. MultiNodeJobId *string // The nextToken value returned from a previous paginated ListJobs request where // maxResults was used and the results exceeded the value of that parameter. // Pagination continues from the end of the previous results that returned the // nextToken value. This value is null when there are no more results to return. // // Treat this token as an opaque identifier that's only used to retrieve the next // items in a list and not for other programmatic purposes. NextToken *string // contains filtered or unexported fields }
Contains the parameters for ListJobs .
type ListJobsOutput ¶
type ListJobsOutput struct { // A list of job summaries that match the request. // // This member is required. JobSummaryList []types.JobSummary // The nextToken value to include in a future ListJobs request. When the results // of a ListJobs request exceed maxResults , this value can be used to retrieve the // next page of results. This value is null when there are no more results to // return. NextToken *string // Metadata pertaining to the operation's result. ResultMetadata middleware.Metadata // contains filtered or unexported fields }
type ListJobsPaginator ¶
type ListJobsPaginator struct {
// contains filtered or unexported fields
}
ListJobsPaginator is a paginator for ListJobs
func NewListJobsPaginator ¶
func NewListJobsPaginator(client ListJobsAPIClient, params *ListJobsInput, optFns ...func(*ListJobsPaginatorOptions)) *ListJobsPaginator
NewListJobsPaginator returns a new ListJobsPaginator
func (*ListJobsPaginator) HasMorePages ¶
func (p *ListJobsPaginator) HasMorePages() bool
HasMorePages returns a boolean indicating whether more pages are available
func (*ListJobsPaginator) NextPage ¶
func (p *ListJobsPaginator) NextPage(ctx context.Context, optFns ...func(*Options)) (*ListJobsOutput, error)
NextPage retrieves the next ListJobs page.
type ListJobsPaginatorOptions ¶
type ListJobsPaginatorOptions struct { // The maximum number of results returned by ListJobs in paginated output. When // this parameter is used, ListJobs only returns maxResults results in a single // page and a nextToken response element. The remaining results of the initial // request can be seen by sending another ListJobs request with the returned // nextToken value. This value can be between 1 and 100. If this parameter isn't // used, then ListJobs returns up to 100 results and a nextToken value if // applicable. Limit int32 // Set to true if pagination should stop if the service returns a pagination token // that matches the most recent token provided to the service. StopOnDuplicateToken bool }
ListJobsPaginatorOptions is the paginator options for ListJobs
type ListSchedulingPoliciesAPIClient ¶
type ListSchedulingPoliciesAPIClient interface {
ListSchedulingPolicies(context.Context, *ListSchedulingPoliciesInput, ...func(*Options)) (*ListSchedulingPoliciesOutput, error)
}
ListSchedulingPoliciesAPIClient is a client that implements the ListSchedulingPolicies operation.
type ListSchedulingPoliciesInput ¶
type ListSchedulingPoliciesInput struct { // The maximum number of results that's returned by ListSchedulingPolicies in // paginated output. When this parameter is used, ListSchedulingPolicies only // returns maxResults results in a single page and a nextToken response element. // You can see the remaining results of the initial request by sending another // ListSchedulingPolicies request with the returned nextToken value. This value // can be between 1 and 100. If this parameter isn't used, ListSchedulingPolicies // returns up to 100 results and a nextToken value if applicable. MaxResults *int32 // The nextToken value that's returned from a previous paginated // ListSchedulingPolicies request where maxResults was used and the results // exceeded the value of that parameter. Pagination continues from the end of the // previous results that returned the nextToken value. This value is null when // there are no more results to return. // // Treat this token as an opaque identifier that's only used to retrieve the next // items in a list and not for other programmatic purposes. NextToken *string // contains filtered or unexported fields }
Contains the parameters for ListSchedulingPolicies .
type ListSchedulingPoliciesOutput ¶
type ListSchedulingPoliciesOutput struct { // The nextToken value to include in a future ListSchedulingPolicies request. When // the results of a ListSchedulingPolicies request exceed maxResults , this value // can be used to retrieve the next page of results. This value is null when there // are no more results to return. NextToken *string // A list of scheduling policies that match the request. SchedulingPolicies []types.SchedulingPolicyListingDetail // Metadata pertaining to the operation's result. ResultMetadata middleware.Metadata // contains filtered or unexported fields }
type ListSchedulingPoliciesPaginator ¶
type ListSchedulingPoliciesPaginator struct {
// contains filtered or unexported fields
}
ListSchedulingPoliciesPaginator is a paginator for ListSchedulingPolicies
func NewListSchedulingPoliciesPaginator ¶
func NewListSchedulingPoliciesPaginator(client ListSchedulingPoliciesAPIClient, params *ListSchedulingPoliciesInput, optFns ...func(*ListSchedulingPoliciesPaginatorOptions)) *ListSchedulingPoliciesPaginator
NewListSchedulingPoliciesPaginator returns a new ListSchedulingPoliciesPaginator
func (*ListSchedulingPoliciesPaginator) HasMorePages ¶
func (p *ListSchedulingPoliciesPaginator) HasMorePages() bool
HasMorePages returns a boolean indicating whether more pages are available
func (*ListSchedulingPoliciesPaginator) NextPage ¶
func (p *ListSchedulingPoliciesPaginator) NextPage(ctx context.Context, optFns ...func(*Options)) (*ListSchedulingPoliciesOutput, error)
NextPage retrieves the next ListSchedulingPolicies page.
type ListSchedulingPoliciesPaginatorOptions ¶
type ListSchedulingPoliciesPaginatorOptions struct { // The maximum number of results that's returned by ListSchedulingPolicies in // paginated output. When this parameter is used, ListSchedulingPolicies only // returns maxResults results in a single page and a nextToken response element. // You can see the remaining results of the initial request by sending another // ListSchedulingPolicies request with the returned nextToken value. This value // can be between 1 and 100. If this parameter isn't used, ListSchedulingPolicies // returns up to 100 results and a nextToken value if applicable. Limit int32 // Set to true if pagination should stop if the service returns a pagination token // that matches the most recent token provided to the service. StopOnDuplicateToken bool }
ListSchedulingPoliciesPaginatorOptions is the paginator options for ListSchedulingPolicies
type ListTagsForResourceInput ¶
type ListTagsForResourceInput struct { // The Amazon Resource Name (ARN) that identifies the resource that tags are // listed for. Batch resources that support tags are compute environments, jobs, // job definitions, job queues, and scheduling policies. ARNs for child jobs of // array and multi-node parallel (MNP) jobs aren't supported. // // This member is required. ResourceArn *string // contains filtered or unexported fields }
Contains the parameters for ListTagsForResource .
type ListTagsForResourceOutput ¶
type ListTagsForResourceOutput struct { // The tags for the resource. Tags map[string]string // Metadata pertaining to the operation's result. ResultMetadata middleware.Metadata // contains filtered or unexported fields }
type Options ¶
type Options struct { // Set of options to modify how an operation is invoked. These apply to all // operations invoked for this client. Use functional options on operation call to // modify this list for per operation behavior. APIOptions []func(*middleware.Stack) error // The optional application specific identifier appended to the User-Agent header. AppID string // This endpoint will be given as input to an EndpointResolverV2. It is used for // providing a custom base endpoint that is subject to modifications by the // processing EndpointResolverV2. BaseEndpoint *string // Configures the events that will be sent to the configured logger. ClientLogMode aws.ClientLogMode // The credentials object to use when signing requests. Credentials aws.CredentialsProvider // The configuration DefaultsMode that the SDK should use when constructing the // clients initial default settings. DefaultsMode aws.DefaultsMode // The endpoint options to be used when attempting to resolve an endpoint. EndpointOptions EndpointResolverOptions // The service endpoint resolver. // // Deprecated: Deprecated: EndpointResolver and WithEndpointResolver. Providing a // value for this field will likely prevent you from using any endpoint-related // service features released after the introduction of EndpointResolverV2 and // BaseEndpoint. // // To migrate an EndpointResolver implementation that uses a custom endpoint, set // the client option BaseEndpoint instead. EndpointResolver EndpointResolver // Resolves the endpoint used for a particular service operation. This should be // used over the deprecated EndpointResolver. EndpointResolverV2 EndpointResolverV2 // Signature Version 4 (SigV4) Signer HTTPSignerV4 HTTPSignerV4 // The logger writer interface to write logging messages to. Logger logging.Logger // The region to send requests to. (Required) Region string // RetryMaxAttempts specifies the maximum number attempts an API client will call // an operation that fails with a retryable error. A value of 0 is ignored, and // will not be used to configure the API client created default retryer, or modify // per operation call's retry max attempts. // // If specified in an operation call's functional options with a value that is // different than the constructed client's Options, the Client's Retryer will be // wrapped to use the operation's specific RetryMaxAttempts value. RetryMaxAttempts int // RetryMode specifies the retry mode the API client will be created with, if // Retryer option is not also specified. // // When creating a new API Clients this member will only be used if the Retryer // Options member is nil. This value will be ignored if Retryer is not nil. // // Currently does not support per operation call overrides, may in the future. RetryMode aws.RetryMode // Retryer guides how HTTP requests should be retried in case of recoverable // failures. When nil the API client will use a default retryer. The kind of // default retry created by the API client can be changed with the RetryMode // option. Retryer aws.Retryer // The RuntimeEnvironment configuration, only populated if the DefaultsMode is set // to DefaultsModeAuto and is initialized using config.LoadDefaultConfig . You // should not populate this structure programmatically, or rely on the values here // within your applications. RuntimeEnvironment aws.RuntimeEnvironment // The HTTP client to invoke API calls with. Defaults to client's default HTTP // implementation if nil. HTTPClient HTTPClient // The auth scheme resolver which determines how to authenticate for each // operation. AuthSchemeResolver AuthSchemeResolver // The list of auth schemes supported by the client. AuthSchemes []smithyhttp.AuthScheme // contains filtered or unexported fields }
func (Options) Copy ¶
Copy creates a clone where the APIOptions list is deep copied.
func (Options) GetIdentityResolver ¶
func (o Options) GetIdentityResolver(schemeID string) smithyauth.IdentityResolver
type RegisterJobDefinitionInput ¶
type RegisterJobDefinitionInput struct { // The name of the job definition to register. It can be up to 128 letters long. // It can contain uppercase and lowercase letters, numbers, hyphens (-), and // underscores (_). // // This member is required. JobDefinitionName *string // The type of job definition. For more information about multi-node parallel // jobs, see [Creating a multi-node parallel job definition]in the Batch User Guide. // // - If the value is container , then one of the following is required: // containerProperties , ecsProperties , or eksProperties . // // - If the value is multinode , then nodeProperties is required. // // If the job is run on Fargate resources, then multinode isn't supported. // // [Creating a multi-node parallel job definition]: https://docs.aws.amazon.com/batch/latest/userguide/multi-node-job-def.html // // This member is required. Type types.JobDefinitionType // An object with properties specific to Amazon ECS-based single-node // container-based jobs. If the job definition's type parameter is container , then // you must specify either containerProperties or nodeProperties . This must not be // specified for Amazon EKS-based job definitions. // // If the job runs on Fargate resources, then you must not specify nodeProperties ; // use only containerProperties . ContainerProperties *types.ContainerProperties // An object with properties that are specific to Amazon ECS-based jobs. This must // not be specified for Amazon EKS-based job definitions. EcsProperties *types.EcsProperties // An object with properties that are specific to Amazon EKS-based jobs. This must // not be specified for Amazon ECS based job definitions. EksProperties *types.EksProperties // An object with properties specific to multi-node parallel jobs. If you specify // node properties for a job, it becomes a multi-node parallel job. For more // information, see [Multi-node Parallel Jobs]in the Batch User Guide. // // If the job runs on Fargate resources, then you must not specify nodeProperties ; // use containerProperties instead. // // If the job runs on Amazon EKS resources, then you must not specify // nodeProperties . // // [Multi-node Parallel Jobs]: https://docs.aws.amazon.com/batch/latest/userguide/multi-node-parallel-jobs.html NodeProperties *types.NodeProperties // Default parameter substitution placeholders to set in the job definition. // Parameters are specified as a key-value pair mapping. Parameters in a SubmitJob // request override any corresponding parameter defaults from the job definition. Parameters map[string]string // The platform capabilities required by the job definition. If no value is // specified, it defaults to EC2 . To run the job on Fargate resources, specify // FARGATE . // // If the job runs on Amazon EKS resources, then you must not specify // platformCapabilities . PlatformCapabilities []types.PlatformCapability // Specifies whether to propagate the tags from the job or job definition to the // corresponding Amazon ECS task. If no value is specified, the tags are not // propagated. Tags can only be propagated to the tasks during task creation. For // tags with the same name, job tags are given priority over job definitions tags. // If the total number of combined tags from the job and job definition is over 50, // the job is moved to the FAILED state. // // If the job runs on Amazon EKS resources, then you must not specify propagateTags // . PropagateTags *bool // The retry strategy to use for failed jobs that are submitted with this job // definition. Any retry strategy that's specified during a SubmitJoboperation overrides // the retry strategy defined here. If a job is terminated due to a timeout, it // isn't retried. RetryStrategy *types.RetryStrategy // The scheduling priority for jobs that are submitted with this job definition. // This only affects jobs in job queues with a fair share policy. Jobs with a // higher scheduling priority are scheduled before jobs with a lower scheduling // priority. // // The minimum supported value is 0 and the maximum supported value is 9999. SchedulingPriority *int32 // The tags that you apply to the job definition to help you categorize and // organize your resources. Each tag consists of a key and an optional value. For // more information, see [Tagging Amazon Web Services Resources]in Batch User Guide. // // [Tagging Amazon Web Services Resources]: https://docs.aws.amazon.com/batch/latest/userguide/using-tags.html Tags map[string]string // The timeout configuration for jobs that are submitted with this job definition, // after which Batch terminates your jobs if they have not finished. If a job is // terminated due to a timeout, it isn't retried. The minimum value for the timeout // is 60 seconds. Any timeout configuration that's specified during a SubmitJoboperation // overrides the timeout configuration defined here. For more information, see [Job Timeouts]in // the Batch User Guide. // // [Job Timeouts]: https://docs.aws.amazon.com/batch/latest/userguide/job_timeouts.html Timeout *types.JobTimeout // contains filtered or unexported fields }
Contains the parameters for RegisterJobDefinition .
type RegisterJobDefinitionOutput ¶
type RegisterJobDefinitionOutput struct { // The Amazon Resource Name (ARN) of the job definition. // // This member is required. JobDefinitionArn *string // The name of the job definition. // // This member is required. JobDefinitionName *string // The revision of the job definition. // // This member is required. Revision *int32 // Metadata pertaining to the operation's result. ResultMetadata middleware.Metadata // contains filtered or unexported fields }
type ResolveEndpoint ¶
type ResolveEndpoint struct { Resolver EndpointResolver Options EndpointResolverOptions }
func (*ResolveEndpoint) HandleSerialize ¶
func (m *ResolveEndpoint) HandleSerialize(ctx context.Context, in middleware.SerializeInput, next middleware.SerializeHandler) ( out middleware.SerializeOutput, metadata middleware.Metadata, err error, )
func (*ResolveEndpoint) ID ¶
func (*ResolveEndpoint) ID() string
type SubmitJobInput ¶
type SubmitJobInput struct { // The job definition used by this job. This value can be one of definition-name , // definition-name:revision , or the Amazon Resource Name (ARN) for the job // definition, with or without the revision ( // arn:aws:batch:region:account:job-definition/definition-name:revision , or // arn:aws:batch:region:account:job-definition/definition-name ). // // If the revision is not specified, then the latest active revision is used. // // This member is required. JobDefinition *string // The name of the job. It can be up to 128 letters long. The first character must // be alphanumeric, can contain uppercase and lowercase letters, numbers, hyphens // (-), and underscores (_). // // This member is required. JobName *string // The job queue where the job is submitted. You can specify either the name or // the Amazon Resource Name (ARN) of the queue. // // This member is required. JobQueue *string // The array properties for the submitted job, such as the size of the array. The // array size can be between 2 and 10,000. If you specify array properties for a // job, it becomes an array job. For more information, see [Array Jobs]in the Batch User Guide. // // [Array Jobs]: https://docs.aws.amazon.com/batch/latest/userguide/array_jobs.html ArrayProperties *types.ArrayProperties // An object with properties that override the defaults for the job definition // that specify the name of a container in the specified job definition and the // overrides it should receive. You can override the default command for a // container, which is specified in the job definition or the Docker image, with a // command override. You can also override existing environment variables on a // container or add new environment variables to it with an environment override. ContainerOverrides *types.ContainerOverrides // A list of dependencies for the job. A job can depend upon a maximum of 20 jobs. // You can specify a SEQUENTIAL type dependency without specifying a job ID for // array jobs so that each child array job completes sequentially, starting at // index 0. You can also specify an N_TO_N type dependency with a job ID for array // jobs. In that case, each index child of this job must wait for the corresponding // index child of each dependency to complete before it can begin. DependsOn []types.JobDependency // An object, with properties that override defaults for the job definition, can // only be specified for jobs that are run on Amazon ECS resources. EcsPropertiesOverride *types.EcsPropertiesOverride // An object, with properties that override defaults for the job definition, can // only be specified for jobs that are run on Amazon EKS resources. EksPropertiesOverride *types.EksPropertiesOverride // A list of node overrides in JSON format that specify the node range to target // and the container overrides for that node range. // // This parameter isn't applicable to jobs that are running on Fargate resources; // use containerOverrides instead. NodeOverrides *types.NodeOverrides // Additional parameters passed to the job that replace parameter substitution // placeholders that are set in the job definition. Parameters are specified as a // key and value pair mapping. Parameters in a SubmitJob request override any // corresponding parameter defaults from the job definition. Parameters map[string]string // Specifies whether to propagate the tags from the job or job definition to the // corresponding Amazon ECS task. If no value is specified, the tags aren't // propagated. Tags can only be propagated to the tasks during task creation. For // tags with the same name, job tags are given priority over job definitions tags. // If the total number of combined tags from the job and job definition is over 50, // the job is moved to the FAILED state. When specified, this overrides the tag // propagation setting in the job definition. PropagateTags *bool // The retry strategy to use for failed jobs from this SubmitJob operation. When a retry // strategy is specified here, it overrides the retry strategy defined in the job // definition. RetryStrategy *types.RetryStrategy // The scheduling priority for the job. This only affects jobs in job queues with // a fair share policy. Jobs with a higher scheduling priority are scheduled before // jobs with a lower scheduling priority. This overrides any scheduling priority in // the job definition and works only within a single share identifier. // // The minimum supported value is 0 and the maximum supported value is 9999. SchedulingPriorityOverride *int32 // doesn't have a scheduling policy. If the job queue has a scheduling policy, then // this parameter must be specified. // // This string is limited to 255 alphanumeric characters, and can be followed by // an asterisk (*). ShareIdentifier *string // The tags that you apply to the job request to help you categorize and organize // your resources. Each tag consists of a key and an optional value. For more // information, see [Tagging Amazon Web Services Resources]in Amazon Web Services General Reference. // // [Tagging Amazon Web Services Resources]: https://docs.aws.amazon.com/general/latest/gr/aws_tagging.html Tags map[string]string // The timeout configuration for this SubmitJob operation. You can specify a timeout // duration after which Batch terminates your jobs if they haven't finished. If a // job is terminated due to a timeout, it isn't retried. The minimum value for the // timeout is 60 seconds. This configuration overrides any timeout configuration // specified in the job definition. For array jobs, child jobs have the same // timeout configuration as the parent job. For more information, see [Job Timeouts]in the // Amazon Elastic Container Service Developer Guide. // // [Job Timeouts]: https://docs.aws.amazon.com/AmazonECS/latest/developerguide/job_timeouts.html Timeout *types.JobTimeout // contains filtered or unexported fields }
Contains the parameters for SubmitJob .
type SubmitJobOutput ¶
type SubmitJobOutput struct { // The unique identifier for the job. // // This member is required. JobId *string // The name of the job. // // This member is required. JobName *string // The Amazon Resource Name (ARN) for the job. JobArn *string // Metadata pertaining to the operation's result. ResultMetadata middleware.Metadata // contains filtered or unexported fields }
type TagResourceInput ¶
type TagResourceInput struct { // The Amazon Resource Name (ARN) of the resource that tags are added to. Batch // resources that support tags are compute environments, jobs, job definitions, job // queues, and scheduling policies. ARNs for child jobs of array and multi-node // parallel (MNP) jobs aren't supported. // // This member is required. ResourceArn *string // The tags that you apply to the resource to help you categorize and organize // your resources. Each tag consists of a key and an optional value. For more // information, see [Tagging Amazon Web Services Resources]in Amazon Web Services General Reference. // // [Tagging Amazon Web Services Resources]: https://docs.aws.amazon.com/general/latest/gr/aws_tagging.html // // This member is required. Tags map[string]string // contains filtered or unexported fields }
Contains the parameters for TagResource .
type TagResourceOutput ¶
type TagResourceOutput struct { // Metadata pertaining to the operation's result. ResultMetadata middleware.Metadata // contains filtered or unexported fields }
type TerminateJobInput ¶
type TerminateJobInput struct { // The Batch job ID of the job to terminate. // // This member is required. JobId *string // A message to attach to the job that explains the reason for canceling it. This // message is returned by future DescribeJobsoperations on the job. This message is also // recorded in the Batch activity logs. // // This member is required. Reason *string // contains filtered or unexported fields }
Contains the parameters for TerminateJob .
type TerminateJobOutput ¶
type TerminateJobOutput struct { // Metadata pertaining to the operation's result. ResultMetadata middleware.Metadata // contains filtered or unexported fields }
type UntagResourceInput ¶
type UntagResourceInput struct { // The Amazon Resource Name (ARN) of the resource from which to delete tags. Batch // resources that support tags are compute environments, jobs, job definitions, job // queues, and scheduling policies. ARNs for child jobs of array and multi-node // parallel (MNP) jobs aren't supported. // // This member is required. ResourceArn *string // The keys of the tags to be removed. // // This member is required. TagKeys []string // contains filtered or unexported fields }
Contains the parameters for UntagResource .
type UntagResourceOutput ¶
type UntagResourceOutput struct { // Metadata pertaining to the operation's result. ResultMetadata middleware.Metadata // contains filtered or unexported fields }
type UpdateComputeEnvironmentInput ¶
type UpdateComputeEnvironmentInput struct { // The name or full Amazon Resource Name (ARN) of the compute environment to // update. // // This member is required. ComputeEnvironment *string // Details of the compute resources managed by the compute environment. Required // for a managed compute environment. For more information, see [Compute Environments]in the Batch User // Guide. // // [Compute Environments]: https://docs.aws.amazon.com/batch/latest/userguide/compute_environments.html ComputeResources *types.ComputeResourceUpdate // The full Amazon Resource Name (ARN) of the IAM role that allows Batch to make // calls to other Amazon Web Services services on your behalf. For more // information, see [Batch service IAM role]in the Batch User Guide. // // If the compute environment has a service-linked role, it can't be changed to // use a regular IAM role. Likewise, if the compute environment has a regular IAM // role, it can't be changed to use a service-linked role. To update the parameters // for the compute environment that require an infrastructure update to change, the // AWSServiceRoleForBatch service-linked role must be used. For more information, // see [Updating compute environments]in the Batch User Guide. // // If your specified role has a path other than / , then you must either specify // the full role ARN (recommended) or prefix the role name with the path. // // Depending on how you created your Batch service role, its ARN might contain the // service-role path prefix. When you only specify the name of the service role, // Batch assumes that your ARN doesn't use the service-role path prefix. Because // of this, we recommend that you specify the full ARN of your service role when // you create compute environments. // // [Updating compute environments]: https://docs.aws.amazon.com/batch/latest/userguide/updating-compute-environments.html // [Batch service IAM role]: https://docs.aws.amazon.com/batch/latest/userguide/service_IAM_role.html ServiceRole *string // The state of the compute environment. Compute environments in the ENABLED state // can accept jobs from a queue and scale in or out automatically based on the // workload demand of its associated queues. // // If the state is ENABLED , then the Batch scheduler can attempt to place jobs // from an associated job queue on the compute resources within the environment. If // the compute environment is managed, then it can scale its instances out or in // automatically, based on the job queue demand. // // If the state is DISABLED , then the Batch scheduler doesn't attempt to place // jobs within the environment. Jobs in a STARTING or RUNNING state continue to // progress normally. Managed compute environments in the DISABLED state don't // scale out. // // Compute environments in a DISABLED state may continue to incur billing charges. // To prevent additional charges, turn off and then delete the compute environment. // For more information, see [State]in the Batch User Guide. // // When an instance is idle, the instance scales down to the minvCpus value. // However, the instance size doesn't change. For example, consider a c5.8xlarge // instance with a minvCpus value of 4 and a desiredvCpus value of 36 . This // instance doesn't scale down to a c5.large instance. // // [State]: https://docs.aws.amazon.com/batch/latest/userguide/compute_environment_parameters.html#compute_environment_state State types.CEState // The maximum number of vCPUs expected to be used for an unmanaged compute // environment. Don't specify this parameter for a managed compute environment. // This parameter is only used for fair share scheduling to reserve vCPU capacity // for new share identifiers. If this parameter isn't provided for a fair share job // queue, no vCPU capacity is reserved. UnmanagedvCpus *int32 // Specifies the updated infrastructure update policy for the compute environment. // For more information about infrastructure updates, see [Updating compute environments]in the Batch User Guide. // // [Updating compute environments]: https://docs.aws.amazon.com/batch/latest/userguide/updating-compute-environments.html UpdatePolicy *types.UpdatePolicy // contains filtered or unexported fields }
Contains the parameters for UpdateComputeEnvironment .
type UpdateComputeEnvironmentOutput ¶
type UpdateComputeEnvironmentOutput struct { // The Amazon Resource Name (ARN) of the compute environment. ComputeEnvironmentArn *string // The name of the compute environment. It can be up to 128 characters long. It // can contain uppercase and lowercase letters, numbers, hyphens (-), and // underscores (_). ComputeEnvironmentName *string // Metadata pertaining to the operation's result. ResultMetadata middleware.Metadata // contains filtered or unexported fields }
type UpdateJobQueueInput ¶
type UpdateJobQueueInput struct { // The name or the Amazon Resource Name (ARN) of the job queue. // // This member is required. JobQueue *string // Details the set of compute environments mapped to a job queue and their order // relative to each other. This is one of the parameters used by the job scheduler // to determine which compute environment runs a given job. Compute environments // must be in the VALID state before you can associate them with a job queue. All // of the compute environments must be either EC2 ( EC2 or SPOT ) or Fargate ( // FARGATE or FARGATE_SPOT ). EC2 and Fargate compute environments can't be mixed. // // All compute environments that are associated with a job queue must share the // same architecture. Batch doesn't support mixing compute environment architecture // types in a single job queue. ComputeEnvironmentOrder []types.ComputeEnvironmentOrder // The set of actions that Batch perform on jobs that remain at the head of the // job queue in the specified state longer than specified times. Batch will perform // each action after maxTimeSeconds has passed. JobStateTimeLimitActions []types.JobStateTimeLimitAction // The priority of the job queue. Job queues with a higher priority (or a higher // integer value for the priority parameter) are evaluated first when associated // with the same compute environment. Priority is determined in descending order. // For example, a job queue with a priority value of 10 is given scheduling // preference over a job queue with a priority value of 1 . All of the compute // environments must be either EC2 ( EC2 or SPOT ) or Fargate ( FARGATE or // FARGATE_SPOT ). EC2 and Fargate compute environments can't be mixed. Priority *int32 // Amazon Resource Name (ARN) of the fair share scheduling policy. Once a job // queue is created, the fair share scheduling policy can be replaced but not // removed. The format is // aws:Partition:batch:Region:Account:scheduling-policy/Name . For example, // aws:aws:batch:us-west-2:123456789012:scheduling-policy/MySchedulingPolicy . SchedulingPolicyArn *string // Describes the queue's ability to accept new jobs. If the job queue state is // ENABLED , it can accept jobs. If the job queue state is DISABLED , new jobs // can't be added to the queue, but jobs already in the queue can finish. State types.JQState // contains filtered or unexported fields }
Contains the parameters for UpdateJobQueue .
type UpdateJobQueueOutput ¶
type UpdateJobQueueOutput struct { // The Amazon Resource Name (ARN) of the job queue. JobQueueArn *string // The name of the job queue. JobQueueName *string // Metadata pertaining to the operation's result. ResultMetadata middleware.Metadata // contains filtered or unexported fields }
type UpdateSchedulingPolicyInput ¶
type UpdateSchedulingPolicyInput struct { // The Amazon Resource Name (ARN) of the scheduling policy to update. // // This member is required. Arn *string FairsharePolicy *types.FairsharePolicy // contains filtered or unexported fields }
Contains the parameters for UpdateSchedulingPolicy .
type UpdateSchedulingPolicyOutput ¶
type UpdateSchedulingPolicyOutput struct { // Metadata pertaining to the operation's result. ResultMetadata middleware.Metadata // contains filtered or unexported fields }
Source Files ¶
- api_client.go
- api_op_CancelJob.go
- api_op_CreateComputeEnvironment.go
- api_op_CreateJobQueue.go
- api_op_CreateSchedulingPolicy.go
- api_op_DeleteComputeEnvironment.go
- api_op_DeleteJobQueue.go
- api_op_DeleteSchedulingPolicy.go
- api_op_DeregisterJobDefinition.go
- api_op_DescribeComputeEnvironments.go
- api_op_DescribeJobDefinitions.go
- api_op_DescribeJobQueues.go
- api_op_DescribeJobs.go
- api_op_DescribeSchedulingPolicies.go
- api_op_ListJobs.go
- api_op_ListSchedulingPolicies.go
- api_op_ListTagsForResource.go
- api_op_RegisterJobDefinition.go
- api_op_SubmitJob.go
- api_op_TagResource.go
- api_op_TerminateJob.go
- api_op_UntagResource.go
- api_op_UpdateComputeEnvironment.go
- api_op_UpdateJobQueue.go
- api_op_UpdateSchedulingPolicy.go
- auth.go
- deserializers.go
- doc.go
- endpoints.go
- go_module_metadata.go
- options.go
- serializers.go
- validators.go