We recommend new projects start with resources from the AWS provider.
aws-native.ecs.Service
Explore with Pulumi AI
We recommend new projects start with resources from the AWS provider.
The AWS::ECS::Service
resource creates an Amazon Elastic Container Service (Amazon ECS) service that runs and maintains the requested number of tasks and associated load balancers.
The stack update fails if you change any properties that require replacement and at least one ECS Service Connect ServiceConnectConfiguration
property the is configured. This is because AWS CloudFormation creates the replacement service first, but each ServiceConnectService
must have a name that is unique in the namespace.
Starting April 15, 2023, AWS; will not onboard new customers to Amazon Elastic Inference (EI), and will help current customers migrate their workloads to options that offer better price and performance. After April 15, 2023, new customers will not be able to launch instances with Amazon EI accelerators in Amazon SageMaker, ECS, or EC2. However, customers who have used Amazon EI at least once during the past 30-day period are considered current customers and will be able to continue using the service.
Create Service Resource
Resources are created with functions called constructors. To learn more about declaring and configuring resources, see Resources.
Constructor syntax
new Service(name: string, args?: ServiceArgs, opts?: CustomResourceOptions);
@overload
def Service(resource_name: str,
args: Optional[ServiceArgs] = None,
opts: Optional[ResourceOptions] = None)
@overload
def Service(resource_name: str,
opts: Optional[ResourceOptions] = None,
availability_zone_rebalancing: Optional[ServiceAvailabilityZoneRebalancing] = None,
capacity_provider_strategy: Optional[Sequence[ServiceCapacityProviderStrategyItemArgs]] = None,
cluster: Optional[str] = None,
deployment_configuration: Optional[ServiceDeploymentConfigurationArgs] = None,
deployment_controller: Optional[ServiceDeploymentControllerArgs] = None,
desired_count: Optional[int] = None,
enable_ecs_managed_tags: Optional[bool] = None,
enable_execute_command: Optional[bool] = None,
health_check_grace_period_seconds: Optional[int] = None,
launch_type: Optional[ServiceLaunchType] = None,
load_balancers: Optional[Sequence[ServiceLoadBalancerArgs]] = None,
network_configuration: Optional[ServiceNetworkConfigurationArgs] = None,
placement_constraints: Optional[Sequence[ServicePlacementConstraintArgs]] = None,
placement_strategies: Optional[Sequence[ServicePlacementStrategyArgs]] = None,
platform_version: Optional[str] = None,
propagate_tags: Optional[ServicePropagateTags] = None,
role: Optional[str] = None,
scheduling_strategy: Optional[ServiceSchedulingStrategy] = None,
service_connect_configuration: Optional[ServiceConnectConfigurationArgs] = None,
service_name: Optional[str] = None,
service_registries: Optional[Sequence[ServiceRegistryArgs]] = None,
tags: Optional[Sequence[_root_inputs.TagArgs]] = None,
task_definition: Optional[str] = None,
volume_configurations: Optional[Sequence[ServiceVolumeConfigurationArgs]] = None,
vpc_lattice_configurations: Optional[Sequence[ServiceVpcLatticeConfigurationArgs]] = None)
func NewService(ctx *Context, name string, args *ServiceArgs, opts ...ResourceOption) (*Service, error)
public Service(string name, ServiceArgs? args = null, CustomResourceOptions? opts = null)
public Service(String name, ServiceArgs args)
public Service(String name, ServiceArgs args, CustomResourceOptions options)
type: aws-native:ecs:Service
properties: # The arguments to resource properties.
options: # Bag of options to control resource's behavior.
Parameters
- name string
- The unique name of the resource.
- args ServiceArgs
- The arguments to resource properties.
- opts CustomResourceOptions
- Bag of options to control resource's behavior.
- resource_name str
- The unique name of the resource.
- args ServiceArgs
- The arguments to resource properties.
- opts ResourceOptions
- Bag of options to control resource's behavior.
- ctx Context
- Context object for the current deployment.
- name string
- The unique name of the resource.
- args ServiceArgs
- The arguments to resource properties.
- opts ResourceOption
- Bag of options to control resource's behavior.
- name string
- The unique name of the resource.
- args ServiceArgs
- The arguments to resource properties.
- opts CustomResourceOptions
- Bag of options to control resource's behavior.
- name String
- The unique name of the resource.
- args ServiceArgs
- The arguments to resource properties.
- options CustomResourceOptions
- Bag of options to control resource's behavior.
Service Resource Properties
To learn more about resource properties and how to use them, see Inputs and Outputs in the Architecture and Concepts docs.
Inputs
In Python, inputs that are objects can be passed either as argument classes or as dictionary literals.
The Service resource accepts the following input properties:
- Availability
Zone Pulumi.Rebalancing Aws Native. Ecs. Service Availability Zone Rebalancing - Capacity
Provider List<Pulumi.Strategy Aws Native. Ecs. Inputs. Service Capacity Provider Strategy Item> - The capacity provider strategy to use for the service.
If a
capacityProviderStrategy
is specified, thelaunchType
parameter must be omitted. If nocapacityProviderStrategy
orlaunchType
is specified, thedefaultCapacityProviderStrategy
for the cluster is used. A capacity provider strategy may contain a maximum of 6 capacity providers. - Cluster string
- The short name or full Amazon Resource Name (ARN) of the cluster that you run your service on. If you do not specify a cluster, the default cluster is assumed.
- Deployment
Configuration Pulumi.Aws Native. Ecs. Inputs. Service Deployment Configuration - Optional deployment parameters that control how many tasks run during the deployment and the ordering of stopping and starting tasks.
- Deployment
Controller Pulumi.Aws Native. Ecs. Inputs. Service Deployment Controller - The deployment controller to use for the service. If no deployment controller is specified, the default value of
ECS
is used. - Desired
Count int - The number of instantiations of the specified task definition to place and keep running in your service.
For new services, if a desired count is not specified, a default value of
1
is used. When using theDAEMON
scheduling strategy, the desired count is not required. For existing services, if a desired count is not specified, it is omitted from the operation. - bool
- Specifies whether to turn on Amazon ECS managed tags for the tasks within the service. For more information, see Tagging your Amazon ECS resources in the Amazon Elastic Container Service Developer Guide.
When you use Amazon ECS managed tags, you need to set the
propagateTags
request parameter. - Enable
Execute boolCommand - Determines whether the execute command functionality is turned on for the service. If
true
, the execute command functionality is turned on for all containers in tasks as part of the service. - Health
Check intGrace Period Seconds - The period of time, in seconds, that the Amazon ECS service scheduler ignores unhealthy Elastic Load Balancing target health checks after a task has first started. This is only used when your service is configured to use a load balancer. If your service has a load balancer defined and you don't specify a health check grace period value, the default value of
0
is used. If you do not use an Elastic Load Balancing, we recommend that you use thestartPeriod
in the task definition health check parameters. For more information, see Health check. If your service's tasks take a while to start and respond to Elastic Load Balancing health checks, you can specify a health check grace period of up to 2,147,483,647 seconds (about 69 years). During that time, the Amazon ECS service scheduler ignores health check status. This grace period can prevent the service scheduler from marking tasks as unhealthy and stopping them before they have time to come up. - Launch
Type Pulumi.Aws Native. Ecs. Service Launch Type - The launch type on which to run your service. For more information, see Amazon ECS Launch Types in the Amazon Elastic Container Service Developer Guide.
- Load
Balancers List<Pulumi.Aws Native. Ecs. Inputs. Service Load Balancer> - A list of load balancer objects to associate with the service. If you specify the
Role
property,LoadBalancers
must be specified as well. For information about the number of load balancers that you can specify per service, see Service Load Balancing in the Amazon Elastic Container Service Developer Guide. - Network
Configuration Pulumi.Aws Native. Ecs. Inputs. Service Network Configuration - The network configuration for the service. This parameter is required for task definitions that use the
awsvpc
network mode to receive their own elastic network interface, and it is not supported for other network modes. For more information, see Task Networking in the Amazon Elastic Container Service Developer Guide. - Placement
Constraints List<Pulumi.Aws Native. Ecs. Inputs. Service Placement Constraint> - An array of placement constraint objects to use for tasks in your service. You can specify a maximum of 10 constraints for each task. This limit includes constraints in the task definition and those specified at runtime.
- Placement
Strategies List<Pulumi.Aws Native. Ecs. Inputs. Service Placement Strategy> - The placement strategy objects to use for tasks in your service. You can specify a maximum of 5 strategy rules for each service.
- Platform
Version string - The platform version that your tasks in the service are running on. A platform version is specified only for tasks using the Fargate launch type. If one isn't specified, the
LATEST
platform version is used. For more information, see platform versions in the Amazon Elastic Container Service Developer Guide. - Pulumi.
Aws Native. Ecs. Service Propagate Tags - Specifies whether to propagate the tags from the task definition to the task. If no value is specified, the tags aren't propagated. Tags can only be propagated to the task during task creation. To add tags to a task after task creation, use the TagResource API action.
You must set this to a value other than
NONE
when you use Cost Explorer. For more information, see Amazon ECS usage reports in the Amazon Elastic Container Service Developer Guide. The default isNONE
. - Role string
- The name or full Amazon Resource Name (ARN) of the IAM role that allows Amazon ECS to make calls to your load balancer on your behalf. This parameter is only permitted if you are using a load balancer with your service and your task definition doesn't use the
awsvpc
network mode. If you specify therole
parameter, you must also specify a load balancer object with theloadBalancers
parameter. If your account has already created the Amazon ECS service-linked role, that role is used for your service unless you specify a role here. The service-linked role is required if your task definition uses theawsvpc
network mode or if the service is configured to use service discovery, an external deployment controller, multiple target groups, or Elastic Inference accelerators in which case you don't specify a role here. For more information, see Using service-linked roles for Amazon ECS in the Amazon Elastic Container Service Developer Guide. If your specified role has a path other than/
, then you must either specify the full role ARN (this is recommended) or prefix the role name with the path. For example, if a role with the namebar
has a path of/foo/
then you would specify/foo/bar
as the role name. For more information, see Friendly names and paths in the IAM User Guide. - Scheduling
Strategy Pulumi.Aws Native. Ecs. Service Scheduling Strategy - The scheduling strategy to use for the service. For more information, see Services.
There are two service scheduler strategies available:
REPLICA
-The replica scheduling strategy places and maintains the desired number of tasks across your cluster. By default, the service scheduler spreads tasks across Availability Zones. You can use task placement strategies and constraints to customize task placement decisions. This scheduler strategy is required if the service uses theCODE_DEPLOY
orEXTERNAL
deployment controller types.DAEMON
-The daemon scheduling strategy deploys exactly one task on each active container instance that meets all of the task placement constraints that you specify in your cluster. The service scheduler also evaluates the task placement constraints for running tasks and will stop tasks that don't meet the placement constraints. When you're using this strategy, you don't need to specify a desired number of tasks, a task placement strategy, or use Service Auto Scaling policies. Tasks using the Fargate launch type or theCODE_DEPLOY
orEXTERNAL
deployment controller types don't support theDAEMON
scheduling strategy.
- Service
Connect Pulumi.Configuration Aws Native. Ecs. Inputs. Service Connect Configuration - The configuration for this service to discover and connect to services, and be discovered by, and connected from, other services within a namespace. Tasks that run in a namespace can use short names to connect to services in the namespace. Tasks can connect to services across all of the clusters in the namespace. Tasks connect through a managed proxy container that collects logs and metrics for increased visibility. Only the tasks that Amazon ECS services create are supported with Service Connect. For more information, see Service Connect in the Amazon Elastic Container Service Developer Guide.
- Service
Name string - The name of your service. Up to 255 letters (uppercase and lowercase), numbers, underscores, and hyphens are allowed. Service names must be unique within a cluster, but you can have similarly named services in multiple clusters within a Region or across multiple Regions.
The stack update fails if you change any properties that require replacement and the
ServiceName
is configured. This is because AWS CloudFormation creates the replacement service first, but eachServiceName
must be unique in the cluster. - Service
Registries List<Pulumi.Aws Native. Ecs. Inputs. Service Registry> - The details of the service discovery registry to associate with this service. For more information, see Service discovery. Each service may be associated with one service registry. Multiple service registries for each service isn't supported.
- List<Pulumi.
Aws Native. Inputs. Tag> - The metadata that you apply to the service to help you categorize and organize them. Each tag consists of a key and an optional value, both of which you define. When a service is deleted, the tags are deleted as well.
The following basic restrictions apply to tags:
- Maximum number of tags per resource - 50
- For each resource, each tag key must be unique, and each tag key can have only one value.
- Maximum key length - 128 Unicode characters in UTF-8
- Maximum value length - 256 Unicode characters in UTF-8
- If your tagging schema is used across multiple services and resources, remember that other services may have restrictions on allowed characters. Generally allowed characters are: letters, numbers, and spaces representable in UTF-8, and the following characters: + - = . _ : / @.
- Tag keys and values are case-sensitive.
- Do not use
aws:
,AWS:
, or any upper or lowercase combination of such as a prefix for either keys or values as it is reserved for AWS use. You cannot edit or delete tag keys or values with this prefix. Tags with this prefix do not count against your tags per resource limit.
- Task
Definition string - The
family
andrevision
(family:revision
) or full ARN of the task definition to run in your service. If arevision
isn't specified, the latestACTIVE
revision is used. A task definition must be specified if the service uses either theECS
orCODE_DEPLOY
deployment controllers. For more information about deployment types, see Amazon ECS deployment types. - Volume
Configurations List<Pulumi.Aws Native. Ecs. Inputs. Service Volume Configuration> - The configuration for a volume specified in the task definition as a volume that is configured at launch time. Currently, the only supported volume type is an Amazon EBS volume.
- Vpc
Lattice List<Pulumi.Configurations Aws Native. Ecs. Inputs. Service Vpc Lattice Configuration>
- Availability
Zone ServiceRebalancing Availability Zone Rebalancing - Capacity
Provider []ServiceStrategy Capacity Provider Strategy Item Args - The capacity provider strategy to use for the service.
If a
capacityProviderStrategy
is specified, thelaunchType
parameter must be omitted. If nocapacityProviderStrategy
orlaunchType
is specified, thedefaultCapacityProviderStrategy
for the cluster is used. A capacity provider strategy may contain a maximum of 6 capacity providers. - Cluster string
- The short name or full Amazon Resource Name (ARN) of the cluster that you run your service on. If you do not specify a cluster, the default cluster is assumed.
- Deployment
Configuration ServiceDeployment Configuration Args - Optional deployment parameters that control how many tasks run during the deployment and the ordering of stopping and starting tasks.
- Deployment
Controller ServiceDeployment Controller Args - The deployment controller to use for the service. If no deployment controller is specified, the default value of
ECS
is used. - Desired
Count int - The number of instantiations of the specified task definition to place and keep running in your service.
For new services, if a desired count is not specified, a default value of
1
is used. When using theDAEMON
scheduling strategy, the desired count is not required. For existing services, if a desired count is not specified, it is omitted from the operation. - bool
- Specifies whether to turn on Amazon ECS managed tags for the tasks within the service. For more information, see Tagging your Amazon ECS resources in the Amazon Elastic Container Service Developer Guide.
When you use Amazon ECS managed tags, you need to set the
propagateTags
request parameter. - Enable
Execute boolCommand - Determines whether the execute command functionality is turned on for the service. If
true
, the execute command functionality is turned on for all containers in tasks as part of the service. - Health
Check intGrace Period Seconds - The period of time, in seconds, that the Amazon ECS service scheduler ignores unhealthy Elastic Load Balancing target health checks after a task has first started. This is only used when your service is configured to use a load balancer. If your service has a load balancer defined and you don't specify a health check grace period value, the default value of
0
is used. If you do not use an Elastic Load Balancing, we recommend that you use thestartPeriod
in the task definition health check parameters. For more information, see Health check. If your service's tasks take a while to start and respond to Elastic Load Balancing health checks, you can specify a health check grace period of up to 2,147,483,647 seconds (about 69 years). During that time, the Amazon ECS service scheduler ignores health check status. This grace period can prevent the service scheduler from marking tasks as unhealthy and stopping them before they have time to come up. - Launch
Type ServiceLaunch Type - The launch type on which to run your service. For more information, see Amazon ECS Launch Types in the Amazon Elastic Container Service Developer Guide.
- Load
Balancers []ServiceLoad Balancer Args - A list of load balancer objects to associate with the service. If you specify the
Role
property,LoadBalancers
must be specified as well. For information about the number of load balancers that you can specify per service, see Service Load Balancing in the Amazon Elastic Container Service Developer Guide. - Network
Configuration ServiceNetwork Configuration Args - The network configuration for the service. This parameter is required for task definitions that use the
awsvpc
network mode to receive their own elastic network interface, and it is not supported for other network modes. For more information, see Task Networking in the Amazon Elastic Container Service Developer Guide. - Placement
Constraints []ServicePlacement Constraint Args - An array of placement constraint objects to use for tasks in your service. You can specify a maximum of 10 constraints for each task. This limit includes constraints in the task definition and those specified at runtime.
- Placement
Strategies []ServicePlacement Strategy Args - The placement strategy objects to use for tasks in your service. You can specify a maximum of 5 strategy rules for each service.
- Platform
Version string - The platform version that your tasks in the service are running on. A platform version is specified only for tasks using the Fargate launch type. If one isn't specified, the
LATEST
platform version is used. For more information, see platform versions in the Amazon Elastic Container Service Developer Guide. - Service
Propagate Tags - Specifies whether to propagate the tags from the task definition to the task. If no value is specified, the tags aren't propagated. Tags can only be propagated to the task during task creation. To add tags to a task after task creation, use the TagResource API action.
You must set this to a value other than
NONE
when you use Cost Explorer. For more information, see Amazon ECS usage reports in the Amazon Elastic Container Service Developer Guide. The default isNONE
. - Role string
- The name or full Amazon Resource Name (ARN) of the IAM role that allows Amazon ECS to make calls to your load balancer on your behalf. This parameter is only permitted if you are using a load balancer with your service and your task definition doesn't use the
awsvpc
network mode. If you specify therole
parameter, you must also specify a load balancer object with theloadBalancers
parameter. If your account has already created the Amazon ECS service-linked role, that role is used for your service unless you specify a role here. The service-linked role is required if your task definition uses theawsvpc
network mode or if the service is configured to use service discovery, an external deployment controller, multiple target groups, or Elastic Inference accelerators in which case you don't specify a role here. For more information, see Using service-linked roles for Amazon ECS in the Amazon Elastic Container Service Developer Guide. If your specified role has a path other than/
, then you must either specify the full role ARN (this is recommended) or prefix the role name with the path. For example, if a role with the namebar
has a path of/foo/
then you would specify/foo/bar
as the role name. For more information, see Friendly names and paths in the IAM User Guide. - Scheduling
Strategy ServiceScheduling Strategy - The scheduling strategy to use for the service. For more information, see Services.
There are two service scheduler strategies available:
REPLICA
-The replica scheduling strategy places and maintains the desired number of tasks across your cluster. By default, the service scheduler spreads tasks across Availability Zones. You can use task placement strategies and constraints to customize task placement decisions. This scheduler strategy is required if the service uses theCODE_DEPLOY
orEXTERNAL
deployment controller types.DAEMON
-The daemon scheduling strategy deploys exactly one task on each active container instance that meets all of the task placement constraints that you specify in your cluster. The service scheduler also evaluates the task placement constraints for running tasks and will stop tasks that don't meet the placement constraints. When you're using this strategy, you don't need to specify a desired number of tasks, a task placement strategy, or use Service Auto Scaling policies. Tasks using the Fargate launch type or theCODE_DEPLOY
orEXTERNAL
deployment controller types don't support theDAEMON
scheduling strategy.
- Service
Connect ServiceConfiguration Connect Configuration Args - The configuration for this service to discover and connect to services, and be discovered by, and connected from, other services within a namespace. Tasks that run in a namespace can use short names to connect to services in the namespace. Tasks can connect to services across all of the clusters in the namespace. Tasks connect through a managed proxy container that collects logs and metrics for increased visibility. Only the tasks that Amazon ECS services create are supported with Service Connect. For more information, see Service Connect in the Amazon Elastic Container Service Developer Guide.
- Service
Name string - The name of your service. Up to 255 letters (uppercase and lowercase), numbers, underscores, and hyphens are allowed. Service names must be unique within a cluster, but you can have similarly named services in multiple clusters within a Region or across multiple Regions.
The stack update fails if you change any properties that require replacement and the
ServiceName
is configured. This is because AWS CloudFormation creates the replacement service first, but eachServiceName
must be unique in the cluster. - Service
Registries []ServiceRegistry Args - The details of the service discovery registry to associate with this service. For more information, see Service discovery. Each service may be associated with one service registry. Multiple service registries for each service isn't supported.
- Tag
Args - The metadata that you apply to the service to help you categorize and organize them. Each tag consists of a key and an optional value, both of which you define. When a service is deleted, the tags are deleted as well.
The following basic restrictions apply to tags:
- Maximum number of tags per resource - 50
- For each resource, each tag key must be unique, and each tag key can have only one value.
- Maximum key length - 128 Unicode characters in UTF-8
- Maximum value length - 256 Unicode characters in UTF-8
- If your tagging schema is used across multiple services and resources, remember that other services may have restrictions on allowed characters. Generally allowed characters are: letters, numbers, and spaces representable in UTF-8, and the following characters: + - = . _ : / @.
- Tag keys and values are case-sensitive.
- Do not use
aws:
,AWS:
, or any upper or lowercase combination of such as a prefix for either keys or values as it is reserved for AWS use. You cannot edit or delete tag keys or values with this prefix. Tags with this prefix do not count against your tags per resource limit.
- Task
Definition string - The
family
andrevision
(family:revision
) or full ARN of the task definition to run in your service. If arevision
isn't specified, the latestACTIVE
revision is used. A task definition must be specified if the service uses either theECS
orCODE_DEPLOY
deployment controllers. For more information about deployment types, see Amazon ECS deployment types. - Volume
Configurations []ServiceVolume Configuration Args - The configuration for a volume specified in the task definition as a volume that is configured at launch time. Currently, the only supported volume type is an Amazon EBS volume.
- Vpc
Lattice []ServiceConfigurations Vpc Lattice Configuration Args
- availability
Zone ServiceRebalancing Availability Zone Rebalancing - capacity
Provider List<ServiceStrategy Capacity Provider Strategy Item> - The capacity provider strategy to use for the service.
If a
capacityProviderStrategy
is specified, thelaunchType
parameter must be omitted. If nocapacityProviderStrategy
orlaunchType
is specified, thedefaultCapacityProviderStrategy
for the cluster is used. A capacity provider strategy may contain a maximum of 6 capacity providers. - cluster String
- The short name or full Amazon Resource Name (ARN) of the cluster that you run your service on. If you do not specify a cluster, the default cluster is assumed.
- deployment
Configuration ServiceDeployment Configuration - Optional deployment parameters that control how many tasks run during the deployment and the ordering of stopping and starting tasks.
- deployment
Controller ServiceDeployment Controller - The deployment controller to use for the service. If no deployment controller is specified, the default value of
ECS
is used. - desired
Count Integer - The number of instantiations of the specified task definition to place and keep running in your service.
For new services, if a desired count is not specified, a default value of
1
is used. When using theDAEMON
scheduling strategy, the desired count is not required. For existing services, if a desired count is not specified, it is omitted from the operation. - Boolean
- Specifies whether to turn on Amazon ECS managed tags for the tasks within the service. For more information, see Tagging your Amazon ECS resources in the Amazon Elastic Container Service Developer Guide.
When you use Amazon ECS managed tags, you need to set the
propagateTags
request parameter. - enable
Execute BooleanCommand - Determines whether the execute command functionality is turned on for the service. If
true
, the execute command functionality is turned on for all containers in tasks as part of the service. - health
Check IntegerGrace Period Seconds - The period of time, in seconds, that the Amazon ECS service scheduler ignores unhealthy Elastic Load Balancing target health checks after a task has first started. This is only used when your service is configured to use a load balancer. If your service has a load balancer defined and you don't specify a health check grace period value, the default value of
0
is used. If you do not use an Elastic Load Balancing, we recommend that you use thestartPeriod
in the task definition health check parameters. For more information, see Health check. If your service's tasks take a while to start and respond to Elastic Load Balancing health checks, you can specify a health check grace period of up to 2,147,483,647 seconds (about 69 years). During that time, the Amazon ECS service scheduler ignores health check status. This grace period can prevent the service scheduler from marking tasks as unhealthy and stopping them before they have time to come up. - launch
Type ServiceLaunch Type - The launch type on which to run your service. For more information, see Amazon ECS Launch Types in the Amazon Elastic Container Service Developer Guide.
- load
Balancers List<ServiceLoad Balancer> - A list of load balancer objects to associate with the service. If you specify the
Role
property,LoadBalancers
must be specified as well. For information about the number of load balancers that you can specify per service, see Service Load Balancing in the Amazon Elastic Container Service Developer Guide. - network
Configuration ServiceNetwork Configuration - The network configuration for the service. This parameter is required for task definitions that use the
awsvpc
network mode to receive their own elastic network interface, and it is not supported for other network modes. For more information, see Task Networking in the Amazon Elastic Container Service Developer Guide. - placement
Constraints List<ServicePlacement Constraint> - An array of placement constraint objects to use for tasks in your service. You can specify a maximum of 10 constraints for each task. This limit includes constraints in the task definition and those specified at runtime.
- placement
Strategies List<ServicePlacement Strategy> - The placement strategy objects to use for tasks in your service. You can specify a maximum of 5 strategy rules for each service.
- platform
Version String - The platform version that your tasks in the service are running on. A platform version is specified only for tasks using the Fargate launch type. If one isn't specified, the
LATEST
platform version is used. For more information, see platform versions in the Amazon Elastic Container Service Developer Guide. - Service
Propagate Tags - Specifies whether to propagate the tags from the task definition to the task. If no value is specified, the tags aren't propagated. Tags can only be propagated to the task during task creation. To add tags to a task after task creation, use the TagResource API action.
You must set this to a value other than
NONE
when you use Cost Explorer. For more information, see Amazon ECS usage reports in the Amazon Elastic Container Service Developer Guide. The default isNONE
. - role String
- The name or full Amazon Resource Name (ARN) of the IAM role that allows Amazon ECS to make calls to your load balancer on your behalf. This parameter is only permitted if you are using a load balancer with your service and your task definition doesn't use the
awsvpc
network mode. If you specify therole
parameter, you must also specify a load balancer object with theloadBalancers
parameter. If your account has already created the Amazon ECS service-linked role, that role is used for your service unless you specify a role here. The service-linked role is required if your task definition uses theawsvpc
network mode or if the service is configured to use service discovery, an external deployment controller, multiple target groups, or Elastic Inference accelerators in which case you don't specify a role here. For more information, see Using service-linked roles for Amazon ECS in the Amazon Elastic Container Service Developer Guide. If your specified role has a path other than/
, then you must either specify the full role ARN (this is recommended) or prefix the role name with the path. For example, if a role with the namebar
has a path of/foo/
then you would specify/foo/bar
as the role name. For more information, see Friendly names and paths in the IAM User Guide. - scheduling
Strategy ServiceScheduling Strategy - The scheduling strategy to use for the service. For more information, see Services.
There are two service scheduler strategies available:
REPLICA
-The replica scheduling strategy places and maintains the desired number of tasks across your cluster. By default, the service scheduler spreads tasks across Availability Zones. You can use task placement strategies and constraints to customize task placement decisions. This scheduler strategy is required if the service uses theCODE_DEPLOY
orEXTERNAL
deployment controller types.DAEMON
-The daemon scheduling strategy deploys exactly one task on each active container instance that meets all of the task placement constraints that you specify in your cluster. The service scheduler also evaluates the task placement constraints for running tasks and will stop tasks that don't meet the placement constraints. When you're using this strategy, you don't need to specify a desired number of tasks, a task placement strategy, or use Service Auto Scaling policies. Tasks using the Fargate launch type or theCODE_DEPLOY
orEXTERNAL
deployment controller types don't support theDAEMON
scheduling strategy.
- service
Connect ServiceConfiguration Connect Configuration - The configuration for this service to discover and connect to services, and be discovered by, and connected from, other services within a namespace. Tasks that run in a namespace can use short names to connect to services in the namespace. Tasks can connect to services across all of the clusters in the namespace. Tasks connect through a managed proxy container that collects logs and metrics for increased visibility. Only the tasks that Amazon ECS services create are supported with Service Connect. For more information, see Service Connect in the Amazon Elastic Container Service Developer Guide.
- service
Name String - The name of your service. Up to 255 letters (uppercase and lowercase), numbers, underscores, and hyphens are allowed. Service names must be unique within a cluster, but you can have similarly named services in multiple clusters within a Region or across multiple Regions.
The stack update fails if you change any properties that require replacement and the
ServiceName
is configured. This is because AWS CloudFormation creates the replacement service first, but eachServiceName
must be unique in the cluster. - service
Registries List<ServiceRegistry> - The details of the service discovery registry to associate with this service. For more information, see Service discovery. Each service may be associated with one service registry. Multiple service registries for each service isn't supported.
- List<Tag>
- The metadata that you apply to the service to help you categorize and organize them. Each tag consists of a key and an optional value, both of which you define. When a service is deleted, the tags are deleted as well.
The following basic restrictions apply to tags:
- Maximum number of tags per resource - 50
- For each resource, each tag key must be unique, and each tag key can have only one value.
- Maximum key length - 128 Unicode characters in UTF-8
- Maximum value length - 256 Unicode characters in UTF-8
- If your tagging schema is used across multiple services and resources, remember that other services may have restrictions on allowed characters. Generally allowed characters are: letters, numbers, and spaces representable in UTF-8, and the following characters: + - = . _ : / @.
- Tag keys and values are case-sensitive.
- Do not use
aws:
,AWS:
, or any upper or lowercase combination of such as a prefix for either keys or values as it is reserved for AWS use. You cannot edit or delete tag keys or values with this prefix. Tags with this prefix do not count against your tags per resource limit.
- task
Definition String - The
family
andrevision
(family:revision
) or full ARN of the task definition to run in your service. If arevision
isn't specified, the latestACTIVE
revision is used. A task definition must be specified if the service uses either theECS
orCODE_DEPLOY
deployment controllers. For more information about deployment types, see Amazon ECS deployment types. - volume
Configurations List<ServiceVolume Configuration> - The configuration for a volume specified in the task definition as a volume that is configured at launch time. Currently, the only supported volume type is an Amazon EBS volume.
- vpc
Lattice List<ServiceConfigurations Vpc Lattice Configuration>
- availability
Zone ServiceRebalancing Availability Zone Rebalancing - capacity
Provider ServiceStrategy Capacity Provider Strategy Item[] - The capacity provider strategy to use for the service.
If a
capacityProviderStrategy
is specified, thelaunchType
parameter must be omitted. If nocapacityProviderStrategy
orlaunchType
is specified, thedefaultCapacityProviderStrategy
for the cluster is used. A capacity provider strategy may contain a maximum of 6 capacity providers. - cluster string
- The short name or full Amazon Resource Name (ARN) of the cluster that you run your service on. If you do not specify a cluster, the default cluster is assumed.
- deployment
Configuration ServiceDeployment Configuration - Optional deployment parameters that control how many tasks run during the deployment and the ordering of stopping and starting tasks.
- deployment
Controller ServiceDeployment Controller - The deployment controller to use for the service. If no deployment controller is specified, the default value of
ECS
is used. - desired
Count number - The number of instantiations of the specified task definition to place and keep running in your service.
For new services, if a desired count is not specified, a default value of
1
is used. When using theDAEMON
scheduling strategy, the desired count is not required. For existing services, if a desired count is not specified, it is omitted from the operation. - boolean
- Specifies whether to turn on Amazon ECS managed tags for the tasks within the service. For more information, see Tagging your Amazon ECS resources in the Amazon Elastic Container Service Developer Guide.
When you use Amazon ECS managed tags, you need to set the
propagateTags
request parameter. - enable
Execute booleanCommand - Determines whether the execute command functionality is turned on for the service. If
true
, the execute command functionality is turned on for all containers in tasks as part of the service. - health
Check numberGrace Period Seconds - The period of time, in seconds, that the Amazon ECS service scheduler ignores unhealthy Elastic Load Balancing target health checks after a task has first started. This is only used when your service is configured to use a load balancer. If your service has a load balancer defined and you don't specify a health check grace period value, the default value of
0
is used. If you do not use an Elastic Load Balancing, we recommend that you use thestartPeriod
in the task definition health check parameters. For more information, see Health check. If your service's tasks take a while to start and respond to Elastic Load Balancing health checks, you can specify a health check grace period of up to 2,147,483,647 seconds (about 69 years). During that time, the Amazon ECS service scheduler ignores health check status. This grace period can prevent the service scheduler from marking tasks as unhealthy and stopping them before they have time to come up. - launch
Type ServiceLaunch Type - The launch type on which to run your service. For more information, see Amazon ECS Launch Types in the Amazon Elastic Container Service Developer Guide.
- load
Balancers ServiceLoad Balancer[] - A list of load balancer objects to associate with the service. If you specify the
Role
property,LoadBalancers
must be specified as well. For information about the number of load balancers that you can specify per service, see Service Load Balancing in the Amazon Elastic Container Service Developer Guide. - network
Configuration ServiceNetwork Configuration - The network configuration for the service. This parameter is required for task definitions that use the
awsvpc
network mode to receive their own elastic network interface, and it is not supported for other network modes. For more information, see Task Networking in the Amazon Elastic Container Service Developer Guide. - placement
Constraints ServicePlacement Constraint[] - An array of placement constraint objects to use for tasks in your service. You can specify a maximum of 10 constraints for each task. This limit includes constraints in the task definition and those specified at runtime.
- placement
Strategies ServicePlacement Strategy[] - The placement strategy objects to use for tasks in your service. You can specify a maximum of 5 strategy rules for each service.
- platform
Version string - The platform version that your tasks in the service are running on. A platform version is specified only for tasks using the Fargate launch type. If one isn't specified, the
LATEST
platform version is used. For more information, see platform versions in the Amazon Elastic Container Service Developer Guide. - Service
Propagate Tags - Specifies whether to propagate the tags from the task definition to the task. If no value is specified, the tags aren't propagated. Tags can only be propagated to the task during task creation. To add tags to a task after task creation, use the TagResource API action.
You must set this to a value other than
NONE
when you use Cost Explorer. For more information, see Amazon ECS usage reports in the Amazon Elastic Container Service Developer Guide. The default isNONE
. - role string
- The name or full Amazon Resource Name (ARN) of the IAM role that allows Amazon ECS to make calls to your load balancer on your behalf. This parameter is only permitted if you are using a load balancer with your service and your task definition doesn't use the
awsvpc
network mode. If you specify therole
parameter, you must also specify a load balancer object with theloadBalancers
parameter. If your account has already created the Amazon ECS service-linked role, that role is used for your service unless you specify a role here. The service-linked role is required if your task definition uses theawsvpc
network mode or if the service is configured to use service discovery, an external deployment controller, multiple target groups, or Elastic Inference accelerators in which case you don't specify a role here. For more information, see Using service-linked roles for Amazon ECS in the Amazon Elastic Container Service Developer Guide. If your specified role has a path other than/
, then you must either specify the full role ARN (this is recommended) or prefix the role name with the path. For example, if a role with the namebar
has a path of/foo/
then you would specify/foo/bar
as the role name. For more information, see Friendly names and paths in the IAM User Guide. - scheduling
Strategy ServiceScheduling Strategy - The scheduling strategy to use for the service. For more information, see Services.
There are two service scheduler strategies available:
REPLICA
-The replica scheduling strategy places and maintains the desired number of tasks across your cluster. By default, the service scheduler spreads tasks across Availability Zones. You can use task placement strategies and constraints to customize task placement decisions. This scheduler strategy is required if the service uses theCODE_DEPLOY
orEXTERNAL
deployment controller types.DAEMON
-The daemon scheduling strategy deploys exactly one task on each active container instance that meets all of the task placement constraints that you specify in your cluster. The service scheduler also evaluates the task placement constraints for running tasks and will stop tasks that don't meet the placement constraints. When you're using this strategy, you don't need to specify a desired number of tasks, a task placement strategy, or use Service Auto Scaling policies. Tasks using the Fargate launch type or theCODE_DEPLOY
orEXTERNAL
deployment controller types don't support theDAEMON
scheduling strategy.
- service
Connect ServiceConfiguration Connect Configuration - The configuration for this service to discover and connect to services, and be discovered by, and connected from, other services within a namespace. Tasks that run in a namespace can use short names to connect to services in the namespace. Tasks can connect to services across all of the clusters in the namespace. Tasks connect through a managed proxy container that collects logs and metrics for increased visibility. Only the tasks that Amazon ECS services create are supported with Service Connect. For more information, see Service Connect in the Amazon Elastic Container Service Developer Guide.
- service
Name string - The name of your service. Up to 255 letters (uppercase and lowercase), numbers, underscores, and hyphens are allowed. Service names must be unique within a cluster, but you can have similarly named services in multiple clusters within a Region or across multiple Regions.
The stack update fails if you change any properties that require replacement and the
ServiceName
is configured. This is because AWS CloudFormation creates the replacement service first, but eachServiceName
must be unique in the cluster. - service
Registries ServiceRegistry[] - The details of the service discovery registry to associate with this service. For more information, see Service discovery. Each service may be associated with one service registry. Multiple service registries for each service isn't supported.
- Tag[]
- The metadata that you apply to the service to help you categorize and organize them. Each tag consists of a key and an optional value, both of which you define. When a service is deleted, the tags are deleted as well.
The following basic restrictions apply to tags:
- Maximum number of tags per resource - 50
- For each resource, each tag key must be unique, and each tag key can have only one value.
- Maximum key length - 128 Unicode characters in UTF-8
- Maximum value length - 256 Unicode characters in UTF-8
- If your tagging schema is used across multiple services and resources, remember that other services may have restrictions on allowed characters. Generally allowed characters are: letters, numbers, and spaces representable in UTF-8, and the following characters: + - = . _ : / @.
- Tag keys and values are case-sensitive.
- Do not use
aws:
,AWS:
, or any upper or lowercase combination of such as a prefix for either keys or values as it is reserved for AWS use. You cannot edit or delete tag keys or values with this prefix. Tags with this prefix do not count against your tags per resource limit.
- task
Definition string - The
family
andrevision
(family:revision
) or full ARN of the task definition to run in your service. If arevision
isn't specified, the latestACTIVE
revision is used. A task definition must be specified if the service uses either theECS
orCODE_DEPLOY
deployment controllers. For more information about deployment types, see Amazon ECS deployment types. - volume
Configurations ServiceVolume Configuration[] - The configuration for a volume specified in the task definition as a volume that is configured at launch time. Currently, the only supported volume type is an Amazon EBS volume.
- vpc
Lattice ServiceConfigurations Vpc Lattice Configuration[]
- availability_
zone_ Servicerebalancing Availability Zone Rebalancing - capacity_
provider_ Sequence[Servicestrategy Capacity Provider Strategy Item Args] - The capacity provider strategy to use for the service.
If a
capacityProviderStrategy
is specified, thelaunchType
parameter must be omitted. If nocapacityProviderStrategy
orlaunchType
is specified, thedefaultCapacityProviderStrategy
for the cluster is used. A capacity provider strategy may contain a maximum of 6 capacity providers. - cluster str
- The short name or full Amazon Resource Name (ARN) of the cluster that you run your service on. If you do not specify a cluster, the default cluster is assumed.
- deployment_
configuration ServiceDeployment Configuration Args - Optional deployment parameters that control how many tasks run during the deployment and the ordering of stopping and starting tasks.
- deployment_
controller ServiceDeployment Controller Args - The deployment controller to use for the service. If no deployment controller is specified, the default value of
ECS
is used. - desired_
count int - The number of instantiations of the specified task definition to place and keep running in your service.
For new services, if a desired count is not specified, a default value of
1
is used. When using theDAEMON
scheduling strategy, the desired count is not required. For existing services, if a desired count is not specified, it is omitted from the operation. - bool
- Specifies whether to turn on Amazon ECS managed tags for the tasks within the service. For more information, see Tagging your Amazon ECS resources in the Amazon Elastic Container Service Developer Guide.
When you use Amazon ECS managed tags, you need to set the
propagateTags
request parameter. - enable_
execute_ boolcommand - Determines whether the execute command functionality is turned on for the service. If
true
, the execute command functionality is turned on for all containers in tasks as part of the service. - health_
check_ intgrace_ period_ seconds - The period of time, in seconds, that the Amazon ECS service scheduler ignores unhealthy Elastic Load Balancing target health checks after a task has first started. This is only used when your service is configured to use a load balancer. If your service has a load balancer defined and you don't specify a health check grace period value, the default value of
0
is used. If you do not use an Elastic Load Balancing, we recommend that you use thestartPeriod
in the task definition health check parameters. For more information, see Health check. If your service's tasks take a while to start and respond to Elastic Load Balancing health checks, you can specify a health check grace period of up to 2,147,483,647 seconds (about 69 years). During that time, the Amazon ECS service scheduler ignores health check status. This grace period can prevent the service scheduler from marking tasks as unhealthy and stopping them before they have time to come up. - launch_
type ServiceLaunch Type - The launch type on which to run your service. For more information, see Amazon ECS Launch Types in the Amazon Elastic Container Service Developer Guide.
- load_
balancers Sequence[ServiceLoad Balancer Args] - A list of load balancer objects to associate with the service. If you specify the
Role
property,LoadBalancers
must be specified as well. For information about the number of load balancers that you can specify per service, see Service Load Balancing in the Amazon Elastic Container Service Developer Guide. - network_
configuration ServiceNetwork Configuration Args - The network configuration for the service. This parameter is required for task definitions that use the
awsvpc
network mode to receive their own elastic network interface, and it is not supported for other network modes. For more information, see Task Networking in the Amazon Elastic Container Service Developer Guide. - placement_
constraints Sequence[ServicePlacement Constraint Args] - An array of placement constraint objects to use for tasks in your service. You can specify a maximum of 10 constraints for each task. This limit includes constraints in the task definition and those specified at runtime.
- placement_
strategies Sequence[ServicePlacement Strategy Args] - The placement strategy objects to use for tasks in your service. You can specify a maximum of 5 strategy rules for each service.
- platform_
version str - The platform version that your tasks in the service are running on. A platform version is specified only for tasks using the Fargate launch type. If one isn't specified, the
LATEST
platform version is used. For more information, see platform versions in the Amazon Elastic Container Service Developer Guide. - Service
Propagate Tags - Specifies whether to propagate the tags from the task definition to the task. If no value is specified, the tags aren't propagated. Tags can only be propagated to the task during task creation. To add tags to a task after task creation, use the TagResource API action.
You must set this to a value other than
NONE
when you use Cost Explorer. For more information, see Amazon ECS usage reports in the Amazon Elastic Container Service Developer Guide. The default isNONE
. - role str
- The name or full Amazon Resource Name (ARN) of the IAM role that allows Amazon ECS to make calls to your load balancer on your behalf. This parameter is only permitted if you are using a load balancer with your service and your task definition doesn't use the
awsvpc
network mode. If you specify therole
parameter, you must also specify a load balancer object with theloadBalancers
parameter. If your account has already created the Amazon ECS service-linked role, that role is used for your service unless you specify a role here. The service-linked role is required if your task definition uses theawsvpc
network mode or if the service is configured to use service discovery, an external deployment controller, multiple target groups, or Elastic Inference accelerators in which case you don't specify a role here. For more information, see Using service-linked roles for Amazon ECS in the Amazon Elastic Container Service Developer Guide. If your specified role has a path other than/
, then you must either specify the full role ARN (this is recommended) or prefix the role name with the path. For example, if a role with the namebar
has a path of/foo/
then you would specify/foo/bar
as the role name. For more information, see Friendly names and paths in the IAM User Guide. - scheduling_
strategy ServiceScheduling Strategy - The scheduling strategy to use for the service. For more information, see Services.
There are two service scheduler strategies available:
REPLICA
-The replica scheduling strategy places and maintains the desired number of tasks across your cluster. By default, the service scheduler spreads tasks across Availability Zones. You can use task placement strategies and constraints to customize task placement decisions. This scheduler strategy is required if the service uses theCODE_DEPLOY
orEXTERNAL
deployment controller types.DAEMON
-The daemon scheduling strategy deploys exactly one task on each active container instance that meets all of the task placement constraints that you specify in your cluster. The service scheduler also evaluates the task placement constraints for running tasks and will stop tasks that don't meet the placement constraints. When you're using this strategy, you don't need to specify a desired number of tasks, a task placement strategy, or use Service Auto Scaling policies. Tasks using the Fargate launch type or theCODE_DEPLOY
orEXTERNAL
deployment controller types don't support theDAEMON
scheduling strategy.
- service_
connect_ Serviceconfiguration Connect Configuration Args - The configuration for this service to discover and connect to services, and be discovered by, and connected from, other services within a namespace. Tasks that run in a namespace can use short names to connect to services in the namespace. Tasks can connect to services across all of the clusters in the namespace. Tasks connect through a managed proxy container that collects logs and metrics for increased visibility. Only the tasks that Amazon ECS services create are supported with Service Connect. For more information, see Service Connect in the Amazon Elastic Container Service Developer Guide.
- service_
name str - The name of your service. Up to 255 letters (uppercase and lowercase), numbers, underscores, and hyphens are allowed. Service names must be unique within a cluster, but you can have similarly named services in multiple clusters within a Region or across multiple Regions.
The stack update fails if you change any properties that require replacement and the
ServiceName
is configured. This is because AWS CloudFormation creates the replacement service first, but eachServiceName
must be unique in the cluster. - service_
registries Sequence[ServiceRegistry Args] - The details of the service discovery registry to associate with this service. For more information, see Service discovery. Each service may be associated with one service registry. Multiple service registries for each service isn't supported.
- Sequence[Tag
Args] - The metadata that you apply to the service to help you categorize and organize them. Each tag consists of a key and an optional value, both of which you define. When a service is deleted, the tags are deleted as well.
The following basic restrictions apply to tags:
- Maximum number of tags per resource - 50
- For each resource, each tag key must be unique, and each tag key can have only one value.
- Maximum key length - 128 Unicode characters in UTF-8
- Maximum value length - 256 Unicode characters in UTF-8
- If your tagging schema is used across multiple services and resources, remember that other services may have restrictions on allowed characters. Generally allowed characters are: letters, numbers, and spaces representable in UTF-8, and the following characters: + - = . _ : / @.
- Tag keys and values are case-sensitive.
- Do not use
aws:
,AWS:
, or any upper or lowercase combination of such as a prefix for either keys or values as it is reserved for AWS use. You cannot edit or delete tag keys or values with this prefix. Tags with this prefix do not count against your tags per resource limit.
- task_
definition str - The
family
andrevision
(family:revision
) or full ARN of the task definition to run in your service. If arevision
isn't specified, the latestACTIVE
revision is used. A task definition must be specified if the service uses either theECS
orCODE_DEPLOY
deployment controllers. For more information about deployment types, see Amazon ECS deployment types. - volume_
configurations Sequence[ServiceVolume Configuration Args] - The configuration for a volume specified in the task definition as a volume that is configured at launch time. Currently, the only supported volume type is an Amazon EBS volume.
- vpc_
lattice_ Sequence[Serviceconfigurations Vpc Lattice Configuration Args]
- availability
Zone "ENABLED" | "DISABLED"Rebalancing - capacity
Provider List<Property Map>Strategy - The capacity provider strategy to use for the service.
If a
capacityProviderStrategy
is specified, thelaunchType
parameter must be omitted. If nocapacityProviderStrategy
orlaunchType
is specified, thedefaultCapacityProviderStrategy
for the cluster is used. A capacity provider strategy may contain a maximum of 6 capacity providers. - cluster String
- The short name or full Amazon Resource Name (ARN) of the cluster that you run your service on. If you do not specify a cluster, the default cluster is assumed.
- deployment
Configuration Property Map - Optional deployment parameters that control how many tasks run during the deployment and the ordering of stopping and starting tasks.
- deployment
Controller Property Map - The deployment controller to use for the service. If no deployment controller is specified, the default value of
ECS
is used. - desired
Count Number - The number of instantiations of the specified task definition to place and keep running in your service.
For new services, if a desired count is not specified, a default value of
1
is used. When using theDAEMON
scheduling strategy, the desired count is not required. For existing services, if a desired count is not specified, it is omitted from the operation. - Boolean
- Specifies whether to turn on Amazon ECS managed tags for the tasks within the service. For more information, see Tagging your Amazon ECS resources in the Amazon Elastic Container Service Developer Guide.
When you use Amazon ECS managed tags, you need to set the
propagateTags
request parameter. - enable
Execute BooleanCommand - Determines whether the execute command functionality is turned on for the service. If
true
, the execute command functionality is turned on for all containers in tasks as part of the service. - health
Check NumberGrace Period Seconds - The period of time, in seconds, that the Amazon ECS service scheduler ignores unhealthy Elastic Load Balancing target health checks after a task has first started. This is only used when your service is configured to use a load balancer. If your service has a load balancer defined and you don't specify a health check grace period value, the default value of
0
is used. If you do not use an Elastic Load Balancing, we recommend that you use thestartPeriod
in the task definition health check parameters. For more information, see Health check. If your service's tasks take a while to start and respond to Elastic Load Balancing health checks, you can specify a health check grace period of up to 2,147,483,647 seconds (about 69 years). During that time, the Amazon ECS service scheduler ignores health check status. This grace period can prevent the service scheduler from marking tasks as unhealthy and stopping them before they have time to come up. - launch
Type "EC2" | "FARGATE" | "EXTERNAL" - The launch type on which to run your service. For more information, see Amazon ECS Launch Types in the Amazon Elastic Container Service Developer Guide.
- load
Balancers List<Property Map> - A list of load balancer objects to associate with the service. If you specify the
Role
property,LoadBalancers
must be specified as well. For information about the number of load balancers that you can specify per service, see Service Load Balancing in the Amazon Elastic Container Service Developer Guide. - network
Configuration Property Map - The network configuration for the service. This parameter is required for task definitions that use the
awsvpc
network mode to receive their own elastic network interface, and it is not supported for other network modes. For more information, see Task Networking in the Amazon Elastic Container Service Developer Guide. - placement
Constraints List<Property Map> - An array of placement constraint objects to use for tasks in your service. You can specify a maximum of 10 constraints for each task. This limit includes constraints in the task definition and those specified at runtime.
- placement
Strategies List<Property Map> - The placement strategy objects to use for tasks in your service. You can specify a maximum of 5 strategy rules for each service.
- platform
Version String - The platform version that your tasks in the service are running on. A platform version is specified only for tasks using the Fargate launch type. If one isn't specified, the
LATEST
platform version is used. For more information, see platform versions in the Amazon Elastic Container Service Developer Guide. - "SERVICE" | "TASK_DEFINITION"
- Specifies whether to propagate the tags from the task definition to the task. If no value is specified, the tags aren't propagated. Tags can only be propagated to the task during task creation. To add tags to a task after task creation, use the TagResource API action.
You must set this to a value other than
NONE
when you use Cost Explorer. For more information, see Amazon ECS usage reports in the Amazon Elastic Container Service Developer Guide. The default isNONE
. - role String
- The name or full Amazon Resource Name (ARN) of the IAM role that allows Amazon ECS to make calls to your load balancer on your behalf. This parameter is only permitted if you are using a load balancer with your service and your task definition doesn't use the
awsvpc
network mode. If you specify therole
parameter, you must also specify a load balancer object with theloadBalancers
parameter. If your account has already created the Amazon ECS service-linked role, that role is used for your service unless you specify a role here. The service-linked role is required if your task definition uses theawsvpc
network mode or if the service is configured to use service discovery, an external deployment controller, multiple target groups, or Elastic Inference accelerators in which case you don't specify a role here. For more information, see Using service-linked roles for Amazon ECS in the Amazon Elastic Container Service Developer Guide. If your specified role has a path other than/
, then you must either specify the full role ARN (this is recommended) or prefix the role name with the path. For example, if a role with the namebar
has a path of/foo/
then you would specify/foo/bar
as the role name. For more information, see Friendly names and paths in the IAM User Guide. - scheduling
Strategy "DAEMON" | "REPLICA" - The scheduling strategy to use for the service. For more information, see Services.
There are two service scheduler strategies available:
REPLICA
-The replica scheduling strategy places and maintains the desired number of tasks across your cluster. By default, the service scheduler spreads tasks across Availability Zones. You can use task placement strategies and constraints to customize task placement decisions. This scheduler strategy is required if the service uses theCODE_DEPLOY
orEXTERNAL
deployment controller types.DAEMON
-The daemon scheduling strategy deploys exactly one task on each active container instance that meets all of the task placement constraints that you specify in your cluster. The service scheduler also evaluates the task placement constraints for running tasks and will stop tasks that don't meet the placement constraints. When you're using this strategy, you don't need to specify a desired number of tasks, a task placement strategy, or use Service Auto Scaling policies. Tasks using the Fargate launch type or theCODE_DEPLOY
orEXTERNAL
deployment controller types don't support theDAEMON
scheduling strategy.
- service
Connect Property MapConfiguration - The configuration for this service to discover and connect to services, and be discovered by, and connected from, other services within a namespace. Tasks that run in a namespace can use short names to connect to services in the namespace. Tasks can connect to services across all of the clusters in the namespace. Tasks connect through a managed proxy container that collects logs and metrics for increased visibility. Only the tasks that Amazon ECS services create are supported with Service Connect. For more information, see Service Connect in the Amazon Elastic Container Service Developer Guide.
- service
Name String - The name of your service. Up to 255 letters (uppercase and lowercase), numbers, underscores, and hyphens are allowed. Service names must be unique within a cluster, but you can have similarly named services in multiple clusters within a Region or across multiple Regions.
The stack update fails if you change any properties that require replacement and the
ServiceName
is configured. This is because AWS CloudFormation creates the replacement service first, but eachServiceName
must be unique in the cluster. - service
Registries List<Property Map> - The details of the service discovery registry to associate with this service. For more information, see Service discovery. Each service may be associated with one service registry. Multiple service registries for each service isn't supported.
- List<Property Map>
- The metadata that you apply to the service to help you categorize and organize them. Each tag consists of a key and an optional value, both of which you define. When a service is deleted, the tags are deleted as well.
The following basic restrictions apply to tags:
- Maximum number of tags per resource - 50
- For each resource, each tag key must be unique, and each tag key can have only one value.
- Maximum key length - 128 Unicode characters in UTF-8
- Maximum value length - 256 Unicode characters in UTF-8
- If your tagging schema is used across multiple services and resources, remember that other services may have restrictions on allowed characters. Generally allowed characters are: letters, numbers, and spaces representable in UTF-8, and the following characters: + - = . _ : / @.
- Tag keys and values are case-sensitive.
- Do not use
aws:
,AWS:
, or any upper or lowercase combination of such as a prefix for either keys or values as it is reserved for AWS use. You cannot edit or delete tag keys or values with this prefix. Tags with this prefix do not count against your tags per resource limit.
- task
Definition String - The
family
andrevision
(family:revision
) or full ARN of the task definition to run in your service. If arevision
isn't specified, the latestACTIVE
revision is used. A task definition must be specified if the service uses either theECS
orCODE_DEPLOY
deployment controllers. For more information about deployment types, see Amazon ECS deployment types. - volume
Configurations List<Property Map> - The configuration for a volume specified in the task definition as a volume that is configured at launch time. Currently, the only supported volume type is an Amazon EBS volume.
- vpc
Lattice List<Property Map>Configurations
Outputs
All input properties are implicitly available as output properties. Additionally, the Service resource produces the following output properties:
- Id string
- The provider-assigned unique ID for this managed resource.
- Name string
- The name of the Amazon ECS service, such as
sample-webapp
. - Service
Arn string - Not currently supported in AWS CloudFormation .
- Id string
- The provider-assigned unique ID for this managed resource.
- Name string
- The name of the Amazon ECS service, such as
sample-webapp
. - Service
Arn string - Not currently supported in AWS CloudFormation .
- id String
- The provider-assigned unique ID for this managed resource.
- name String
- The name of the Amazon ECS service, such as
sample-webapp
. - service
Arn String - Not currently supported in AWS CloudFormation .
- id string
- The provider-assigned unique ID for this managed resource.
- name string
- The name of the Amazon ECS service, such as
sample-webapp
. - service
Arn string - Not currently supported in AWS CloudFormation .
- id str
- The provider-assigned unique ID for this managed resource.
- name str
- The name of the Amazon ECS service, such as
sample-webapp
. - service_
arn str - Not currently supported in AWS CloudFormation .
- id String
- The provider-assigned unique ID for this managed resource.
- name String
- The name of the Amazon ECS service, such as
sample-webapp
. - service
Arn String - Not currently supported in AWS CloudFormation .
Supporting Types
ServiceAvailabilityZoneRebalancing, ServiceAvailabilityZoneRebalancingArgs
- Enabled
- ENABLED
- Disabled
- DISABLED
- Service
Availability Zone Rebalancing Enabled - ENABLED
- Service
Availability Zone Rebalancing Disabled - DISABLED
- Enabled
- ENABLED
- Disabled
- DISABLED
- Enabled
- ENABLED
- Disabled
- DISABLED
- ENABLED
- ENABLED
- DISABLED
- DISABLED
- "ENABLED"
- ENABLED
- "DISABLED"
- DISABLED
ServiceAwsVpcConfiguration, ServiceAwsVpcConfigurationArgs
- Assign
Public Pulumi.Ip Aws Native. Ecs. Service Aws Vpc Configuration Assign Public Ip - Whether the task's elastic network interface receives a public IP address. The default value is
DISABLED
. - Security
Groups List<string> - The IDs of the security groups associated with the task or service. If you don't specify a security group, the default security group for the VPC is used. There's a limit of 5 security groups that can be specified per
awsvpcConfiguration
. All specified security groups must be from the same VPC. - Subnets List<string>
- The IDs of the subnets associated with the task or service. There's a limit of 16 subnets that can be specified per
awsvpcConfiguration
. All specified subnets must be from the same VPC.
- Assign
Public ServiceIp Aws Vpc Configuration Assign Public Ip - Whether the task's elastic network interface receives a public IP address. The default value is
DISABLED
. - Security
Groups []string - The IDs of the security groups associated with the task or service. If you don't specify a security group, the default security group for the VPC is used. There's a limit of 5 security groups that can be specified per
awsvpcConfiguration
. All specified security groups must be from the same VPC. - Subnets []string
- The IDs of the subnets associated with the task or service. There's a limit of 16 subnets that can be specified per
awsvpcConfiguration
. All specified subnets must be from the same VPC.
- assign
Public ServiceIp Aws Vpc Configuration Assign Public Ip - Whether the task's elastic network interface receives a public IP address. The default value is
DISABLED
. - security
Groups List<String> - The IDs of the security groups associated with the task or service. If you don't specify a security group, the default security group for the VPC is used. There's a limit of 5 security groups that can be specified per
awsvpcConfiguration
. All specified security groups must be from the same VPC. - subnets List<String>
- The IDs of the subnets associated with the task or service. There's a limit of 16 subnets that can be specified per
awsvpcConfiguration
. All specified subnets must be from the same VPC.
- assign
Public ServiceIp Aws Vpc Configuration Assign Public Ip - Whether the task's elastic network interface receives a public IP address. The default value is
DISABLED
. - security
Groups string[] - The IDs of the security groups associated with the task or service. If you don't specify a security group, the default security group for the VPC is used. There's a limit of 5 security groups that can be specified per
awsvpcConfiguration
. All specified security groups must be from the same VPC. - subnets string[]
- The IDs of the subnets associated with the task or service. There's a limit of 16 subnets that can be specified per
awsvpcConfiguration
. All specified subnets must be from the same VPC.
- assign_
public_ Serviceip Aws Vpc Configuration Assign Public Ip - Whether the task's elastic network interface receives a public IP address. The default value is
DISABLED
. - security_
groups Sequence[str] - The IDs of the security groups associated with the task or service. If you don't specify a security group, the default security group for the VPC is used. There's a limit of 5 security groups that can be specified per
awsvpcConfiguration
. All specified security groups must be from the same VPC. - subnets Sequence[str]
- The IDs of the subnets associated with the task or service. There's a limit of 16 subnets that can be specified per
awsvpcConfiguration
. All specified subnets must be from the same VPC.
- assign
Public "DISABLED" | "ENABLED"Ip - Whether the task's elastic network interface receives a public IP address. The default value is
DISABLED
. - security
Groups List<String> - The IDs of the security groups associated with the task or service. If you don't specify a security group, the default security group for the VPC is used. There's a limit of 5 security groups that can be specified per
awsvpcConfiguration
. All specified security groups must be from the same VPC. - subnets List<String>
- The IDs of the subnets associated with the task or service. There's a limit of 16 subnets that can be specified per
awsvpcConfiguration
. All specified subnets must be from the same VPC.
ServiceAwsVpcConfigurationAssignPublicIp, ServiceAwsVpcConfigurationAssignPublicIpArgs
- Disabled
- DISABLED
- Enabled
- ENABLED
- Service
Aws Vpc Configuration Assign Public Ip Disabled - DISABLED
- Service
Aws Vpc Configuration Assign Public Ip Enabled - ENABLED
- Disabled
- DISABLED
- Enabled
- ENABLED
- Disabled
- DISABLED
- Enabled
- ENABLED
- DISABLED
- DISABLED
- ENABLED
- ENABLED
- "DISABLED"
- DISABLED
- "ENABLED"
- ENABLED
ServiceCapacityProviderStrategyItem, ServiceCapacityProviderStrategyItemArgs
- Base int
- The base value designates how many tasks, at a minimum, to run on the specified capacity provider. Only one capacity provider in a capacity provider strategy can have a base defined. If no value is specified, the default value of
0
is used. - Capacity
Provider string - The short name of the capacity provider.
- Weight int
- The weight value designates the relative percentage of the total number of tasks launched that should use the specified capacity provider. The
weight
value is taken into consideration after thebase
value, if defined, is satisfied. If noweight
value is specified, the default value of0
is used. When multiple capacity providers are specified within a capacity provider strategy, at least one of the capacity providers must have a weight value greater than zero and any capacity providers with a weight of0
can't be used to place tasks. If you specify multiple capacity providers in a strategy that all have a weight of0
, anyRunTask
orCreateService
actions using the capacity provider strategy will fail. An example scenario for using weights is defining a strategy that contains two capacity providers and both have a weight of1
, then when thebase
is satisfied, the tasks will be split evenly across the two capacity providers. Using that same logic, if you specify a weight of1
for capacityProviderA and a weight of4
for capacityProviderB, then for every one task that's run using capacityProviderA, four tasks would use capacityProviderB.
- Base int
- The base value designates how many tasks, at a minimum, to run on the specified capacity provider. Only one capacity provider in a capacity provider strategy can have a base defined. If no value is specified, the default value of
0
is used. - Capacity
Provider string - The short name of the capacity provider.
- Weight int
- The weight value designates the relative percentage of the total number of tasks launched that should use the specified capacity provider. The
weight
value is taken into consideration after thebase
value, if defined, is satisfied. If noweight
value is specified, the default value of0
is used. When multiple capacity providers are specified within a capacity provider strategy, at least one of the capacity providers must have a weight value greater than zero and any capacity providers with a weight of0
can't be used to place tasks. If you specify multiple capacity providers in a strategy that all have a weight of0
, anyRunTask
orCreateService
actions using the capacity provider strategy will fail. An example scenario for using weights is defining a strategy that contains two capacity providers and both have a weight of1
, then when thebase
is satisfied, the tasks will be split evenly across the two capacity providers. Using that same logic, if you specify a weight of1
for capacityProviderA and a weight of4
for capacityProviderB, then for every one task that's run using capacityProviderA, four tasks would use capacityProviderB.
- base Integer
- The base value designates how many tasks, at a minimum, to run on the specified capacity provider. Only one capacity provider in a capacity provider strategy can have a base defined. If no value is specified, the default value of
0
is used. - capacity
Provider String - The short name of the capacity provider.
- weight Integer
- The weight value designates the relative percentage of the total number of tasks launched that should use the specified capacity provider. The
weight
value is taken into consideration after thebase
value, if defined, is satisfied. If noweight
value is specified, the default value of0
is used. When multiple capacity providers are specified within a capacity provider strategy, at least one of the capacity providers must have a weight value greater than zero and any capacity providers with a weight of0
can't be used to place tasks. If you specify multiple capacity providers in a strategy that all have a weight of0
, anyRunTask
orCreateService
actions using the capacity provider strategy will fail. An example scenario for using weights is defining a strategy that contains two capacity providers and both have a weight of1
, then when thebase
is satisfied, the tasks will be split evenly across the two capacity providers. Using that same logic, if you specify a weight of1
for capacityProviderA and a weight of4
for capacityProviderB, then for every one task that's run using capacityProviderA, four tasks would use capacityProviderB.
- base number
- The base value designates how many tasks, at a minimum, to run on the specified capacity provider. Only one capacity provider in a capacity provider strategy can have a base defined. If no value is specified, the default value of
0
is used. - capacity
Provider string - The short name of the capacity provider.
- weight number
- The weight value designates the relative percentage of the total number of tasks launched that should use the specified capacity provider. The
weight
value is taken into consideration after thebase
value, if defined, is satisfied. If noweight
value is specified, the default value of0
is used. When multiple capacity providers are specified within a capacity provider strategy, at least one of the capacity providers must have a weight value greater than zero and any capacity providers with a weight of0
can't be used to place tasks. If you specify multiple capacity providers in a strategy that all have a weight of0
, anyRunTask
orCreateService
actions using the capacity provider strategy will fail. An example scenario for using weights is defining a strategy that contains two capacity providers and both have a weight of1
, then when thebase
is satisfied, the tasks will be split evenly across the two capacity providers. Using that same logic, if you specify a weight of1
for capacityProviderA and a weight of4
for capacityProviderB, then for every one task that's run using capacityProviderA, four tasks would use capacityProviderB.
- base int
- The base value designates how many tasks, at a minimum, to run on the specified capacity provider. Only one capacity provider in a capacity provider strategy can have a base defined. If no value is specified, the default value of
0
is used. - capacity_
provider str - The short name of the capacity provider.
- weight int
- The weight value designates the relative percentage of the total number of tasks launched that should use the specified capacity provider. The
weight
value is taken into consideration after thebase
value, if defined, is satisfied. If noweight
value is specified, the default value of0
is used. When multiple capacity providers are specified within a capacity provider strategy, at least one of the capacity providers must have a weight value greater than zero and any capacity providers with a weight of0
can't be used to place tasks. If you specify multiple capacity providers in a strategy that all have a weight of0
, anyRunTask
orCreateService
actions using the capacity provider strategy will fail. An example scenario for using weights is defining a strategy that contains two capacity providers and both have a weight of1
, then when thebase
is satisfied, the tasks will be split evenly across the two capacity providers. Using that same logic, if you specify a weight of1
for capacityProviderA and a weight of4
for capacityProviderB, then for every one task that's run using capacityProviderA, four tasks would use capacityProviderB.
- base Number
- The base value designates how many tasks, at a minimum, to run on the specified capacity provider. Only one capacity provider in a capacity provider strategy can have a base defined. If no value is specified, the default value of
0
is used. - capacity
Provider String - The short name of the capacity provider.
- weight Number
- The weight value designates the relative percentage of the total number of tasks launched that should use the specified capacity provider. The
weight
value is taken into consideration after thebase
value, if defined, is satisfied. If noweight
value is specified, the default value of0
is used. When multiple capacity providers are specified within a capacity provider strategy, at least one of the capacity providers must have a weight value greater than zero and any capacity providers with a weight of0
can't be used to place tasks. If you specify multiple capacity providers in a strategy that all have a weight of0
, anyRunTask
orCreateService
actions using the capacity provider strategy will fail. An example scenario for using weights is defining a strategy that contains two capacity providers and both have a weight of1
, then when thebase
is satisfied, the tasks will be split evenly across the two capacity providers. Using that same logic, if you specify a weight of1
for capacityProviderA and a weight of4
for capacityProviderB, then for every one task that's run using capacityProviderA, four tasks would use capacityProviderB.
ServiceConnectClientAlias, ServiceConnectClientAliasArgs
- Port int
- The listening port number for the Service Connect proxy. This port is available inside of all of the tasks within the same namespace. To avoid changing your applications in client Amazon ECS services, set this to the same port that the client application uses by default. For more information, see Service Connect in the Amazon Elastic Container Service Developer Guide.
- Dns
Name string - The
dnsName
is the name that you use in the applications of client tasks to connect to this service. The name must be a valid DNS name but doesn't need to be fully-qualified. The name can include up to 127 characters. The name can include lowercase letters, numbers, underscores (_), hyphens (-), and periods (.). The name can't start with a hyphen. If this parameter isn't specified, the default value ofdiscoveryName.namespace
is used. If thediscoveryName
isn't specified, the port mapping name from the task definition is used inportName.namespace
. To avoid changing your applications in client Amazon ECS services, set this to the same name that the client application uses by default. For example, a few common names aredatabase
,db
, or the lowercase name of a database, such asmysql
orredis
. For more information, see Service Connect in the Amazon Elastic Container Service Developer Guide.
- Port int
- The listening port number for the Service Connect proxy. This port is available inside of all of the tasks within the same namespace. To avoid changing your applications in client Amazon ECS services, set this to the same port that the client application uses by default. For more information, see Service Connect in the Amazon Elastic Container Service Developer Guide.
- Dns
Name string - The
dnsName
is the name that you use in the applications of client tasks to connect to this service. The name must be a valid DNS name but doesn't need to be fully-qualified. The name can include up to 127 characters. The name can include lowercase letters, numbers, underscores (_), hyphens (-), and periods (.). The name can't start with a hyphen. If this parameter isn't specified, the default value ofdiscoveryName.namespace
is used. If thediscoveryName
isn't specified, the port mapping name from the task definition is used inportName.namespace
. To avoid changing your applications in client Amazon ECS services, set this to the same name that the client application uses by default. For example, a few common names aredatabase
,db
, or the lowercase name of a database, such asmysql
orredis
. For more information, see Service Connect in the Amazon Elastic Container Service Developer Guide.
- port Integer
- The listening port number for the Service Connect proxy. This port is available inside of all of the tasks within the same namespace. To avoid changing your applications in client Amazon ECS services, set this to the same port that the client application uses by default. For more information, see Service Connect in the Amazon Elastic Container Service Developer Guide.
- dns
Name String - The
dnsName
is the name that you use in the applications of client tasks to connect to this service. The name must be a valid DNS name but doesn't need to be fully-qualified. The name can include up to 127 characters. The name can include lowercase letters, numbers, underscores (_), hyphens (-), and periods (.). The name can't start with a hyphen. If this parameter isn't specified, the default value ofdiscoveryName.namespace
is used. If thediscoveryName
isn't specified, the port mapping name from the task definition is used inportName.namespace
. To avoid changing your applications in client Amazon ECS services, set this to the same name that the client application uses by default. For example, a few common names aredatabase
,db
, or the lowercase name of a database, such asmysql
orredis
. For more information, see Service Connect in the Amazon Elastic Container Service Developer Guide.
- port number
- The listening port number for the Service Connect proxy. This port is available inside of all of the tasks within the same namespace. To avoid changing your applications in client Amazon ECS services, set this to the same port that the client application uses by default. For more information, see Service Connect in the Amazon Elastic Container Service Developer Guide.
- dns
Name string - The
dnsName
is the name that you use in the applications of client tasks to connect to this service. The name must be a valid DNS name but doesn't need to be fully-qualified. The name can include up to 127 characters. The name can include lowercase letters, numbers, underscores (_), hyphens (-), and periods (.). The name can't start with a hyphen. If this parameter isn't specified, the default value ofdiscoveryName.namespace
is used. If thediscoveryName
isn't specified, the port mapping name from the task definition is used inportName.namespace
. To avoid changing your applications in client Amazon ECS services, set this to the same name that the client application uses by default. For example, a few common names aredatabase
,db
, or the lowercase name of a database, such asmysql
orredis
. For more information, see Service Connect in the Amazon Elastic Container Service Developer Guide.
- port int
- The listening port number for the Service Connect proxy. This port is available inside of all of the tasks within the same namespace. To avoid changing your applications in client Amazon ECS services, set this to the same port that the client application uses by default. For more information, see Service Connect in the Amazon Elastic Container Service Developer Guide.
- dns_
name str - The
dnsName
is the name that you use in the applications of client tasks to connect to this service. The name must be a valid DNS name but doesn't need to be fully-qualified. The name can include up to 127 characters. The name can include lowercase letters, numbers, underscores (_), hyphens (-), and periods (.). The name can't start with a hyphen. If this parameter isn't specified, the default value ofdiscoveryName.namespace
is used. If thediscoveryName
isn't specified, the port mapping name from the task definition is used inportName.namespace
. To avoid changing your applications in client Amazon ECS services, set this to the same name that the client application uses by default. For example, a few common names aredatabase
,db
, or the lowercase name of a database, such asmysql
orredis
. For more information, see Service Connect in the Amazon Elastic Container Service Developer Guide.
- port Number
- The listening port number for the Service Connect proxy. This port is available inside of all of the tasks within the same namespace. To avoid changing your applications in client Amazon ECS services, set this to the same port that the client application uses by default. For more information, see Service Connect in the Amazon Elastic Container Service Developer Guide.
- dns
Name String - The
dnsName
is the name that you use in the applications of client tasks to connect to this service. The name must be a valid DNS name but doesn't need to be fully-qualified. The name can include up to 127 characters. The name can include lowercase letters, numbers, underscores (_), hyphens (-), and periods (.). The name can't start with a hyphen. If this parameter isn't specified, the default value ofdiscoveryName.namespace
is used. If thediscoveryName
isn't specified, the port mapping name from the task definition is used inportName.namespace
. To avoid changing your applications in client Amazon ECS services, set this to the same name that the client application uses by default. For example, a few common names aredatabase
,db
, or the lowercase name of a database, such asmysql
orredis
. For more information, see Service Connect in the Amazon Elastic Container Service Developer Guide.
ServiceConnectConfiguration, ServiceConnectConfigurationArgs
- Enabled bool
- Specifies whether to use Service Connect with this service.
- Log
Configuration Pulumi.Aws Native. Ecs. Inputs. Service Log Configuration - The log configuration for the container. This parameter maps to
LogConfig
in the docker container create command and the--log-driver
option to docker run. By default, containers use the same logging driver that the Docker daemon uses. However, the container might use a different logging driver than the Docker daemon by specifying a log driver configuration in the container definition. Understand the following when specifying a log configuration for your containers.- Amazon ECS currently supports a subset of the logging drivers available to the Docker daemon. Additional log drivers may be available in future releases of the Amazon ECS container agent.
For tasks on FARGATElong, the supported log drivers are
awslogs
,splunk
, andawsfirelens
. For tasks hosted on Amazon EC2 instances, the supported log drivers areawslogs
,fluentd
,gelf
,json-file
,journald
,syslog
,splunk
, andawsfirelens
. - This parameter requires version 1.18 of the Docker Remote API or greater on your container instance.
- For tasks that are hosted on Amazon EC2 instances, the Amazon ECS container agent must register the available logging drivers with the
ECS_AVAILABLE_LOGGING_DRIVERS
environment variable before containers placed on that instance can use these log configuration options. For more information, see Amazon ECS container agent configuration in the Amazon Elastic Container Service Developer Guide. - For tasks that are on FARGATElong, because you don't have access to the underlying infrastructure your tasks are hosted on, any additional software needed must be installed outside of the task. For example, the Fluentd output aggregators or a remote host running Logstash to send Gelf logs to.
- Amazon ECS currently supports a subset of the logging drivers available to the Docker daemon. Additional log drivers may be available in future releases of the Amazon ECS container agent.
For tasks on FARGATElong, the supported log drivers are
- Namespace string
- The namespace name or full Amazon Resource Name (ARN) of the CMAPlong namespace for use with Service Connect. The namespace must be in the same AWS Region as the Amazon ECS service and cluster. The type of namespace doesn't affect Service Connect. For more information about CMAPlong, see Working with Services in the Developer Guide.
- Services
List<Pulumi.
Aws Native. Ecs. Inputs. Service Connect Service> - The list of Service Connect service objects. These are names and aliases (also known as endpoints) that are used by other Amazon ECS services to connect to this service. This field is not required for a "client" Amazon ECS service that's a member of a namespace only to connect to other services within the namespace. An example of this would be a frontend application that accepts incoming requests from either a load balancer that's attached to the service or by other means. An object selects a port from the task definition, assigns a name for the CMAPlong service, and a list of aliases (endpoints) and ports for client applications to refer to this service.
- Enabled bool
- Specifies whether to use Service Connect with this service.
- Log
Configuration ServiceLog Configuration - The log configuration for the container. This parameter maps to
LogConfig
in the docker container create command and the--log-driver
option to docker run. By default, containers use the same logging driver that the Docker daemon uses. However, the container might use a different logging driver than the Docker daemon by specifying a log driver configuration in the container definition. Understand the following when specifying a log configuration for your containers.- Amazon ECS currently supports a subset of the logging drivers available to the Docker daemon. Additional log drivers may be available in future releases of the Amazon ECS container agent.
For tasks on FARGATElong, the supported log drivers are
awslogs
,splunk
, andawsfirelens
. For tasks hosted on Amazon EC2 instances, the supported log drivers areawslogs
,fluentd
,gelf
,json-file
,journald
,syslog
,splunk
, andawsfirelens
. - This parameter requires version 1.18 of the Docker Remote API or greater on your container instance.
- For tasks that are hosted on Amazon EC2 instances, the Amazon ECS container agent must register the available logging drivers with the
ECS_AVAILABLE_LOGGING_DRIVERS
environment variable before containers placed on that instance can use these log configuration options. For more information, see Amazon ECS container agent configuration in the Amazon Elastic Container Service Developer Guide. - For tasks that are on FARGATElong, because you don't have access to the underlying infrastructure your tasks are hosted on, any additional software needed must be installed outside of the task. For example, the Fluentd output aggregators or a remote host running Logstash to send Gelf logs to.
- Amazon ECS currently supports a subset of the logging drivers available to the Docker daemon. Additional log drivers may be available in future releases of the Amazon ECS container agent.
For tasks on FARGATElong, the supported log drivers are
- Namespace string
- The namespace name or full Amazon Resource Name (ARN) of the CMAPlong namespace for use with Service Connect. The namespace must be in the same AWS Region as the Amazon ECS service and cluster. The type of namespace doesn't affect Service Connect. For more information about CMAPlong, see Working with Services in the Developer Guide.
- Services
[]Service
Connect Service - The list of Service Connect service objects. These are names and aliases (also known as endpoints) that are used by other Amazon ECS services to connect to this service. This field is not required for a "client" Amazon ECS service that's a member of a namespace only to connect to other services within the namespace. An example of this would be a frontend application that accepts incoming requests from either a load balancer that's attached to the service or by other means. An object selects a port from the task definition, assigns a name for the CMAPlong service, and a list of aliases (endpoints) and ports for client applications to refer to this service.
- enabled Boolean
- Specifies whether to use Service Connect with this service.
- log
Configuration ServiceLog Configuration - The log configuration for the container. This parameter maps to
LogConfig
in the docker container create command and the--log-driver
option to docker run. By default, containers use the same logging driver that the Docker daemon uses. However, the container might use a different logging driver than the Docker daemon by specifying a log driver configuration in the container definition. Understand the following when specifying a log configuration for your containers.- Amazon ECS currently supports a subset of the logging drivers available to the Docker daemon. Additional log drivers may be available in future releases of the Amazon ECS container agent.
For tasks on FARGATElong, the supported log drivers are
awslogs
,splunk
, andawsfirelens
. For tasks hosted on Amazon EC2 instances, the supported log drivers areawslogs
,fluentd
,gelf
,json-file
,journald
,syslog
,splunk
, andawsfirelens
. - This parameter requires version 1.18 of the Docker Remote API or greater on your container instance.
- For tasks that are hosted on Amazon EC2 instances, the Amazon ECS container agent must register the available logging drivers with the
ECS_AVAILABLE_LOGGING_DRIVERS
environment variable before containers placed on that instance can use these log configuration options. For more information, see Amazon ECS container agent configuration in the Amazon Elastic Container Service Developer Guide. - For tasks that are on FARGATElong, because you don't have access to the underlying infrastructure your tasks are hosted on, any additional software needed must be installed outside of the task. For example, the Fluentd output aggregators or a remote host running Logstash to send Gelf logs to.
- Amazon ECS currently supports a subset of the logging drivers available to the Docker daemon. Additional log drivers may be available in future releases of the Amazon ECS container agent.
For tasks on FARGATElong, the supported log drivers are
- namespace String
- The namespace name or full Amazon Resource Name (ARN) of the CMAPlong namespace for use with Service Connect. The namespace must be in the same AWS Region as the Amazon ECS service and cluster. The type of namespace doesn't affect Service Connect. For more information about CMAPlong, see Working with Services in the Developer Guide.
- services
List<Service
Connect Service> - The list of Service Connect service objects. These are names and aliases (also known as endpoints) that are used by other Amazon ECS services to connect to this service. This field is not required for a "client" Amazon ECS service that's a member of a namespace only to connect to other services within the namespace. An example of this would be a frontend application that accepts incoming requests from either a load balancer that's attached to the service or by other means. An object selects a port from the task definition, assigns a name for the CMAPlong service, and a list of aliases (endpoints) and ports for client applications to refer to this service.
- enabled boolean
- Specifies whether to use Service Connect with this service.
- log
Configuration ServiceLog Configuration - The log configuration for the container. This parameter maps to
LogConfig
in the docker container create command and the--log-driver
option to docker run. By default, containers use the same logging driver that the Docker daemon uses. However, the container might use a different logging driver than the Docker daemon by specifying a log driver configuration in the container definition. Understand the following when specifying a log configuration for your containers.- Amazon ECS currently supports a subset of the logging drivers available to the Docker daemon. Additional log drivers may be available in future releases of the Amazon ECS container agent.
For tasks on FARGATElong, the supported log drivers are
awslogs
,splunk
, andawsfirelens
. For tasks hosted on Amazon EC2 instances, the supported log drivers areawslogs
,fluentd
,gelf
,json-file
,journald
,syslog
,splunk
, andawsfirelens
. - This parameter requires version 1.18 of the Docker Remote API or greater on your container instance.
- For tasks that are hosted on Amazon EC2 instances, the Amazon ECS container agent must register the available logging drivers with the
ECS_AVAILABLE_LOGGING_DRIVERS
environment variable before containers placed on that instance can use these log configuration options. For more information, see Amazon ECS container agent configuration in the Amazon Elastic Container Service Developer Guide. - For tasks that are on FARGATElong, because you don't have access to the underlying infrastructure your tasks are hosted on, any additional software needed must be installed outside of the task. For example, the Fluentd output aggregators or a remote host running Logstash to send Gelf logs to.
- Amazon ECS currently supports a subset of the logging drivers available to the Docker daemon. Additional log drivers may be available in future releases of the Amazon ECS container agent.
For tasks on FARGATElong, the supported log drivers are
- namespace string
- The namespace name or full Amazon Resource Name (ARN) of the CMAPlong namespace for use with Service Connect. The namespace must be in the same AWS Region as the Amazon ECS service and cluster. The type of namespace doesn't affect Service Connect. For more information about CMAPlong, see Working with Services in the Developer Guide.
- services
Service
Connect Service[] - The list of Service Connect service objects. These are names and aliases (also known as endpoints) that are used by other Amazon ECS services to connect to this service. This field is not required for a "client" Amazon ECS service that's a member of a namespace only to connect to other services within the namespace. An example of this would be a frontend application that accepts incoming requests from either a load balancer that's attached to the service or by other means. An object selects a port from the task definition, assigns a name for the CMAPlong service, and a list of aliases (endpoints) and ports for client applications to refer to this service.
- enabled bool
- Specifies whether to use Service Connect with this service.
- log_
configuration ServiceLog Configuration - The log configuration for the container. This parameter maps to
LogConfig
in the docker container create command and the--log-driver
option to docker run. By default, containers use the same logging driver that the Docker daemon uses. However, the container might use a different logging driver than the Docker daemon by specifying a log driver configuration in the container definition. Understand the following when specifying a log configuration for your containers.- Amazon ECS currently supports a subset of the logging drivers available to the Docker daemon. Additional log drivers may be available in future releases of the Amazon ECS container agent.
For tasks on FARGATElong, the supported log drivers are
awslogs
,splunk
, andawsfirelens
. For tasks hosted on Amazon EC2 instances, the supported log drivers areawslogs
,fluentd
,gelf
,json-file
,journald
,syslog
,splunk
, andawsfirelens
. - This parameter requires version 1.18 of the Docker Remote API or greater on your container instance.
- For tasks that are hosted on Amazon EC2 instances, the Amazon ECS container agent must register the available logging drivers with the
ECS_AVAILABLE_LOGGING_DRIVERS
environment variable before containers placed on that instance can use these log configuration options. For more information, see Amazon ECS container agent configuration in the Amazon Elastic Container Service Developer Guide. - For tasks that are on FARGATElong, because you don't have access to the underlying infrastructure your tasks are hosted on, any additional software needed must be installed outside of the task. For example, the Fluentd output aggregators or a remote host running Logstash to send Gelf logs to.
- Amazon ECS currently supports a subset of the logging drivers available to the Docker daemon. Additional log drivers may be available in future releases of the Amazon ECS container agent.
For tasks on FARGATElong, the supported log drivers are
- namespace str
- The namespace name or full Amazon Resource Name (ARN) of the CMAPlong namespace for use with Service Connect. The namespace must be in the same AWS Region as the Amazon ECS service and cluster. The type of namespace doesn't affect Service Connect. For more information about CMAPlong, see Working with Services in the Developer Guide.
- services
Sequence[Service
Connect Service] - The list of Service Connect service objects. These are names and aliases (also known as endpoints) that are used by other Amazon ECS services to connect to this service. This field is not required for a "client" Amazon ECS service that's a member of a namespace only to connect to other services within the namespace. An example of this would be a frontend application that accepts incoming requests from either a load balancer that's attached to the service or by other means. An object selects a port from the task definition, assigns a name for the CMAPlong service, and a list of aliases (endpoints) and ports for client applications to refer to this service.
- enabled Boolean
- Specifies whether to use Service Connect with this service.
- log
Configuration Property Map - The log configuration for the container. This parameter maps to
LogConfig
in the docker container create command and the--log-driver
option to docker run. By default, containers use the same logging driver that the Docker daemon uses. However, the container might use a different logging driver than the Docker daemon by specifying a log driver configuration in the container definition. Understand the following when specifying a log configuration for your containers.- Amazon ECS currently supports a subset of the logging drivers available to the Docker daemon. Additional log drivers may be available in future releases of the Amazon ECS container agent.
For tasks on FARGATElong, the supported log drivers are
awslogs
,splunk
, andawsfirelens
. For tasks hosted on Amazon EC2 instances, the supported log drivers areawslogs
,fluentd
,gelf
,json-file
,journald
,syslog
,splunk
, andawsfirelens
. - This parameter requires version 1.18 of the Docker Remote API or greater on your container instance.
- For tasks that are hosted on Amazon EC2 instances, the Amazon ECS container agent must register the available logging drivers with the
ECS_AVAILABLE_LOGGING_DRIVERS
environment variable before containers placed on that instance can use these log configuration options. For more information, see Amazon ECS container agent configuration in the Amazon Elastic Container Service Developer Guide. - For tasks that are on FARGATElong, because you don't have access to the underlying infrastructure your tasks are hosted on, any additional software needed must be installed outside of the task. For example, the Fluentd output aggregators or a remote host running Logstash to send Gelf logs to.
- Amazon ECS currently supports a subset of the logging drivers available to the Docker daemon. Additional log drivers may be available in future releases of the Amazon ECS container agent.
For tasks on FARGATElong, the supported log drivers are
- namespace String
- The namespace name or full Amazon Resource Name (ARN) of the CMAPlong namespace for use with Service Connect. The namespace must be in the same AWS Region as the Amazon ECS service and cluster. The type of namespace doesn't affect Service Connect. For more information about CMAPlong, see Working with Services in the Developer Guide.
- services List<Property Map>
- The list of Service Connect service objects. These are names and aliases (also known as endpoints) that are used by other Amazon ECS services to connect to this service. This field is not required for a "client" Amazon ECS service that's a member of a namespace only to connect to other services within the namespace. An example of this would be a frontend application that accepts incoming requests from either a load balancer that's attached to the service or by other means. An object selects a port from the task definition, assigns a name for the CMAPlong service, and a list of aliases (endpoints) and ports for client applications to refer to this service.
ServiceConnectService, ServiceConnectServiceArgs
- Port
Name string - The
portName
must match the name of one of theportMappings
from all the containers in the task definition of this Amazon ECS service. - Client
Aliases List<Pulumi.Aws Native. Ecs. Inputs. Service Connect Client Alias> - The list of client aliases for this Service Connect service. You use these to assign names that can be used by client applications. The maximum number of client aliases that you can have in this list is 1.
Each alias ("endpoint") is a fully-qualified name and port number that other Amazon ECS tasks ("clients") can use to connect to this service.
Each name and port mapping must be unique within the namespace.
For each
ServiceConnectService
, you must provide at least oneclientAlias
with oneport
. - Discovery
Name string - The
discoveryName
is the name of the new CMAP service that Amazon ECS creates for this Amazon ECS service. This must be unique within the CMAP namespace. The name can contain up to 64 characters. The name can include lowercase letters, numbers, underscores (_), and hyphens (-). The name can't start with a hyphen. If thediscoveryName
isn't specified, the port mapping name from the task definition is used inportName.namespace
. - Ingress
Port intOverride - The port number for the Service Connect proxy to listen on.
Use the value of this field to bypass the proxy for traffic on the port number specified in the named
portMapping
in the task definition of this application, and then use it in your VPC security groups to allow traffic into the proxy for this Amazon ECS service. Inawsvpc
mode and Fargate, the default value is the container port number. The container port number is in theportMapping
in the task definition. In bridge mode, the default value is the ephemeral port of the Service Connect proxy. - Timeout
Pulumi.
Aws Native. Ecs. Inputs. Service Timeout Configuration - A reference to an object that represents the configured timeouts for Service Connect.
- Tls
Pulumi.
Aws Native. Ecs. Inputs. Service Connect Tls Configuration - A reference to an object that represents a Transport Layer Security (TLS) configuration.
- Port
Name string - The
portName
must match the name of one of theportMappings
from all the containers in the task definition of this Amazon ECS service. - Client
Aliases []ServiceConnect Client Alias - The list of client aliases for this Service Connect service. You use these to assign names that can be used by client applications. The maximum number of client aliases that you can have in this list is 1.
Each alias ("endpoint") is a fully-qualified name and port number that other Amazon ECS tasks ("clients") can use to connect to this service.
Each name and port mapping must be unique within the namespace.
For each
ServiceConnectService
, you must provide at least oneclientAlias
with oneport
. - Discovery
Name string - The
discoveryName
is the name of the new CMAP service that Amazon ECS creates for this Amazon ECS service. This must be unique within the CMAP namespace. The name can contain up to 64 characters. The name can include lowercase letters, numbers, underscores (_), and hyphens (-). The name can't start with a hyphen. If thediscoveryName
isn't specified, the port mapping name from the task definition is used inportName.namespace
. - Ingress
Port intOverride - The port number for the Service Connect proxy to listen on.
Use the value of this field to bypass the proxy for traffic on the port number specified in the named
portMapping
in the task definition of this application, and then use it in your VPC security groups to allow traffic into the proxy for this Amazon ECS service. Inawsvpc
mode and Fargate, the default value is the container port number. The container port number is in theportMapping
in the task definition. In bridge mode, the default value is the ephemeral port of the Service Connect proxy. - Timeout
Service
Timeout Configuration - A reference to an object that represents the configured timeouts for Service Connect.
- Tls
Service
Connect Tls Configuration - A reference to an object that represents a Transport Layer Security (TLS) configuration.
- port
Name String - The
portName
must match the name of one of theportMappings
from all the containers in the task definition of this Amazon ECS service. - client
Aliases List<ServiceConnect Client Alias> - The list of client aliases for this Service Connect service. You use these to assign names that can be used by client applications. The maximum number of client aliases that you can have in this list is 1.
Each alias ("endpoint") is a fully-qualified name and port number that other Amazon ECS tasks ("clients") can use to connect to this service.
Each name and port mapping must be unique within the namespace.
For each
ServiceConnectService
, you must provide at least oneclientAlias
with oneport
. - discovery
Name String - The
discoveryName
is the name of the new CMAP service that Amazon ECS creates for this Amazon ECS service. This must be unique within the CMAP namespace. The name can contain up to 64 characters. The name can include lowercase letters, numbers, underscores (_), and hyphens (-). The name can't start with a hyphen. If thediscoveryName
isn't specified, the port mapping name from the task definition is used inportName.namespace
. - ingress
Port IntegerOverride - The port number for the Service Connect proxy to listen on.
Use the value of this field to bypass the proxy for traffic on the port number specified in the named
portMapping
in the task definition of this application, and then use it in your VPC security groups to allow traffic into the proxy for this Amazon ECS service. Inawsvpc
mode and Fargate, the default value is the container port number. The container port number is in theportMapping
in the task definition. In bridge mode, the default value is the ephemeral port of the Service Connect proxy. - timeout
Service
Timeout Configuration - A reference to an object that represents the configured timeouts for Service Connect.
- tls
Service
Connect Tls Configuration - A reference to an object that represents a Transport Layer Security (TLS) configuration.
- port
Name string - The
portName
must match the name of one of theportMappings
from all the containers in the task definition of this Amazon ECS service. - client
Aliases ServiceConnect Client Alias[] - The list of client aliases for this Service Connect service. You use these to assign names that can be used by client applications. The maximum number of client aliases that you can have in this list is 1.
Each alias ("endpoint") is a fully-qualified name and port number that other Amazon ECS tasks ("clients") can use to connect to this service.
Each name and port mapping must be unique within the namespace.
For each
ServiceConnectService
, you must provide at least oneclientAlias
with oneport
. - discovery
Name string - The
discoveryName
is the name of the new CMAP service that Amazon ECS creates for this Amazon ECS service. This must be unique within the CMAP namespace. The name can contain up to 64 characters. The name can include lowercase letters, numbers, underscores (_), and hyphens (-). The name can't start with a hyphen. If thediscoveryName
isn't specified, the port mapping name from the task definition is used inportName.namespace
. - ingress
Port numberOverride - The port number for the Service Connect proxy to listen on.
Use the value of this field to bypass the proxy for traffic on the port number specified in the named
portMapping
in the task definition of this application, and then use it in your VPC security groups to allow traffic into the proxy for this Amazon ECS service. Inawsvpc
mode and Fargate, the default value is the container port number. The container port number is in theportMapping
in the task definition. In bridge mode, the default value is the ephemeral port of the Service Connect proxy. - timeout
Service
Timeout Configuration - A reference to an object that represents the configured timeouts for Service Connect.
- tls
Service
Connect Tls Configuration - A reference to an object that represents a Transport Layer Security (TLS) configuration.
- port_
name str - The
portName
must match the name of one of theportMappings
from all the containers in the task definition of this Amazon ECS service. - client_
aliases Sequence[ServiceConnect Client Alias] - The list of client aliases for this Service Connect service. You use these to assign names that can be used by client applications. The maximum number of client aliases that you can have in this list is 1.
Each alias ("endpoint") is a fully-qualified name and port number that other Amazon ECS tasks ("clients") can use to connect to this service.
Each name and port mapping must be unique within the namespace.
For each
ServiceConnectService
, you must provide at least oneclientAlias
with oneport
. - discovery_
name str - The
discoveryName
is the name of the new CMAP service that Amazon ECS creates for this Amazon ECS service. This must be unique within the CMAP namespace. The name can contain up to 64 characters. The name can include lowercase letters, numbers, underscores (_), and hyphens (-). The name can't start with a hyphen. If thediscoveryName
isn't specified, the port mapping name from the task definition is used inportName.namespace
. - ingress_
port_ intoverride - The port number for the Service Connect proxy to listen on.
Use the value of this field to bypass the proxy for traffic on the port number specified in the named
portMapping
in the task definition of this application, and then use it in your VPC security groups to allow traffic into the proxy for this Amazon ECS service. Inawsvpc
mode and Fargate, the default value is the container port number. The container port number is in theportMapping
in the task definition. In bridge mode, the default value is the ephemeral port of the Service Connect proxy. - timeout
Service
Timeout Configuration - A reference to an object that represents the configured timeouts for Service Connect.
- tls
Service
Connect Tls Configuration - A reference to an object that represents a Transport Layer Security (TLS) configuration.
- port
Name String - The
portName
must match the name of one of theportMappings
from all the containers in the task definition of this Amazon ECS service. - client
Aliases List<Property Map> - The list of client aliases for this Service Connect service. You use these to assign names that can be used by client applications. The maximum number of client aliases that you can have in this list is 1.
Each alias ("endpoint") is a fully-qualified name and port number that other Amazon ECS tasks ("clients") can use to connect to this service.
Each name and port mapping must be unique within the namespace.
For each
ServiceConnectService
, you must provide at least oneclientAlias
with oneport
. - discovery
Name String - The
discoveryName
is the name of the new CMAP service that Amazon ECS creates for this Amazon ECS service. This must be unique within the CMAP namespace. The name can contain up to 64 characters. The name can include lowercase letters, numbers, underscores (_), and hyphens (-). The name can't start with a hyphen. If thediscoveryName
isn't specified, the port mapping name from the task definition is used inportName.namespace
. - ingress
Port NumberOverride - The port number for the Service Connect proxy to listen on.
Use the value of this field to bypass the proxy for traffic on the port number specified in the named
portMapping
in the task definition of this application, and then use it in your VPC security groups to allow traffic into the proxy for this Amazon ECS service. Inawsvpc
mode and Fargate, the default value is the container port number. The container port number is in theportMapping
in the task definition. In bridge mode, the default value is the ephemeral port of the Service Connect proxy. - timeout Property Map
- A reference to an object that represents the configured timeouts for Service Connect.
- tls Property Map
- A reference to an object that represents a Transport Layer Security (TLS) configuration.
ServiceConnectTlsCertificateAuthority, ServiceConnectTlsCertificateAuthorityArgs
- string
- The ARN of the AWS Private Certificate Authority certificate.
- string
- The ARN of the AWS Private Certificate Authority certificate.
- String
- The ARN of the AWS Private Certificate Authority certificate.
- string
- The ARN of the AWS Private Certificate Authority certificate.
- str
- The ARN of the AWS Private Certificate Authority certificate.
- String
- The ARN of the AWS Private Certificate Authority certificate.
ServiceConnectTlsConfiguration, ServiceConnectTlsConfigurationArgs
- Pulumi.
Aws Native. Ecs. Inputs. Service Connect Tls Certificate Authority - The signer certificate authority.
- Kms
Key string - The AWS Key Management Service key.
- Role
Arn string - The Amazon Resource Name (ARN) of the IAM role that's associated with the Service Connect TLS.
- Service
Connect Tls Certificate Authority - The signer certificate authority.
- Kms
Key string - The AWS Key Management Service key.
- Role
Arn string - The Amazon Resource Name (ARN) of the IAM role that's associated with the Service Connect TLS.
- Service
Connect Tls Certificate Authority - The signer certificate authority.
- kms
Key String - The AWS Key Management Service key.
- role
Arn String - The Amazon Resource Name (ARN) of the IAM role that's associated with the Service Connect TLS.
- Service
Connect Tls Certificate Authority - The signer certificate authority.
- kms
Key string - The AWS Key Management Service key.
- role
Arn string - The Amazon Resource Name (ARN) of the IAM role that's associated with the Service Connect TLS.
- Service
Connect Tls Certificate Authority - The signer certificate authority.
- kms_
key str - The AWS Key Management Service key.
- role_
arn str - The Amazon Resource Name (ARN) of the IAM role that's associated with the Service Connect TLS.
- Property Map
- The signer certificate authority.
- kms
Key String - The AWS Key Management Service key.
- role
Arn String - The Amazon Resource Name (ARN) of the IAM role that's associated with the Service Connect TLS.
ServiceDeploymentAlarms, ServiceDeploymentAlarmsArgs
- Alarm
Names List<string> - One or more CloudWatch alarm names. Use a "," to separate the alarms.
- Enable bool
- Determines whether to use the CloudWatch alarm option in the service deployment process.
- Rollback bool
- Determines whether to configure Amazon ECS to roll back the service if a service deployment fails. If rollback is used, when a service deployment fails, the service is rolled back to the last deployment that completed successfully.
- Alarm
Names []string - One or more CloudWatch alarm names. Use a "," to separate the alarms.
- Enable bool
- Determines whether to use the CloudWatch alarm option in the service deployment process.
- Rollback bool
- Determines whether to configure Amazon ECS to roll back the service if a service deployment fails. If rollback is used, when a service deployment fails, the service is rolled back to the last deployment that completed successfully.
- alarm
Names List<String> - One or more CloudWatch alarm names. Use a "," to separate the alarms.
- enable Boolean
- Determines whether to use the CloudWatch alarm option in the service deployment process.
- rollback Boolean
- Determines whether to configure Amazon ECS to roll back the service if a service deployment fails. If rollback is used, when a service deployment fails, the service is rolled back to the last deployment that completed successfully.
- alarm
Names string[] - One or more CloudWatch alarm names. Use a "," to separate the alarms.
- enable boolean
- Determines whether to use the CloudWatch alarm option in the service deployment process.
- rollback boolean
- Determines whether to configure Amazon ECS to roll back the service if a service deployment fails. If rollback is used, when a service deployment fails, the service is rolled back to the last deployment that completed successfully.
- alarm_
names Sequence[str] - One or more CloudWatch alarm names. Use a "," to separate the alarms.
- enable bool
- Determines whether to use the CloudWatch alarm option in the service deployment process.
- rollback bool
- Determines whether to configure Amazon ECS to roll back the service if a service deployment fails. If rollback is used, when a service deployment fails, the service is rolled back to the last deployment that completed successfully.
- alarm
Names List<String> - One or more CloudWatch alarm names. Use a "," to separate the alarms.
- enable Boolean
- Determines whether to use the CloudWatch alarm option in the service deployment process.
- rollback Boolean
- Determines whether to configure Amazon ECS to roll back the service if a service deployment fails. If rollback is used, when a service deployment fails, the service is rolled back to the last deployment that completed successfully.
ServiceDeploymentCircuitBreaker, ServiceDeploymentCircuitBreakerArgs
- Enable bool
- Determines whether to use the deployment circuit breaker logic for the service.
- Rollback bool
- Determines whether to configure Amazon ECS to roll back the service if a service deployment fails. If rollback is on, when a service deployment fails, the service is rolled back to the last deployment that completed successfully.
- Enable bool
- Determines whether to use the deployment circuit breaker logic for the service.
- Rollback bool
- Determines whether to configure Amazon ECS to roll back the service if a service deployment fails. If rollback is on, when a service deployment fails, the service is rolled back to the last deployment that completed successfully.
- enable Boolean
- Determines whether to use the deployment circuit breaker logic for the service.
- rollback Boolean
- Determines whether to configure Amazon ECS to roll back the service if a service deployment fails. If rollback is on, when a service deployment fails, the service is rolled back to the last deployment that completed successfully.
- enable boolean
- Determines whether to use the deployment circuit breaker logic for the service.
- rollback boolean
- Determines whether to configure Amazon ECS to roll back the service if a service deployment fails. If rollback is on, when a service deployment fails, the service is rolled back to the last deployment that completed successfully.
- enable bool
- Determines whether to use the deployment circuit breaker logic for the service.
- rollback bool
- Determines whether to configure Amazon ECS to roll back the service if a service deployment fails. If rollback is on, when a service deployment fails, the service is rolled back to the last deployment that completed successfully.
- enable Boolean
- Determines whether to use the deployment circuit breaker logic for the service.
- rollback Boolean
- Determines whether to configure Amazon ECS to roll back the service if a service deployment fails. If rollback is on, when a service deployment fails, the service is rolled back to the last deployment that completed successfully.
ServiceDeploymentConfiguration, ServiceDeploymentConfigurationArgs
- Alarms
Pulumi.
Aws Native. Ecs. Inputs. Service Deployment Alarms - Information about the CloudWatch alarms.
- Deployment
Circuit Pulumi.Breaker Aws Native. Ecs. Inputs. Service Deployment Circuit Breaker - The deployment circuit breaker can only be used for services using the rolling update (
ECS
) deployment type. The deployment circuit breaker determines whether a service deployment will fail if the service can't reach a steady state. If you use the deployment circuit breaker, a service deployment will transition to a failed state and stop launching new tasks. If you use the rollback option, when a service deployment fails, the service is rolled back to the last deployment that completed successfully. For more information, see Rolling update in the Amazon Elastic Container Service Developer Guide - Maximum
Percent int - If a service is using the rolling update (
ECS
) deployment type, themaximumPercent
parameter represents an upper limit on the number of your service's tasks that are allowed in theRUNNING
orPENDING
state during a deployment, as a percentage of thedesiredCount
(rounded down to the nearest integer). This parameter enables you to define the deployment batch size. For example, if your service is using theREPLICA
service scheduler and has adesiredCount
of four tasks and amaximumPercent
value of 200%, the scheduler may start four new tasks before stopping the four older tasks (provided that the cluster resources required to do this are available). The defaultmaximumPercent
value for a service using theREPLICA
service scheduler is 200%. If a service is using either the blue/green (CODE_DEPLOY
) orEXTERNAL
deployment types, and tasks in the service use the EC2 launch type, the maximum percent value is set to the default value. The maximum percent value is used to define the upper limit on the number of the tasks in the service that remain in theRUNNING
state while the container instances are in theDRAINING
state. You can't specify a custommaximumPercent
value for a service that uses either the blue/green (CODE_DEPLOY
) orEXTERNAL
deployment types and has tasks that use the EC2 launch type. If the tasks in the service use the Fargate launch type, the maximum percent value is not used, although it is returned when describing your service. - Minimum
Healthy intPercent If a service is using the rolling update (
ECS
) deployment type, theminimumHealthyPercent
represents a lower limit on the number of your service's tasks that must remain in theRUNNING
state during a deployment, as a percentage of thedesiredCount
(rounded up to the nearest integer). This parameter enables you to deploy without using additional cluster capacity. For example, if your service has adesiredCount
of four tasks and aminimumHealthyPercent
of 50%, the service scheduler may stop two existing tasks to free up cluster capacity before starting two new tasks. For services that do not use a load balancer, the following should be noted:- A service is considered healthy if all essential containers within the tasks in the service pass their health checks.
- If a task has no essential containers with a health check defined, the service scheduler will wait for 40 seconds after a task reaches a
RUNNING
state before the task is counted towards the minimum healthy percent total. - If a task has one or more essential containers with a health check defined, the service scheduler will wait for the task to reach a healthy status before counting it towards the minimum healthy percent total. A task is considered healthy when all essential containers within the task have passed their health checks. The amount of time the service scheduler can wait for is determined by the container health check settings.
For services that do use a load balancer, the following should be noted:
- If a task has no essential containers with a health check defined, the service scheduler will wait for the load balancer target group health check to return a healthy status before counting the task towards the minimum healthy percent total.
- If a task has an essential container with a health check defined, the service scheduler will wait for both the task to reach a healthy status and the load balancer target group health check to return a healthy status before counting the task towards the minimum healthy percent total.
The default value for a replica service for
minimumHealthyPercent
is 100%. The defaultminimumHealthyPercent
value for a service using theDAEMON
service schedule is 0% for the CLI, the AWS SDKs, and the APIs and 50% for the AWS Management Console. The minimum number of healthy tasks during a deployment is thedesiredCount
multiplied by theminimumHealthyPercent
/100, rounded up to the nearest integer value. If a service is using either the blue/green (CODE_DEPLOY
) orEXTERNAL
deployment types and is running tasks that use the EC2 launch type, the minimum healthy percent value is set to the default value. The minimum healthy percent value is used to define the lower limit on the number of the tasks in the service that remain in theRUNNING
state while the container instances are in theDRAINING
state. You can't specify a customminimumHealthyPercent
value for a service that uses either the blue/green (CODE_DEPLOY
) orEXTERNAL
deployment types and has tasks that use the EC2 launch type. If a service is using either the blue/green (CODE_DEPLOY
) orEXTERNAL
deployment types and is running tasks that use the Fargate launch type, the minimum healthy percent value is not used, although it is returned when describing your service.
- Alarms
Service
Deployment Alarms - Information about the CloudWatch alarms.
- Deployment
Circuit ServiceBreaker Deployment Circuit Breaker - The deployment circuit breaker can only be used for services using the rolling update (
ECS
) deployment type. The deployment circuit breaker determines whether a service deployment will fail if the service can't reach a steady state. If you use the deployment circuit breaker, a service deployment will transition to a failed state and stop launching new tasks. If you use the rollback option, when a service deployment fails, the service is rolled back to the last deployment that completed successfully. For more information, see Rolling update in the Amazon Elastic Container Service Developer Guide - Maximum
Percent int - If a service is using the rolling update (
ECS
) deployment type, themaximumPercent
parameter represents an upper limit on the number of your service's tasks that are allowed in theRUNNING
orPENDING
state during a deployment, as a percentage of thedesiredCount
(rounded down to the nearest integer). This parameter enables you to define the deployment batch size. For example, if your service is using theREPLICA
service scheduler and has adesiredCount
of four tasks and amaximumPercent
value of 200%, the scheduler may start four new tasks before stopping the four older tasks (provided that the cluster resources required to do this are available). The defaultmaximumPercent
value for a service using theREPLICA
service scheduler is 200%. If a service is using either the blue/green (CODE_DEPLOY
) orEXTERNAL
deployment types, and tasks in the service use the EC2 launch type, the maximum percent value is set to the default value. The maximum percent value is used to define the upper limit on the number of the tasks in the service that remain in theRUNNING
state while the container instances are in theDRAINING
state. You can't specify a custommaximumPercent
value for a service that uses either the blue/green (CODE_DEPLOY
) orEXTERNAL
deployment types and has tasks that use the EC2 launch type. If the tasks in the service use the Fargate launch type, the maximum percent value is not used, although it is returned when describing your service. - Minimum
Healthy intPercent If a service is using the rolling update (
ECS
) deployment type, theminimumHealthyPercent
represents a lower limit on the number of your service's tasks that must remain in theRUNNING
state during a deployment, as a percentage of thedesiredCount
(rounded up to the nearest integer). This parameter enables you to deploy without using additional cluster capacity. For example, if your service has adesiredCount
of four tasks and aminimumHealthyPercent
of 50%, the service scheduler may stop two existing tasks to free up cluster capacity before starting two new tasks. For services that do not use a load balancer, the following should be noted:- A service is considered healthy if all essential containers within the tasks in the service pass their health checks.
- If a task has no essential containers with a health check defined, the service scheduler will wait for 40 seconds after a task reaches a
RUNNING
state before the task is counted towards the minimum healthy percent total. - If a task has one or more essential containers with a health check defined, the service scheduler will wait for the task to reach a healthy status before counting it towards the minimum healthy percent total. A task is considered healthy when all essential containers within the task have passed their health checks. The amount of time the service scheduler can wait for is determined by the container health check settings.
For services that do use a load balancer, the following should be noted:
- If a task has no essential containers with a health check defined, the service scheduler will wait for the load balancer target group health check to return a healthy status before counting the task towards the minimum healthy percent total.
- If a task has an essential container with a health check defined, the service scheduler will wait for both the task to reach a healthy status and the load balancer target group health check to return a healthy status before counting the task towards the minimum healthy percent total.
The default value for a replica service for
minimumHealthyPercent
is 100%. The defaultminimumHealthyPercent
value for a service using theDAEMON
service schedule is 0% for the CLI, the AWS SDKs, and the APIs and 50% for the AWS Management Console. The minimum number of healthy tasks during a deployment is thedesiredCount
multiplied by theminimumHealthyPercent
/100, rounded up to the nearest integer value. If a service is using either the blue/green (CODE_DEPLOY
) orEXTERNAL
deployment types and is running tasks that use the EC2 launch type, the minimum healthy percent value is set to the default value. The minimum healthy percent value is used to define the lower limit on the number of the tasks in the service that remain in theRUNNING
state while the container instances are in theDRAINING
state. You can't specify a customminimumHealthyPercent
value for a service that uses either the blue/green (CODE_DEPLOY
) orEXTERNAL
deployment types and has tasks that use the EC2 launch type. If a service is using either the blue/green (CODE_DEPLOY
) orEXTERNAL
deployment types and is running tasks that use the Fargate launch type, the minimum healthy percent value is not used, although it is returned when describing your service.
- alarms
Service
Deployment Alarms - Information about the CloudWatch alarms.
- deployment
Circuit ServiceBreaker Deployment Circuit Breaker - The deployment circuit breaker can only be used for services using the rolling update (
ECS
) deployment type. The deployment circuit breaker determines whether a service deployment will fail if the service can't reach a steady state. If you use the deployment circuit breaker, a service deployment will transition to a failed state and stop launching new tasks. If you use the rollback option, when a service deployment fails, the service is rolled back to the last deployment that completed successfully. For more information, see Rolling update in the Amazon Elastic Container Service Developer Guide - maximum
Percent Integer - If a service is using the rolling update (
ECS
) deployment type, themaximumPercent
parameter represents an upper limit on the number of your service's tasks that are allowed in theRUNNING
orPENDING
state during a deployment, as a percentage of thedesiredCount
(rounded down to the nearest integer). This parameter enables you to define the deployment batch size. For example, if your service is using theREPLICA
service scheduler and has adesiredCount
of four tasks and amaximumPercent
value of 200%, the scheduler may start four new tasks before stopping the four older tasks (provided that the cluster resources required to do this are available). The defaultmaximumPercent
value for a service using theREPLICA
service scheduler is 200%. If a service is using either the blue/green (CODE_DEPLOY
) orEXTERNAL
deployment types, and tasks in the service use the EC2 launch type, the maximum percent value is set to the default value. The maximum percent value is used to define the upper limit on the number of the tasks in the service that remain in theRUNNING
state while the container instances are in theDRAINING
state. You can't specify a custommaximumPercent
value for a service that uses either the blue/green (CODE_DEPLOY
) orEXTERNAL
deployment types and has tasks that use the EC2 launch type. If the tasks in the service use the Fargate launch type, the maximum percent value is not used, although it is returned when describing your service. - minimum
Healthy IntegerPercent If a service is using the rolling update (
ECS
) deployment type, theminimumHealthyPercent
represents a lower limit on the number of your service's tasks that must remain in theRUNNING
state during a deployment, as a percentage of thedesiredCount
(rounded up to the nearest integer). This parameter enables you to deploy without using additional cluster capacity. For example, if your service has adesiredCount
of four tasks and aminimumHealthyPercent
of 50%, the service scheduler may stop two existing tasks to free up cluster capacity before starting two new tasks. For services that do not use a load balancer, the following should be noted:- A service is considered healthy if all essential containers within the tasks in the service pass their health checks.
- If a task has no essential containers with a health check defined, the service scheduler will wait for 40 seconds after a task reaches a
RUNNING
state before the task is counted towards the minimum healthy percent total. - If a task has one or more essential containers with a health check defined, the service scheduler will wait for the task to reach a healthy status before counting it towards the minimum healthy percent total. A task is considered healthy when all essential containers within the task have passed their health checks. The amount of time the service scheduler can wait for is determined by the container health check settings.
For services that do use a load balancer, the following should be noted:
- If a task has no essential containers with a health check defined, the service scheduler will wait for the load balancer target group health check to return a healthy status before counting the task towards the minimum healthy percent total.
- If a task has an essential container with a health check defined, the service scheduler will wait for both the task to reach a healthy status and the load balancer target group health check to return a healthy status before counting the task towards the minimum healthy percent total.
The default value for a replica service for
minimumHealthyPercent
is 100%. The defaultminimumHealthyPercent
value for a service using theDAEMON
service schedule is 0% for the CLI, the AWS SDKs, and the APIs and 50% for the AWS Management Console. The minimum number of healthy tasks during a deployment is thedesiredCount
multiplied by theminimumHealthyPercent
/100, rounded up to the nearest integer value. If a service is using either the blue/green (CODE_DEPLOY
) orEXTERNAL
deployment types and is running tasks that use the EC2 launch type, the minimum healthy percent value is set to the default value. The minimum healthy percent value is used to define the lower limit on the number of the tasks in the service that remain in theRUNNING
state while the container instances are in theDRAINING
state. You can't specify a customminimumHealthyPercent
value for a service that uses either the blue/green (CODE_DEPLOY
) orEXTERNAL
deployment types and has tasks that use the EC2 launch type. If a service is using either the blue/green (CODE_DEPLOY
) orEXTERNAL
deployment types and is running tasks that use the Fargate launch type, the minimum healthy percent value is not used, although it is returned when describing your service.
- alarms
Service
Deployment Alarms - Information about the CloudWatch alarms.
- deployment
Circuit ServiceBreaker Deployment Circuit Breaker - The deployment circuit breaker can only be used for services using the rolling update (
ECS
) deployment type. The deployment circuit breaker determines whether a service deployment will fail if the service can't reach a steady state. If you use the deployment circuit breaker, a service deployment will transition to a failed state and stop launching new tasks. If you use the rollback option, when a service deployment fails, the service is rolled back to the last deployment that completed successfully. For more information, see Rolling update in the Amazon Elastic Container Service Developer Guide - maximum
Percent number - If a service is using the rolling update (
ECS
) deployment type, themaximumPercent
parameter represents an upper limit on the number of your service's tasks that are allowed in theRUNNING
orPENDING
state during a deployment, as a percentage of thedesiredCount
(rounded down to the nearest integer). This parameter enables you to define the deployment batch size. For example, if your service is using theREPLICA
service scheduler and has adesiredCount
of four tasks and amaximumPercent
value of 200%, the scheduler may start four new tasks before stopping the four older tasks (provided that the cluster resources required to do this are available). The defaultmaximumPercent
value for a service using theREPLICA
service scheduler is 200%. If a service is using either the blue/green (CODE_DEPLOY
) orEXTERNAL
deployment types, and tasks in the service use the EC2 launch type, the maximum percent value is set to the default value. The maximum percent value is used to define the upper limit on the number of the tasks in the service that remain in theRUNNING
state while the container instances are in theDRAINING
state. You can't specify a custommaximumPercent
value for a service that uses either the blue/green (CODE_DEPLOY
) orEXTERNAL
deployment types and has tasks that use the EC2 launch type. If the tasks in the service use the Fargate launch type, the maximum percent value is not used, although it is returned when describing your service. - minimum
Healthy numberPercent If a service is using the rolling update (
ECS
) deployment type, theminimumHealthyPercent
represents a lower limit on the number of your service's tasks that must remain in theRUNNING
state during a deployment, as a percentage of thedesiredCount
(rounded up to the nearest integer). This parameter enables you to deploy without using additional cluster capacity. For example, if your service has adesiredCount
of four tasks and aminimumHealthyPercent
of 50%, the service scheduler may stop two existing tasks to free up cluster capacity before starting two new tasks. For services that do not use a load balancer, the following should be noted:- A service is considered healthy if all essential containers within the tasks in the service pass their health checks.
- If a task has no essential containers with a health check defined, the service scheduler will wait for 40 seconds after a task reaches a
RUNNING
state before the task is counted towards the minimum healthy percent total. - If a task has one or more essential containers with a health check defined, the service scheduler will wait for the task to reach a healthy status before counting it towards the minimum healthy percent total. A task is considered healthy when all essential containers within the task have passed their health checks. The amount of time the service scheduler can wait for is determined by the container health check settings.
For services that do use a load balancer, the following should be noted:
- If a task has no essential containers with a health check defined, the service scheduler will wait for the load balancer target group health check to return a healthy status before counting the task towards the minimum healthy percent total.
- If a task has an essential container with a health check defined, the service scheduler will wait for both the task to reach a healthy status and the load balancer target group health check to return a healthy status before counting the task towards the minimum healthy percent total.
The default value for a replica service for
minimumHealthyPercent
is 100%. The defaultminimumHealthyPercent
value for a service using theDAEMON
service schedule is 0% for the CLI, the AWS SDKs, and the APIs and 50% for the AWS Management Console. The minimum number of healthy tasks during a deployment is thedesiredCount
multiplied by theminimumHealthyPercent
/100, rounded up to the nearest integer value. If a service is using either the blue/green (CODE_DEPLOY
) orEXTERNAL
deployment types and is running tasks that use the EC2 launch type, the minimum healthy percent value is set to the default value. The minimum healthy percent value is used to define the lower limit on the number of the tasks in the service that remain in theRUNNING
state while the container instances are in theDRAINING
state. You can't specify a customminimumHealthyPercent
value for a service that uses either the blue/green (CODE_DEPLOY
) orEXTERNAL
deployment types and has tasks that use the EC2 launch type. If a service is using either the blue/green (CODE_DEPLOY
) orEXTERNAL
deployment types and is running tasks that use the Fargate launch type, the minimum healthy percent value is not used, although it is returned when describing your service.
- alarms
Service
Deployment Alarms - Information about the CloudWatch alarms.
- deployment_
circuit_ Servicebreaker Deployment Circuit Breaker - The deployment circuit breaker can only be used for services using the rolling update (
ECS
) deployment type. The deployment circuit breaker determines whether a service deployment will fail if the service can't reach a steady state. If you use the deployment circuit breaker, a service deployment will transition to a failed state and stop launching new tasks. If you use the rollback option, when a service deployment fails, the service is rolled back to the last deployment that completed successfully. For more information, see Rolling update in the Amazon Elastic Container Service Developer Guide - maximum_
percent int - If a service is using the rolling update (
ECS
) deployment type, themaximumPercent
parameter represents an upper limit on the number of your service's tasks that are allowed in theRUNNING
orPENDING
state during a deployment, as a percentage of thedesiredCount
(rounded down to the nearest integer). This parameter enables you to define the deployment batch size. For example, if your service is using theREPLICA
service scheduler and has adesiredCount
of four tasks and amaximumPercent
value of 200%, the scheduler may start four new tasks before stopping the four older tasks (provided that the cluster resources required to do this are available). The defaultmaximumPercent
value for a service using theREPLICA
service scheduler is 200%. If a service is using either the blue/green (CODE_DEPLOY
) orEXTERNAL
deployment types, and tasks in the service use the EC2 launch type, the maximum percent value is set to the default value. The maximum percent value is used to define the upper limit on the number of the tasks in the service that remain in theRUNNING
state while the container instances are in theDRAINING
state. You can't specify a custommaximumPercent
value for a service that uses either the blue/green (CODE_DEPLOY
) orEXTERNAL
deployment types and has tasks that use the EC2 launch type. If the tasks in the service use the Fargate launch type, the maximum percent value is not used, although it is returned when describing your service. - minimum_
healthy_ intpercent If a service is using the rolling update (
ECS
) deployment type, theminimumHealthyPercent
represents a lower limit on the number of your service's tasks that must remain in theRUNNING
state during a deployment, as a percentage of thedesiredCount
(rounded up to the nearest integer). This parameter enables you to deploy without using additional cluster capacity. For example, if your service has adesiredCount
of four tasks and aminimumHealthyPercent
of 50%, the service scheduler may stop two existing tasks to free up cluster capacity before starting two new tasks. For services that do not use a load balancer, the following should be noted:- A service is considered healthy if all essential containers within the tasks in the service pass their health checks.
- If a task has no essential containers with a health check defined, the service scheduler will wait for 40 seconds after a task reaches a
RUNNING
state before the task is counted towards the minimum healthy percent total. - If a task has one or more essential containers with a health check defined, the service scheduler will wait for the task to reach a healthy status before counting it towards the minimum healthy percent total. A task is considered healthy when all essential containers within the task have passed their health checks. The amount of time the service scheduler can wait for is determined by the container health check settings.
For services that do use a load balancer, the following should be noted:
- If a task has no essential containers with a health check defined, the service scheduler will wait for the load balancer target group health check to return a healthy status before counting the task towards the minimum healthy percent total.
- If a task has an essential container with a health check defined, the service scheduler will wait for both the task to reach a healthy status and the load balancer target group health check to return a healthy status before counting the task towards the minimum healthy percent total.
The default value for a replica service for
minimumHealthyPercent
is 100%. The defaultminimumHealthyPercent
value for a service using theDAEMON
service schedule is 0% for the CLI, the AWS SDKs, and the APIs and 50% for the AWS Management Console. The minimum number of healthy tasks during a deployment is thedesiredCount
multiplied by theminimumHealthyPercent
/100, rounded up to the nearest integer value. If a service is using either the blue/green (CODE_DEPLOY
) orEXTERNAL
deployment types and is running tasks that use the EC2 launch type, the minimum healthy percent value is set to the default value. The minimum healthy percent value is used to define the lower limit on the number of the tasks in the service that remain in theRUNNING
state while the container instances are in theDRAINING
state. You can't specify a customminimumHealthyPercent
value for a service that uses either the blue/green (CODE_DEPLOY
) orEXTERNAL
deployment types and has tasks that use the EC2 launch type. If a service is using either the blue/green (CODE_DEPLOY
) orEXTERNAL
deployment types and is running tasks that use the Fargate launch type, the minimum healthy percent value is not used, although it is returned when describing your service.
- alarms Property Map
- Information about the CloudWatch alarms.
- deployment
Circuit Property MapBreaker - The deployment circuit breaker can only be used for services using the rolling update (
ECS
) deployment type. The deployment circuit breaker determines whether a service deployment will fail if the service can't reach a steady state. If you use the deployment circuit breaker, a service deployment will transition to a failed state and stop launching new tasks. If you use the rollback option, when a service deployment fails, the service is rolled back to the last deployment that completed successfully. For more information, see Rolling update in the Amazon Elastic Container Service Developer Guide - maximum
Percent Number - If a service is using the rolling update (
ECS
) deployment type, themaximumPercent
parameter represents an upper limit on the number of your service's tasks that are allowed in theRUNNING
orPENDING
state during a deployment, as a percentage of thedesiredCount
(rounded down to the nearest integer). This parameter enables you to define the deployment batch size. For example, if your service is using theREPLICA
service scheduler and has adesiredCount
of four tasks and amaximumPercent
value of 200%, the scheduler may start four new tasks before stopping the four older tasks (provided that the cluster resources required to do this are available). The defaultmaximumPercent
value for a service using theREPLICA
service scheduler is 200%. If a service is using either the blue/green (CODE_DEPLOY
) orEXTERNAL
deployment types, and tasks in the service use the EC2 launch type, the maximum percent value is set to the default value. The maximum percent value is used to define the upper limit on the number of the tasks in the service that remain in theRUNNING
state while the container instances are in theDRAINING
state. You can't specify a custommaximumPercent
value for a service that uses either the blue/green (CODE_DEPLOY
) orEXTERNAL
deployment types and has tasks that use the EC2 launch type. If the tasks in the service use the Fargate launch type, the maximum percent value is not used, although it is returned when describing your service. - minimum
Healthy NumberPercent If a service is using the rolling update (
ECS
) deployment type, theminimumHealthyPercent
represents a lower limit on the number of your service's tasks that must remain in theRUNNING
state during a deployment, as a percentage of thedesiredCount
(rounded up to the nearest integer). This parameter enables you to deploy without using additional cluster capacity. For example, if your service has adesiredCount
of four tasks and aminimumHealthyPercent
of 50%, the service scheduler may stop two existing tasks to free up cluster capacity before starting two new tasks. For services that do not use a load balancer, the following should be noted:- A service is considered healthy if all essential containers within the tasks in the service pass their health checks.
- If a task has no essential containers with a health check defined, the service scheduler will wait for 40 seconds after a task reaches a
RUNNING
state before the task is counted towards the minimum healthy percent total. - If a task has one or more essential containers with a health check defined, the service scheduler will wait for the task to reach a healthy status before counting it towards the minimum healthy percent total. A task is considered healthy when all essential containers within the task have passed their health checks. The amount of time the service scheduler can wait for is determined by the container health check settings.
For services that do use a load balancer, the following should be noted:
- If a task has no essential containers with a health check defined, the service scheduler will wait for the load balancer target group health check to return a healthy status before counting the task towards the minimum healthy percent total.
- If a task has an essential container with a health check defined, the service scheduler will wait for both the task to reach a healthy status and the load balancer target group health check to return a healthy status before counting the task towards the minimum healthy percent total.
The default value for a replica service for
minimumHealthyPercent
is 100%. The defaultminimumHealthyPercent
value for a service using theDAEMON
service schedule is 0% for the CLI, the AWS SDKs, and the APIs and 50% for the AWS Management Console. The minimum number of healthy tasks during a deployment is thedesiredCount
multiplied by theminimumHealthyPercent
/100, rounded up to the nearest integer value. If a service is using either the blue/green (CODE_DEPLOY
) orEXTERNAL
deployment types and is running tasks that use the EC2 launch type, the minimum healthy percent value is set to the default value. The minimum healthy percent value is used to define the lower limit on the number of the tasks in the service that remain in theRUNNING
state while the container instances are in theDRAINING
state. You can't specify a customminimumHealthyPercent
value for a service that uses either the blue/green (CODE_DEPLOY
) orEXTERNAL
deployment types and has tasks that use the EC2 launch type. If a service is using either the blue/green (CODE_DEPLOY
) orEXTERNAL
deployment types and is running tasks that use the Fargate launch type, the minimum healthy percent value is not used, although it is returned when describing your service.
ServiceDeploymentController, ServiceDeploymentControllerArgs
- Type
Pulumi.
Aws Native. Ecs. Service Deployment Controller Type - The deployment controller type to use. There are three deployment controller types available:
- ECS The rolling update (ECS) deployment type involves replacing the current running version of the container with the latest version. The number of containers Amazon ECS adds or removes from the service during a rolling update is controlled by adjusting the minimum and maximum number of healthy tasks allowed during a service deployment, as specified in the DeploymentConfiguration. + CODE_DEPLOY The blue/green (CODE_DEPLOY) deployment type uses the blue/green deployment model powered by , which allows you to verify a new deployment of a service before sending production traffic to it. + EXTERNAL The external (EXTERNAL) deployment type enables you to use any third-party deployment controller for full control over the deployment process for an Amazon ECS service.
- Type
Service
Deployment Controller Type - The deployment controller type to use. There are three deployment controller types available:
- ECS The rolling update (ECS) deployment type involves replacing the current running version of the container with the latest version. The number of containers Amazon ECS adds or removes from the service during a rolling update is controlled by adjusting the minimum and maximum number of healthy tasks allowed during a service deployment, as specified in the DeploymentConfiguration. + CODE_DEPLOY The blue/green (CODE_DEPLOY) deployment type uses the blue/green deployment model powered by , which allows you to verify a new deployment of a service before sending production traffic to it. + EXTERNAL The external (EXTERNAL) deployment type enables you to use any third-party deployment controller for full control over the deployment process for an Amazon ECS service.
- type
Service
Deployment Controller Type - The deployment controller type to use. There are three deployment controller types available:
- ECS The rolling update (ECS) deployment type involves replacing the current running version of the container with the latest version. The number of containers Amazon ECS adds or removes from the service during a rolling update is controlled by adjusting the minimum and maximum number of healthy tasks allowed during a service deployment, as specified in the DeploymentConfiguration. + CODE_DEPLOY The blue/green (CODE_DEPLOY) deployment type uses the blue/green deployment model powered by , which allows you to verify a new deployment of a service before sending production traffic to it. + EXTERNAL The external (EXTERNAL) deployment type enables you to use any third-party deployment controller for full control over the deployment process for an Amazon ECS service.
- type
Service
Deployment Controller Type - The deployment controller type to use. There are three deployment controller types available:
- ECS The rolling update (ECS) deployment type involves replacing the current running version of the container with the latest version. The number of containers Amazon ECS adds or removes from the service during a rolling update is controlled by adjusting the minimum and maximum number of healthy tasks allowed during a service deployment, as specified in the DeploymentConfiguration. + CODE_DEPLOY The blue/green (CODE_DEPLOY) deployment type uses the blue/green deployment model powered by , which allows you to verify a new deployment of a service before sending production traffic to it. + EXTERNAL The external (EXTERNAL) deployment type enables you to use any third-party deployment controller for full control over the deployment process for an Amazon ECS service.
- type
Service
Deployment Controller Type - The deployment controller type to use. There are three deployment controller types available:
- ECS The rolling update (ECS) deployment type involves replacing the current running version of the container with the latest version. The number of containers Amazon ECS adds or removes from the service during a rolling update is controlled by adjusting the minimum and maximum number of healthy tasks allowed during a service deployment, as specified in the DeploymentConfiguration. + CODE_DEPLOY The blue/green (CODE_DEPLOY) deployment type uses the blue/green deployment model powered by , which allows you to verify a new deployment of a service before sending production traffic to it. + EXTERNAL The external (EXTERNAL) deployment type enables you to use any third-party deployment controller for full control over the deployment process for an Amazon ECS service.
- type "CODE_DEPLOY" | "ECS" | "EXTERNAL"
- The deployment controller type to use. There are three deployment controller types available:
- ECS The rolling update (ECS) deployment type involves replacing the current running version of the container with the latest version. The number of containers Amazon ECS adds or removes from the service during a rolling update is controlled by adjusting the minimum and maximum number of healthy tasks allowed during a service deployment, as specified in the DeploymentConfiguration. + CODE_DEPLOY The blue/green (CODE_DEPLOY) deployment type uses the blue/green deployment model powered by , which allows you to verify a new deployment of a service before sending production traffic to it. + EXTERNAL The external (EXTERNAL) deployment type enables you to use any third-party deployment controller for full control over the deployment process for an Amazon ECS service.
ServiceDeploymentControllerType, ServiceDeploymentControllerTypeArgs
- Code
Deploy - CODE_DEPLOY
- Ecs
- ECS
- External
- EXTERNAL
- Service
Deployment Controller Type Code Deploy - CODE_DEPLOY
- Service
Deployment Controller Type Ecs - ECS
- Service
Deployment Controller Type External - EXTERNAL
- Code
Deploy - CODE_DEPLOY
- Ecs
- ECS
- External
- EXTERNAL
- Code
Deploy - CODE_DEPLOY
- Ecs
- ECS
- External
- EXTERNAL
- CODE_DEPLOY
- CODE_DEPLOY
- ECS
- ECS
- EXTERNAL
- EXTERNAL
- "CODE_DEPLOY"
- CODE_DEPLOY
- "ECS"
- ECS
- "EXTERNAL"
- EXTERNAL
ServiceEbsTagSpecification, ServiceEbsTagSpecificationArgs
- Resource
Type string - The type of volume resource.
- Pulumi.
Aws Native. Ecs. Service Ebs Tag Specification Propagate Tags - Determines whether to propagate the tags from the task definition to
the Amazon EBS volume. Tags can only propagate to a
SERVICE
specified inServiceVolumeConfiguration
. If no value is specified, the tags aren't propagated. - List<Pulumi.
Aws Native. Ecs. Inputs. Service Tag> - The tags applied to this Amazon EBS volume.
AmazonECSCreated
andAmazonECSManaged
are reserved tags that can't be used.
- Resource
Type string - The type of volume resource.
- Service
Ebs Tag Specification Propagate Tags - Determines whether to propagate the tags from the task definition to
the Amazon EBS volume. Tags can only propagate to a
SERVICE
specified inServiceVolumeConfiguration
. If no value is specified, the tags aren't propagated. - []Service
Tag - The tags applied to this Amazon EBS volume.
AmazonECSCreated
andAmazonECSManaged
are reserved tags that can't be used.
- resource
Type String - The type of volume resource.
- Service
Ebs Tag Specification Propagate Tags - Determines whether to propagate the tags from the task definition to
the Amazon EBS volume. Tags can only propagate to a
SERVICE
specified inServiceVolumeConfiguration
. If no value is specified, the tags aren't propagated. - List<Service
Tag> - The tags applied to this Amazon EBS volume.
AmazonECSCreated
andAmazonECSManaged
are reserved tags that can't be used.
- resource
Type string - The type of volume resource.
- Service
Ebs Tag Specification Propagate Tags - Determines whether to propagate the tags from the task definition to
the Amazon EBS volume. Tags can only propagate to a
SERVICE
specified inServiceVolumeConfiguration
. If no value is specified, the tags aren't propagated. - Service
Tag[] - The tags applied to this Amazon EBS volume.
AmazonECSCreated
andAmazonECSManaged
are reserved tags that can't be used.
- resource_
type str - The type of volume resource.
- Service
Ebs Tag Specification Propagate Tags - Determines whether to propagate the tags from the task definition to
the Amazon EBS volume. Tags can only propagate to a
SERVICE
specified inServiceVolumeConfiguration
. If no value is specified, the tags aren't propagated. - Sequence[Service
Tag] - The tags applied to this Amazon EBS volume.
AmazonECSCreated
andAmazonECSManaged
are reserved tags that can't be used.
- resource
Type String - The type of volume resource.
- "SERVICE" | "TASK_DEFINITION"
- Determines whether to propagate the tags from the task definition to
the Amazon EBS volume. Tags can only propagate to a
SERVICE
specified inServiceVolumeConfiguration
. If no value is specified, the tags aren't propagated. - List<Property Map>
- The tags applied to this Amazon EBS volume.
AmazonECSCreated
andAmazonECSManaged
are reserved tags that can't be used.
ServiceEbsTagSpecificationPropagateTags, ServiceEbsTagSpecificationPropagateTagsArgs
- Service
- SERVICE
- Task
Definition - TASK_DEFINITION
- Service
Ebs Tag Specification Propagate Tags Service - SERVICE
- Service
Ebs Tag Specification Propagate Tags Task Definition - TASK_DEFINITION
- Service
- SERVICE
- Task
Definition - TASK_DEFINITION
- Service
- SERVICE
- Task
Definition - TASK_DEFINITION
- SERVICE
- SERVICE
- TASK_DEFINITION
- TASK_DEFINITION
- "SERVICE"
- SERVICE
- "TASK_DEFINITION"
- TASK_DEFINITION
ServiceLaunchType, ServiceLaunchTypeArgs
- Ec2
- EC2
- Fargate
- FARGATE
- External
- EXTERNAL
- Service
Launch Type Ec2 - EC2
- Service
Launch Type Fargate - FARGATE
- Service
Launch Type External - EXTERNAL
- Ec2
- EC2
- Fargate
- FARGATE
- External
- EXTERNAL
- Ec2
- EC2
- Fargate
- FARGATE
- External
- EXTERNAL
- EC2
- EC2
- FARGATE
- FARGATE
- EXTERNAL
- EXTERNAL
- "EC2"
- EC2
- "FARGATE"
- FARGATE
- "EXTERNAL"
- EXTERNAL
ServiceLoadBalancer, ServiceLoadBalancerArgs
- Container
Name string - The name of the container (as it appears in a container definition) to associate with the load balancer. You need to specify the container name when configuring the target group for an Amazon ECS load balancer.
- Container
Port int - The port on the container to associate with the load balancer. This port must correspond to a
containerPort
in the task definition the tasks in the service are using. For tasks that use the EC2 launch type, the container instance they're launched on must allow ingress traffic on thehostPort
of the port mapping. - Load
Balancer stringName - The name of the load balancer to associate with the Amazon ECS service or task set. If you are using an Application Load Balancer or a Network Load Balancer the load balancer name parameter should be omitted.
- Target
Group stringArn - The full Amazon Resource Name (ARN) of the Elastic Load Balancing target group or groups associated with a service or task set.
A target group ARN is only specified when using an Application Load Balancer or Network Load Balancer.
For services using the
ECS
deployment controller, you can specify one or multiple target groups. For more information, see Registering multiple target groups with a service in the Amazon Elastic Container Service Developer Guide. For services using theCODE_DEPLOY
deployment controller, you're required to define two target groups for the load balancer. For more information, see Blue/green deployment with CodeDeploy in the Amazon Elastic Container Service Developer Guide. If your service's task definition uses theawsvpc
network mode, you must chooseip
as the target type, notinstance
. Do this when creating your target groups because tasks that use theawsvpc
network mode are associated with an elastic network interface, not an Amazon EC2 instance. This network mode is required for the Fargate launch type.
- Container
Name string - The name of the container (as it appears in a container definition) to associate with the load balancer. You need to specify the container name when configuring the target group for an Amazon ECS load balancer.
- Container
Port int - The port on the container to associate with the load balancer. This port must correspond to a
containerPort
in the task definition the tasks in the service are using. For tasks that use the EC2 launch type, the container instance they're launched on must allow ingress traffic on thehostPort
of the port mapping. - Load
Balancer stringName - The name of the load balancer to associate with the Amazon ECS service or task set. If you are using an Application Load Balancer or a Network Load Balancer the load balancer name parameter should be omitted.
- Target
Group stringArn - The full Amazon Resource Name (ARN) of the Elastic Load Balancing target group or groups associated with a service or task set.
A target group ARN is only specified when using an Application Load Balancer or Network Load Balancer.
For services using the
ECS
deployment controller, you can specify one or multiple target groups. For more information, see Registering multiple target groups with a service in the Amazon Elastic Container Service Developer Guide. For services using theCODE_DEPLOY
deployment controller, you're required to define two target groups for the load balancer. For more information, see Blue/green deployment with CodeDeploy in the Amazon Elastic Container Service Developer Guide. If your service's task definition uses theawsvpc
network mode, you must chooseip
as the target type, notinstance
. Do this when creating your target groups because tasks that use theawsvpc
network mode are associated with an elastic network interface, not an Amazon EC2 instance. This network mode is required for the Fargate launch type.
- container
Name String - The name of the container (as it appears in a container definition) to associate with the load balancer. You need to specify the container name when configuring the target group for an Amazon ECS load balancer.
- container
Port Integer - The port on the container to associate with the load balancer. This port must correspond to a
containerPort
in the task definition the tasks in the service are using. For tasks that use the EC2 launch type, the container instance they're launched on must allow ingress traffic on thehostPort
of the port mapping. - load
Balancer StringName - The name of the load balancer to associate with the Amazon ECS service or task set. If you are using an Application Load Balancer or a Network Load Balancer the load balancer name parameter should be omitted.
- target
Group StringArn - The full Amazon Resource Name (ARN) of the Elastic Load Balancing target group or groups associated with a service or task set.
A target group ARN is only specified when using an Application Load Balancer or Network Load Balancer.
For services using the
ECS
deployment controller, you can specify one or multiple target groups. For more information, see Registering multiple target groups with a service in the Amazon Elastic Container Service Developer Guide. For services using theCODE_DEPLOY
deployment controller, you're required to define two target groups for the load balancer. For more information, see Blue/green deployment with CodeDeploy in the Amazon Elastic Container Service Developer Guide. If your service's task definition uses theawsvpc
network mode, you must chooseip
as the target type, notinstance
. Do this when creating your target groups because tasks that use theawsvpc
network mode are associated with an elastic network interface, not an Amazon EC2 instance. This network mode is required for the Fargate launch type.
- container
Name string - The name of the container (as it appears in a container definition) to associate with the load balancer. You need to specify the container name when configuring the target group for an Amazon ECS load balancer.
- container
Port number - The port on the container to associate with the load balancer. This port must correspond to a
containerPort
in the task definition the tasks in the service are using. For tasks that use the EC2 launch type, the container instance they're launched on must allow ingress traffic on thehostPort
of the port mapping. - load
Balancer stringName - The name of the load balancer to associate with the Amazon ECS service or task set. If you are using an Application Load Balancer or a Network Load Balancer the load balancer name parameter should be omitted.
- target
Group stringArn - The full Amazon Resource Name (ARN) of the Elastic Load Balancing target group or groups associated with a service or task set.
A target group ARN is only specified when using an Application Load Balancer or Network Load Balancer.
For services using the
ECS
deployment controller, you can specify one or multiple target groups. For more information, see Registering multiple target groups with a service in the Amazon Elastic Container Service Developer Guide. For services using theCODE_DEPLOY
deployment controller, you're required to define two target groups for the load balancer. For more information, see Blue/green deployment with CodeDeploy in the Amazon Elastic Container Service Developer Guide. If your service's task definition uses theawsvpc
network mode, you must chooseip
as the target type, notinstance
. Do this when creating your target groups because tasks that use theawsvpc
network mode are associated with an elastic network interface, not an Amazon EC2 instance. This network mode is required for the Fargate launch type.
- container_
name str - The name of the container (as it appears in a container definition) to associate with the load balancer. You need to specify the container name when configuring the target group for an Amazon ECS load balancer.
- container_
port int - The port on the container to associate with the load balancer. This port must correspond to a
containerPort
in the task definition the tasks in the service are using. For tasks that use the EC2 launch type, the container instance they're launched on must allow ingress traffic on thehostPort
of the port mapping. - load_
balancer_ strname - The name of the load balancer to associate with the Amazon ECS service or task set. If you are using an Application Load Balancer or a Network Load Balancer the load balancer name parameter should be omitted.
- target_
group_ strarn - The full Amazon Resource Name (ARN) of the Elastic Load Balancing target group or groups associated with a service or task set.
A target group ARN is only specified when using an Application Load Balancer or Network Load Balancer.
For services using the
ECS
deployment controller, you can specify one or multiple target groups. For more information, see Registering multiple target groups with a service in the Amazon Elastic Container Service Developer Guide. For services using theCODE_DEPLOY
deployment controller, you're required to define two target groups for the load balancer. For more information, see Blue/green deployment with CodeDeploy in the Amazon Elastic Container Service Developer Guide. If your service's task definition uses theawsvpc
network mode, you must chooseip
as the target type, notinstance
. Do this when creating your target groups because tasks that use theawsvpc
network mode are associated with an elastic network interface, not an Amazon EC2 instance. This network mode is required for the Fargate launch type.
- container
Name String - The name of the container (as it appears in a container definition) to associate with the load balancer. You need to specify the container name when configuring the target group for an Amazon ECS load balancer.
- container
Port Number - The port on the container to associate with the load balancer. This port must correspond to a
containerPort
in the task definition the tasks in the service are using. For tasks that use the EC2 launch type, the container instance they're launched on must allow ingress traffic on thehostPort
of the port mapping. - load
Balancer StringName - The name of the load balancer to associate with the Amazon ECS service or task set. If you are using an Application Load Balancer or a Network Load Balancer the load balancer name parameter should be omitted.
- target
Group StringArn - The full Amazon Resource Name (ARN) of the Elastic Load Balancing target group or groups associated with a service or task set.
A target group ARN is only specified when using an Application Load Balancer or Network Load Balancer.
For services using the
ECS
deployment controller, you can specify one or multiple target groups. For more information, see Registering multiple target groups with a service in the Amazon Elastic Container Service Developer Guide. For services using theCODE_DEPLOY
deployment controller, you're required to define two target groups for the load balancer. For more information, see Blue/green deployment with CodeDeploy in the Amazon Elastic Container Service Developer Guide. If your service's task definition uses theawsvpc
network mode, you must chooseip
as the target type, notinstance
. Do this when creating your target groups because tasks that use theawsvpc
network mode are associated with an elastic network interface, not an Amazon EC2 instance. This network mode is required for the Fargate launch type.
ServiceLogConfiguration, ServiceLogConfigurationArgs
- Log
Driver string - The log driver to use for the container.
For tasks on FARGATElong, the supported log drivers are
awslogs
,splunk
, andawsfirelens
. For tasks hosted on Amazon EC2 instances, the supported log drivers areawslogs
,fluentd
,gelf
,json-file
,journald
,syslog
,splunk
, andawsfirelens
. For more information about using theawslogs
log driver, see Send Amazon ECS logs to CloudWatch in the Amazon Elastic Container Service Developer Guide. For more information about using theawsfirelens
log driver, see Send Amazon ECS logs to an service or Partner. If you have a custom driver that isn't listed, you can fork the Amazon ECS container agent project that's available on GitHub and customize it to work with that driver. We encourage you to submit pull requests for changes that you would like to have included. However, we don't currently provide support for running modified copies of this software. - Options Dictionary<string, string>
- The configuration options to send to the log driver.
The options you can specify depend on the log driver. Some of the options you can specify when you use the
awslogs
log driver to route logs to Amazon CloudWatch include the following:- awslogs-create-group Required: No Specify whether you want the log group to be created automatically. If this option isn't specified, it defaults to false. Your IAM policy must include the logs:CreateLogGroup permission before you attempt to use awslogs-create-group. + awslogs-region Required: Yes Specify the Region that the awslogs log driver is to send your Docker logs to. You can choose to send all of your logs from clusters in different Regions to a single region in CloudWatch Logs. This is so that they're all visible in one location. Otherwise, you can separate them by Region for more granularity. Make sure that the specified log group exists in the Region that you specify with this option. + awslogs-group Required: Yes Make sure to specify a log group that the awslogs log driver sends its log streams to. + awslogs-stream-prefix Required: Yes, when using the Fargate launch type.Optional for the EC2 launch type, required for the Fargate launch type. Use the awslogs-stream-prefix option to associate a log stream with the specified prefix, the container name, and the ID of the Amazon ECS task that the container belongs to. If you specify a prefix with this option, then the log stream takes the format prefix-name/container-name/ecs-task-id. If you don't specify a prefix with this option, then the log stream is named after the container ID that's assigned by the Docker daemon on the container instance. Because it's difficult to trace logs back to the container that sent them with just the Docker container ID (which is only available on the container instance), we recommend that you specify a prefix with this option. For Amazon ECS services, you can use the service name as the prefix. Doing so, you can trace log streams to the service that the container belongs to, the name of the container that sent them, and the ID of the task that the container belongs to. You must specify a stream-prefix for your logs to have your logs appear in the Log pane when using the Amazon ECS console. + awslogs-datetime-format Required: No This option defines a multiline start pattern in Python strftime format. A log message consists of a line that matches the pattern and any following lines that don’t match the pattern. The matched line is the delimiter between log messages. One example of a use case for using this format is for parsing output such as a stack dump, which might otherwise be logged in multiple entries. The correct pattern allows it to be captured in a single entry. For more information, see awslogs-datetime-format. You cannot configure both the awslogs-datetime-format and awslogs-multiline-pattern options. Multiline logging performs regular expression parsing and matching of all log messages. This might have a negative impact on logging performance. + awslogs-multiline-pattern Required: No This option defines a multiline start pattern that uses a regular expression. A log message consists of a line that matches the pattern and any following lines that don’t match the pattern. The matched line is the delimiter between log messages. For more information, see awslogs-multiline-pattern. This option is ignored if awslogs-datetime-format is also configured. You cannot configure both the awslogs-datetime-format and awslogs-multiline-pattern options. Multiline logging performs regular expression parsing and matching of all log messages. This might have a negative impact on logging performance. + mode Required: No Valid values: non-blocking | blocking This option defines the delivery mode of log messages from the container to CloudWatch Logs. The delivery mode you choose affects application availability when the flow of logs from container to CloudWatch is interrupted. If you use the blocking mode and the flow of logs to CloudWatch is interrupted, calls from container code to write to the stdout and stderr streams will block. The logging thread of the application will block as a result. This may cause the application to become unresponsive and lead to container healthcheck failure. If you use the non-blocking mode, the container's logs are instead stored in an in-memory intermediate buffer configured with the max-buffer-size option. This prevents the application from becoming unresponsive when logs cannot be sent to CloudWatch. We recommend using this mode if you want to ensure service availability and are okay with some log loss. For more information, see Preventing log loss with non-blocking mode in the awslogs container log driver. + max-buffer-size Required: No Default value: 1m When non-blocking mode is used, the max-buffer-size log option controls the size of the buffer that's used for intermediate message storage. Make sure to specify an adequate buffer size based on your application. When the buffer fills up, further logs cannot be stored. Logs that cannot be stored are lost.
To route logs using the
splunk
log router, you need to specify asplunk-token
and asplunk-url
. When you use theawsfirelens
log router to route logs to an AWS Service or AWS Partner Network destination for log storage and analytics, you can set thelog-driver-buffer-limit
option to limit the number of events that are buffered in memory, before being sent to the log router container. It can help to resolve potential log loss issue because high throughput might result in memory running out for the buffer inside of Docker. Other options you can specify when usingawsfirelens
to route logs depend on the destination. When you export logs to Amazon Data Firehose, you can specify the AWS Region withregion
and a name for the log stream withdelivery_stream
. When you export logs to Amazon Kinesis Data Streams, you can specify an AWS Region withregion
and a data stream name withstream
. When you export logs to Amazon OpenSearch Service, you can specify options likeName
,Host
(OpenSearch Service endpoint without protocol),Port
,Index
,Type
,Aws_auth
,Aws_region
,Suppress_Type_Name
, andtls
. When you export logs to Amazon S3, you can specify the bucket using thebucket
option. You can also specifyregion
,total_file_size
,upload_timeout
, anduse_put_object
as options. This parameter requires version 1.19 of the Docker Remote API or greater on your container instance. To check the Docker Remote API version on your container instance, log in to your container instance and run the following command:sudo docker version --format '{{.Server.APIVersion}}'
- awslogs-create-group Required: No Specify whether you want the log group to be created automatically. If this option isn't specified, it defaults to false. Your IAM policy must include the logs:CreateLogGroup permission before you attempt to use awslogs-create-group. + awslogs-region Required: Yes Specify the Region that the awslogs log driver is to send your Docker logs to. You can choose to send all of your logs from clusters in different Regions to a single region in CloudWatch Logs. This is so that they're all visible in one location. Otherwise, you can separate them by Region for more granularity. Make sure that the specified log group exists in the Region that you specify with this option. + awslogs-group Required: Yes Make sure to specify a log group that the awslogs log driver sends its log streams to. + awslogs-stream-prefix Required: Yes, when using the Fargate launch type.Optional for the EC2 launch type, required for the Fargate launch type. Use the awslogs-stream-prefix option to associate a log stream with the specified prefix, the container name, and the ID of the Amazon ECS task that the container belongs to. If you specify a prefix with this option, then the log stream takes the format prefix-name/container-name/ecs-task-id. If you don't specify a prefix with this option, then the log stream is named after the container ID that's assigned by the Docker daemon on the container instance. Because it's difficult to trace logs back to the container that sent them with just the Docker container ID (which is only available on the container instance), we recommend that you specify a prefix with this option. For Amazon ECS services, you can use the service name as the prefix. Doing so, you can trace log streams to the service that the container belongs to, the name of the container that sent them, and the ID of the task that the container belongs to. You must specify a stream-prefix for your logs to have your logs appear in the Log pane when using the Amazon ECS console. + awslogs-datetime-format Required: No This option defines a multiline start pattern in Python strftime format. A log message consists of a line that matches the pattern and any following lines that don’t match the pattern. The matched line is the delimiter between log messages. One example of a use case for using this format is for parsing output such as a stack dump, which might otherwise be logged in multiple entries. The correct pattern allows it to be captured in a single entry. For more information, see awslogs-datetime-format. You cannot configure both the awslogs-datetime-format and awslogs-multiline-pattern options. Multiline logging performs regular expression parsing and matching of all log messages. This might have a negative impact on logging performance. + awslogs-multiline-pattern Required: No This option defines a multiline start pattern that uses a regular expression. A log message consists of a line that matches the pattern and any following lines that don’t match the pattern. The matched line is the delimiter between log messages. For more information, see awslogs-multiline-pattern. This option is ignored if awslogs-datetime-format is also configured. You cannot configure both the awslogs-datetime-format and awslogs-multiline-pattern options. Multiline logging performs regular expression parsing and matching of all log messages. This might have a negative impact on logging performance. + mode Required: No Valid values: non-blocking | blocking This option defines the delivery mode of log messages from the container to CloudWatch Logs. The delivery mode you choose affects application availability when the flow of logs from container to CloudWatch is interrupted. If you use the blocking mode and the flow of logs to CloudWatch is interrupted, calls from container code to write to the stdout and stderr streams will block. The logging thread of the application will block as a result. This may cause the application to become unresponsive and lead to container healthcheck failure. If you use the non-blocking mode, the container's logs are instead stored in an in-memory intermediate buffer configured with the max-buffer-size option. This prevents the application from becoming unresponsive when logs cannot be sent to CloudWatch. We recommend using this mode if you want to ensure service availability and are okay with some log loss. For more information, see Preventing log loss with non-blocking mode in the awslogs container log driver. + max-buffer-size Required: No Default value: 1m When non-blocking mode is used, the max-buffer-size log option controls the size of the buffer that's used for intermediate message storage. Make sure to specify an adequate buffer size based on your application. When the buffer fills up, further logs cannot be stored. Logs that cannot be stored are lost.
To route logs using the
- Secret
Options List<Pulumi.Aws Native. Ecs. Inputs. Service Secret> - The secrets to pass to the log configuration. For more information, see Specifying sensitive data in the Amazon Elastic Container Service Developer Guide.
- Log
Driver string - The log driver to use for the container.
For tasks on FARGATElong, the supported log drivers are
awslogs
,splunk
, andawsfirelens
. For tasks hosted on Amazon EC2 instances, the supported log drivers areawslogs
,fluentd
,gelf
,json-file
,journald
,syslog
,splunk
, andawsfirelens
. For more information about using theawslogs
log driver, see Send Amazon ECS logs to CloudWatch in the Amazon Elastic Container Service Developer Guide. For more information about using theawsfirelens
log driver, see Send Amazon ECS logs to an service or Partner. If you have a custom driver that isn't listed, you can fork the Amazon ECS container agent project that's available on GitHub and customize it to work with that driver. We encourage you to submit pull requests for changes that you would like to have included. However, we don't currently provide support for running modified copies of this software. - Options map[string]string
- The configuration options to send to the log driver.
The options you can specify depend on the log driver. Some of the options you can specify when you use the
awslogs
log driver to route logs to Amazon CloudWatch include the following:- awslogs-create-group Required: No Specify whether you want the log group to be created automatically. If this option isn't specified, it defaults to false. Your IAM policy must include the logs:CreateLogGroup permission before you attempt to use awslogs-create-group. + awslogs-region Required: Yes Specify the Region that the awslogs log driver is to send your Docker logs to. You can choose to send all of your logs from clusters in different Regions to a single region in CloudWatch Logs. This is so that they're all visible in one location. Otherwise, you can separate them by Region for more granularity. Make sure that the specified log group exists in the Region that you specify with this option. + awslogs-group Required: Yes Make sure to specify a log group that the awslogs log driver sends its log streams to. + awslogs-stream-prefix Required: Yes, when using the Fargate launch type.Optional for the EC2 launch type, required for the Fargate launch type. Use the awslogs-stream-prefix option to associate a log stream with the specified prefix, the container name, and the ID of the Amazon ECS task that the container belongs to. If you specify a prefix with this option, then the log stream takes the format prefix-name/container-name/ecs-task-id. If you don't specify a prefix with this option, then the log stream is named after the container ID that's assigned by the Docker daemon on the container instance. Because it's difficult to trace logs back to the container that sent them with just the Docker container ID (which is only available on the container instance), we recommend that you specify a prefix with this option. For Amazon ECS services, you can use the service name as the prefix. Doing so, you can trace log streams to the service that the container belongs to, the name of the container that sent them, and the ID of the task that the container belongs to. You must specify a stream-prefix for your logs to have your logs appear in the Log pane when using the Amazon ECS console. + awslogs-datetime-format Required: No This option defines a multiline start pattern in Python strftime format. A log message consists of a line that matches the pattern and any following lines that don’t match the pattern. The matched line is the delimiter between log messages. One example of a use case for using this format is for parsing output such as a stack dump, which might otherwise be logged in multiple entries. The correct pattern allows it to be captured in a single entry. For more information, see awslogs-datetime-format. You cannot configure both the awslogs-datetime-format and awslogs-multiline-pattern options. Multiline logging performs regular expression parsing and matching of all log messages. This might have a negative impact on logging performance. + awslogs-multiline-pattern Required: No This option defines a multiline start pattern that uses a regular expression. A log message consists of a line that matches the pattern and any following lines that don’t match the pattern. The matched line is the delimiter between log messages. For more information, see awslogs-multiline-pattern. This option is ignored if awslogs-datetime-format is also configured. You cannot configure both the awslogs-datetime-format and awslogs-multiline-pattern options. Multiline logging performs regular expression parsing and matching of all log messages. This might have a negative impact on logging performance. + mode Required: No Valid values: non-blocking | blocking This option defines the delivery mode of log messages from the container to CloudWatch Logs. The delivery mode you choose affects application availability when the flow of logs from container to CloudWatch is interrupted. If you use the blocking mode and the flow of logs to CloudWatch is interrupted, calls from container code to write to the stdout and stderr streams will block. The logging thread of the application will block as a result. This may cause the application to become unresponsive and lead to container healthcheck failure. If you use the non-blocking mode, the container's logs are instead stored in an in-memory intermediate buffer configured with the max-buffer-size option. This prevents the application from becoming unresponsive when logs cannot be sent to CloudWatch. We recommend using this mode if you want to ensure service availability and are okay with some log loss. For more information, see Preventing log loss with non-blocking mode in the awslogs container log driver. + max-buffer-size Required: No Default value: 1m When non-blocking mode is used, the max-buffer-size log option controls the size of the buffer that's used for intermediate message storage. Make sure to specify an adequate buffer size based on your application. When the buffer fills up, further logs cannot be stored. Logs that cannot be stored are lost.
To route logs using the
splunk
log router, you need to specify asplunk-token
and asplunk-url
. When you use theawsfirelens
log router to route logs to an AWS Service or AWS Partner Network destination for log storage and analytics, you can set thelog-driver-buffer-limit
option to limit the number of events that are buffered in memory, before being sent to the log router container. It can help to resolve potential log loss issue because high throughput might result in memory running out for the buffer inside of Docker. Other options you can specify when usingawsfirelens
to route logs depend on the destination. When you export logs to Amazon Data Firehose, you can specify the AWS Region withregion
and a name for the log stream withdelivery_stream
. When you export logs to Amazon Kinesis Data Streams, you can specify an AWS Region withregion
and a data stream name withstream
. When you export logs to Amazon OpenSearch Service, you can specify options likeName
,Host
(OpenSearch Service endpoint without protocol),Port
,Index
,Type
,Aws_auth
,Aws_region
,Suppress_Type_Name
, andtls
. When you export logs to Amazon S3, you can specify the bucket using thebucket
option. You can also specifyregion
,total_file_size
,upload_timeout
, anduse_put_object
as options. This parameter requires version 1.19 of the Docker Remote API or greater on your container instance. To check the Docker Remote API version on your container instance, log in to your container instance and run the following command:sudo docker version --format '{{.Server.APIVersion}}'
- awslogs-create-group Required: No Specify whether you want the log group to be created automatically. If this option isn't specified, it defaults to false. Your IAM policy must include the logs:CreateLogGroup permission before you attempt to use awslogs-create-group. + awslogs-region Required: Yes Specify the Region that the awslogs log driver is to send your Docker logs to. You can choose to send all of your logs from clusters in different Regions to a single region in CloudWatch Logs. This is so that they're all visible in one location. Otherwise, you can separate them by Region for more granularity. Make sure that the specified log group exists in the Region that you specify with this option. + awslogs-group Required: Yes Make sure to specify a log group that the awslogs log driver sends its log streams to. + awslogs-stream-prefix Required: Yes, when using the Fargate launch type.Optional for the EC2 launch type, required for the Fargate launch type. Use the awslogs-stream-prefix option to associate a log stream with the specified prefix, the container name, and the ID of the Amazon ECS task that the container belongs to. If you specify a prefix with this option, then the log stream takes the format prefix-name/container-name/ecs-task-id. If you don't specify a prefix with this option, then the log stream is named after the container ID that's assigned by the Docker daemon on the container instance. Because it's difficult to trace logs back to the container that sent them with just the Docker container ID (which is only available on the container instance), we recommend that you specify a prefix with this option. For Amazon ECS services, you can use the service name as the prefix. Doing so, you can trace log streams to the service that the container belongs to, the name of the container that sent them, and the ID of the task that the container belongs to. You must specify a stream-prefix for your logs to have your logs appear in the Log pane when using the Amazon ECS console. + awslogs-datetime-format Required: No This option defines a multiline start pattern in Python strftime format. A log message consists of a line that matches the pattern and any following lines that don’t match the pattern. The matched line is the delimiter between log messages. One example of a use case for using this format is for parsing output such as a stack dump, which might otherwise be logged in multiple entries. The correct pattern allows it to be captured in a single entry. For more information, see awslogs-datetime-format. You cannot configure both the awslogs-datetime-format and awslogs-multiline-pattern options. Multiline logging performs regular expression parsing and matching of all log messages. This might have a negative impact on logging performance. + awslogs-multiline-pattern Required: No This option defines a multiline start pattern that uses a regular expression. A log message consists of a line that matches the pattern and any following lines that don’t match the pattern. The matched line is the delimiter between log messages. For more information, see awslogs-multiline-pattern. This option is ignored if awslogs-datetime-format is also configured. You cannot configure both the awslogs-datetime-format and awslogs-multiline-pattern options. Multiline logging performs regular expression parsing and matching of all log messages. This might have a negative impact on logging performance. + mode Required: No Valid values: non-blocking | blocking This option defines the delivery mode of log messages from the container to CloudWatch Logs. The delivery mode you choose affects application availability when the flow of logs from container to CloudWatch is interrupted. If you use the blocking mode and the flow of logs to CloudWatch is interrupted, calls from container code to write to the stdout and stderr streams will block. The logging thread of the application will block as a result. This may cause the application to become unresponsive and lead to container healthcheck failure. If you use the non-blocking mode, the container's logs are instead stored in an in-memory intermediate buffer configured with the max-buffer-size option. This prevents the application from becoming unresponsive when logs cannot be sent to CloudWatch. We recommend using this mode if you want to ensure service availability and are okay with some log loss. For more information, see Preventing log loss with non-blocking mode in the awslogs container log driver. + max-buffer-size Required: No Default value: 1m When non-blocking mode is used, the max-buffer-size log option controls the size of the buffer that's used for intermediate message storage. Make sure to specify an adequate buffer size based on your application. When the buffer fills up, further logs cannot be stored. Logs that cannot be stored are lost.
To route logs using the
- Secret
Options []ServiceSecret - The secrets to pass to the log configuration. For more information, see Specifying sensitive data in the Amazon Elastic Container Service Developer Guide.
- log
Driver String - The log driver to use for the container.
For tasks on FARGATElong, the supported log drivers are
awslogs
,splunk
, andawsfirelens
. For tasks hosted on Amazon EC2 instances, the supported log drivers areawslogs
,fluentd
,gelf
,json-file
,journald
,syslog
,splunk
, andawsfirelens
. For more information about using theawslogs
log driver, see Send Amazon ECS logs to CloudWatch in the Amazon Elastic Container Service Developer Guide. For more information about using theawsfirelens
log driver, see Send Amazon ECS logs to an service or Partner. If you have a custom driver that isn't listed, you can fork the Amazon ECS container agent project that's available on GitHub and customize it to work with that driver. We encourage you to submit pull requests for changes that you would like to have included. However, we don't currently provide support for running modified copies of this software. - options Map<String,String>
- The configuration options to send to the log driver.
The options you can specify depend on the log driver. Some of the options you can specify when you use the
awslogs
log driver to route logs to Amazon CloudWatch include the following:- awslogs-create-group Required: No Specify whether you want the log group to be created automatically. If this option isn't specified, it defaults to false. Your IAM policy must include the logs:CreateLogGroup permission before you attempt to use awslogs-create-group. + awslogs-region Required: Yes Specify the Region that the awslogs log driver is to send your Docker logs to. You can choose to send all of your logs from clusters in different Regions to a single region in CloudWatch Logs. This is so that they're all visible in one location. Otherwise, you can separate them by Region for more granularity. Make sure that the specified log group exists in the Region that you specify with this option. + awslogs-group Required: Yes Make sure to specify a log group that the awslogs log driver sends its log streams to. + awslogs-stream-prefix Required: Yes, when using the Fargate launch type.Optional for the EC2 launch type, required for the Fargate launch type. Use the awslogs-stream-prefix option to associate a log stream with the specified prefix, the container name, and the ID of the Amazon ECS task that the container belongs to. If you specify a prefix with this option, then the log stream takes the format prefix-name/container-name/ecs-task-id. If you don't specify a prefix with this option, then the log stream is named after the container ID that's assigned by the Docker daemon on the container instance. Because it's difficult to trace logs back to the container that sent them with just the Docker container ID (which is only available on the container instance), we recommend that you specify a prefix with this option. For Amazon ECS services, you can use the service name as the prefix. Doing so, you can trace log streams to the service that the container belongs to, the name of the container that sent them, and the ID of the task that the container belongs to. You must specify a stream-prefix for your logs to have your logs appear in the Log pane when using the Amazon ECS console. + awslogs-datetime-format Required: No This option defines a multiline start pattern in Python strftime format. A log message consists of a line that matches the pattern and any following lines that don’t match the pattern. The matched line is the delimiter between log messages. One example of a use case for using this format is for parsing output such as a stack dump, which might otherwise be logged in multiple entries. The correct pattern allows it to be captured in a single entry. For more information, see awslogs-datetime-format. You cannot configure both the awslogs-datetime-format and awslogs-multiline-pattern options. Multiline logging performs regular expression parsing and matching of all log messages. This might have a negative impact on logging performance. + awslogs-multiline-pattern Required: No This option defines a multiline start pattern that uses a regular expression. A log message consists of a line that matches the pattern and any following lines that don’t match the pattern. The matched line is the delimiter between log messages. For more information, see awslogs-multiline-pattern. This option is ignored if awslogs-datetime-format is also configured. You cannot configure both the awslogs-datetime-format and awslogs-multiline-pattern options. Multiline logging performs regular expression parsing and matching of all log messages. This might have a negative impact on logging performance. + mode Required: No Valid values: non-blocking | blocking This option defines the delivery mode of log messages from the container to CloudWatch Logs. The delivery mode you choose affects application availability when the flow of logs from container to CloudWatch is interrupted. If you use the blocking mode and the flow of logs to CloudWatch is interrupted, calls from container code to write to the stdout and stderr streams will block. The logging thread of the application will block as a result. This may cause the application to become unresponsive and lead to container healthcheck failure. If you use the non-blocking mode, the container's logs are instead stored in an in-memory intermediate buffer configured with the max-buffer-size option. This prevents the application from becoming unresponsive when logs cannot be sent to CloudWatch. We recommend using this mode if you want to ensure service availability and are okay with some log loss. For more information, see Preventing log loss with non-blocking mode in the awslogs container log driver. + max-buffer-size Required: No Default value: 1m When non-blocking mode is used, the max-buffer-size log option controls the size of the buffer that's used for intermediate message storage. Make sure to specify an adequate buffer size based on your application. When the buffer fills up, further logs cannot be stored. Logs that cannot be stored are lost.
To route logs using the
splunk
log router, you need to specify asplunk-token
and asplunk-url
. When you use theawsfirelens
log router to route logs to an AWS Service or AWS Partner Network destination for log storage and analytics, you can set thelog-driver-buffer-limit
option to limit the number of events that are buffered in memory, before being sent to the log router container. It can help to resolve potential log loss issue because high throughput might result in memory running out for the buffer inside of Docker. Other options you can specify when usingawsfirelens
to route logs depend on the destination. When you export logs to Amazon Data Firehose, you can specify the AWS Region withregion
and a name for the log stream withdelivery_stream
. When you export logs to Amazon Kinesis Data Streams, you can specify an AWS Region withregion
and a data stream name withstream
. When you export logs to Amazon OpenSearch Service, you can specify options likeName
,Host
(OpenSearch Service endpoint without protocol),Port
,Index
,Type
,Aws_auth
,Aws_region
,Suppress_Type_Name
, andtls
. When you export logs to Amazon S3, you can specify the bucket using thebucket
option. You can also specifyregion
,total_file_size
,upload_timeout
, anduse_put_object
as options. This parameter requires version 1.19 of the Docker Remote API or greater on your container instance. To check the Docker Remote API version on your container instance, log in to your container instance and run the following command:sudo docker version --format '{{.Server.APIVersion}}'
- awslogs-create-group Required: No Specify whether you want the log group to be created automatically. If this option isn't specified, it defaults to false. Your IAM policy must include the logs:CreateLogGroup permission before you attempt to use awslogs-create-group. + awslogs-region Required: Yes Specify the Region that the awslogs log driver is to send your Docker logs to. You can choose to send all of your logs from clusters in different Regions to a single region in CloudWatch Logs. This is so that they're all visible in one location. Otherwise, you can separate them by Region for more granularity. Make sure that the specified log group exists in the Region that you specify with this option. + awslogs-group Required: Yes Make sure to specify a log group that the awslogs log driver sends its log streams to. + awslogs-stream-prefix Required: Yes, when using the Fargate launch type.Optional for the EC2 launch type, required for the Fargate launch type. Use the awslogs-stream-prefix option to associate a log stream with the specified prefix, the container name, and the ID of the Amazon ECS task that the container belongs to. If you specify a prefix with this option, then the log stream takes the format prefix-name/container-name/ecs-task-id. If you don't specify a prefix with this option, then the log stream is named after the container ID that's assigned by the Docker daemon on the container instance. Because it's difficult to trace logs back to the container that sent them with just the Docker container ID (which is only available on the container instance), we recommend that you specify a prefix with this option. For Amazon ECS services, you can use the service name as the prefix. Doing so, you can trace log streams to the service that the container belongs to, the name of the container that sent them, and the ID of the task that the container belongs to. You must specify a stream-prefix for your logs to have your logs appear in the Log pane when using the Amazon ECS console. + awslogs-datetime-format Required: No This option defines a multiline start pattern in Python strftime format. A log message consists of a line that matches the pattern and any following lines that don’t match the pattern. The matched line is the delimiter between log messages. One example of a use case for using this format is for parsing output such as a stack dump, which might otherwise be logged in multiple entries. The correct pattern allows it to be captured in a single entry. For more information, see awslogs-datetime-format. You cannot configure both the awslogs-datetime-format and awslogs-multiline-pattern options. Multiline logging performs regular expression parsing and matching of all log messages. This might have a negative impact on logging performance. + awslogs-multiline-pattern Required: No This option defines a multiline start pattern that uses a regular expression. A log message consists of a line that matches the pattern and any following lines that don’t match the pattern. The matched line is the delimiter between log messages. For more information, see awslogs-multiline-pattern. This option is ignored if awslogs-datetime-format is also configured. You cannot configure both the awslogs-datetime-format and awslogs-multiline-pattern options. Multiline logging performs regular expression parsing and matching of all log messages. This might have a negative impact on logging performance. + mode Required: No Valid values: non-blocking | blocking This option defines the delivery mode of log messages from the container to CloudWatch Logs. The delivery mode you choose affects application availability when the flow of logs from container to CloudWatch is interrupted. If you use the blocking mode and the flow of logs to CloudWatch is interrupted, calls from container code to write to the stdout and stderr streams will block. The logging thread of the application will block as a result. This may cause the application to become unresponsive and lead to container healthcheck failure. If you use the non-blocking mode, the container's logs are instead stored in an in-memory intermediate buffer configured with the max-buffer-size option. This prevents the application from becoming unresponsive when logs cannot be sent to CloudWatch. We recommend using this mode if you want to ensure service availability and are okay with some log loss. For more information, see Preventing log loss with non-blocking mode in the awslogs container log driver. + max-buffer-size Required: No Default value: 1m When non-blocking mode is used, the max-buffer-size log option controls the size of the buffer that's used for intermediate message storage. Make sure to specify an adequate buffer size based on your application. When the buffer fills up, further logs cannot be stored. Logs that cannot be stored are lost.
To route logs using the
- secret
Options List<ServiceSecret> - The secrets to pass to the log configuration. For more information, see Specifying sensitive data in the Amazon Elastic Container Service Developer Guide.
- log
Driver string - The log driver to use for the container.
For tasks on FARGATElong, the supported log drivers are
awslogs
,splunk
, andawsfirelens
. For tasks hosted on Amazon EC2 instances, the supported log drivers areawslogs
,fluentd
,gelf
,json-file
,journald
,syslog
,splunk
, andawsfirelens
. For more information about using theawslogs
log driver, see Send Amazon ECS logs to CloudWatch in the Amazon Elastic Container Service Developer Guide. For more information about using theawsfirelens
log driver, see Send Amazon ECS logs to an service or Partner. If you have a custom driver that isn't listed, you can fork the Amazon ECS container agent project that's available on GitHub and customize it to work with that driver. We encourage you to submit pull requests for changes that you would like to have included. However, we don't currently provide support for running modified copies of this software. - options {[key: string]: string}
- The configuration options to send to the log driver.
The options you can specify depend on the log driver. Some of the options you can specify when you use the
awslogs
log driver to route logs to Amazon CloudWatch include the following:- awslogs-create-group Required: No Specify whether you want the log group to be created automatically. If this option isn't specified, it defaults to false. Your IAM policy must include the logs:CreateLogGroup permission before you attempt to use awslogs-create-group. + awslogs-region Required: Yes Specify the Region that the awslogs log driver is to send your Docker logs to. You can choose to send all of your logs from clusters in different Regions to a single region in CloudWatch Logs. This is so that they're all visible in one location. Otherwise, you can separate them by Region for more granularity. Make sure that the specified log group exists in the Region that you specify with this option. + awslogs-group Required: Yes Make sure to specify a log group that the awslogs log driver sends its log streams to. + awslogs-stream-prefix Required: Yes, when using the Fargate launch type.Optional for the EC2 launch type, required for the Fargate launch type. Use the awslogs-stream-prefix option to associate a log stream with the specified prefix, the container name, and the ID of the Amazon ECS task that the container belongs to. If you specify a prefix with this option, then the log stream takes the format prefix-name/container-name/ecs-task-id. If you don't specify a prefix with this option, then the log stream is named after the container ID that's assigned by the Docker daemon on the container instance. Because it's difficult to trace logs back to the container that sent them with just the Docker container ID (which is only available on the container instance), we recommend that you specify a prefix with this option. For Amazon ECS services, you can use the service name as the prefix. Doing so, you can trace log streams to the service that the container belongs to, the name of the container that sent them, and the ID of the task that the container belongs to. You must specify a stream-prefix for your logs to have your logs appear in the Log pane when using the Amazon ECS console. + awslogs-datetime-format Required: No This option defines a multiline start pattern in Python strftime format. A log message consists of a line that matches the pattern and any following lines that don’t match the pattern. The matched line is the delimiter between log messages. One example of a use case for using this format is for parsing output such as a stack dump, which might otherwise be logged in multiple entries. The correct pattern allows it to be captured in a single entry. For more information, see awslogs-datetime-format. You cannot configure both the awslogs-datetime-format and awslogs-multiline-pattern options. Multiline logging performs regular expression parsing and matching of all log messages. This might have a negative impact on logging performance. + awslogs-multiline-pattern Required: No This option defines a multiline start pattern that uses a regular expression. A log message consists of a line that matches the pattern and any following lines that don’t match the pattern. The matched line is the delimiter between log messages. For more information, see awslogs-multiline-pattern. This option is ignored if awslogs-datetime-format is also configured. You cannot configure both the awslogs-datetime-format and awslogs-multiline-pattern options. Multiline logging performs regular expression parsing and matching of all log messages. This might have a negative impact on logging performance. + mode Required: No Valid values: non-blocking | blocking This option defines the delivery mode of log messages from the container to CloudWatch Logs. The delivery mode you choose affects application availability when the flow of logs from container to CloudWatch is interrupted. If you use the blocking mode and the flow of logs to CloudWatch is interrupted, calls from container code to write to the stdout and stderr streams will block. The logging thread of the application will block as a result. This may cause the application to become unresponsive and lead to container healthcheck failure. If you use the non-blocking mode, the container's logs are instead stored in an in-memory intermediate buffer configured with the max-buffer-size option. This prevents the application from becoming unresponsive when logs cannot be sent to CloudWatch. We recommend using this mode if you want to ensure service availability and are okay with some log loss. For more information, see Preventing log loss with non-blocking mode in the awslogs container log driver. + max-buffer-size Required: No Default value: 1m When non-blocking mode is used, the max-buffer-size log option controls the size of the buffer that's used for intermediate message storage. Make sure to specify an adequate buffer size based on your application. When the buffer fills up, further logs cannot be stored. Logs that cannot be stored are lost.
To route logs using the
splunk
log router, you need to specify asplunk-token
and asplunk-url
. When you use theawsfirelens
log router to route logs to an AWS Service or AWS Partner Network destination for log storage and analytics, you can set thelog-driver-buffer-limit
option to limit the number of events that are buffered in memory, before being sent to the log router container. It can help to resolve potential log loss issue because high throughput might result in memory running out for the buffer inside of Docker. Other options you can specify when usingawsfirelens
to route logs depend on the destination. When you export logs to Amazon Data Firehose, you can specify the AWS Region withregion
and a name for the log stream withdelivery_stream
. When you export logs to Amazon Kinesis Data Streams, you can specify an AWS Region withregion
and a data stream name withstream
. When you export logs to Amazon OpenSearch Service, you can specify options likeName
,Host
(OpenSearch Service endpoint without protocol),Port
,Index
,Type
,Aws_auth
,Aws_region
,Suppress_Type_Name
, andtls
. When you export logs to Amazon S3, you can specify the bucket using thebucket
option. You can also specifyregion
,total_file_size
,upload_timeout
, anduse_put_object
as options. This parameter requires version 1.19 of the Docker Remote API or greater on your container instance. To check the Docker Remote API version on your container instance, log in to your container instance and run the following command:sudo docker version --format '{{.Server.APIVersion}}'
- awslogs-create-group Required: No Specify whether you want the log group to be created automatically. If this option isn't specified, it defaults to false. Your IAM policy must include the logs:CreateLogGroup permission before you attempt to use awslogs-create-group. + awslogs-region Required: Yes Specify the Region that the awslogs log driver is to send your Docker logs to. You can choose to send all of your logs from clusters in different Regions to a single region in CloudWatch Logs. This is so that they're all visible in one location. Otherwise, you can separate them by Region for more granularity. Make sure that the specified log group exists in the Region that you specify with this option. + awslogs-group Required: Yes Make sure to specify a log group that the awslogs log driver sends its log streams to. + awslogs-stream-prefix Required: Yes, when using the Fargate launch type.Optional for the EC2 launch type, required for the Fargate launch type. Use the awslogs-stream-prefix option to associate a log stream with the specified prefix, the container name, and the ID of the Amazon ECS task that the container belongs to. If you specify a prefix with this option, then the log stream takes the format prefix-name/container-name/ecs-task-id. If you don't specify a prefix with this option, then the log stream is named after the container ID that's assigned by the Docker daemon on the container instance. Because it's difficult to trace logs back to the container that sent them with just the Docker container ID (which is only available on the container instance), we recommend that you specify a prefix with this option. For Amazon ECS services, you can use the service name as the prefix. Doing so, you can trace log streams to the service that the container belongs to, the name of the container that sent them, and the ID of the task that the container belongs to. You must specify a stream-prefix for your logs to have your logs appear in the Log pane when using the Amazon ECS console. + awslogs-datetime-format Required: No This option defines a multiline start pattern in Python strftime format. A log message consists of a line that matches the pattern and any following lines that don’t match the pattern. The matched line is the delimiter between log messages. One example of a use case for using this format is for parsing output such as a stack dump, which might otherwise be logged in multiple entries. The correct pattern allows it to be captured in a single entry. For more information, see awslogs-datetime-format. You cannot configure both the awslogs-datetime-format and awslogs-multiline-pattern options. Multiline logging performs regular expression parsing and matching of all log messages. This might have a negative impact on logging performance. + awslogs-multiline-pattern Required: No This option defines a multiline start pattern that uses a regular expression. A log message consists of a line that matches the pattern and any following lines that don’t match the pattern. The matched line is the delimiter between log messages. For more information, see awslogs-multiline-pattern. This option is ignored if awslogs-datetime-format is also configured. You cannot configure both the awslogs-datetime-format and awslogs-multiline-pattern options. Multiline logging performs regular expression parsing and matching of all log messages. This might have a negative impact on logging performance. + mode Required: No Valid values: non-blocking | blocking This option defines the delivery mode of log messages from the container to CloudWatch Logs. The delivery mode you choose affects application availability when the flow of logs from container to CloudWatch is interrupted. If you use the blocking mode and the flow of logs to CloudWatch is interrupted, calls from container code to write to the stdout and stderr streams will block. The logging thread of the application will block as a result. This may cause the application to become unresponsive and lead to container healthcheck failure. If you use the non-blocking mode, the container's logs are instead stored in an in-memory intermediate buffer configured with the max-buffer-size option. This prevents the application from becoming unresponsive when logs cannot be sent to CloudWatch. We recommend using this mode if you want to ensure service availability and are okay with some log loss. For more information, see Preventing log loss with non-blocking mode in the awslogs container log driver. + max-buffer-size Required: No Default value: 1m When non-blocking mode is used, the max-buffer-size log option controls the size of the buffer that's used for intermediate message storage. Make sure to specify an adequate buffer size based on your application. When the buffer fills up, further logs cannot be stored. Logs that cannot be stored are lost.
To route logs using the
- secret
Options ServiceSecret[] - The secrets to pass to the log configuration. For more information, see Specifying sensitive data in the Amazon Elastic Container Service Developer Guide.
- log_
driver str - The log driver to use for the container.
For tasks on FARGATElong, the supported log drivers are
awslogs
,splunk
, andawsfirelens
. For tasks hosted on Amazon EC2 instances, the supported log drivers areawslogs
,fluentd
,gelf
,json-file
,journald
,syslog
,splunk
, andawsfirelens
. For more information about using theawslogs
log driver, see Send Amazon ECS logs to CloudWatch in the Amazon Elastic Container Service Developer Guide. For more information about using theawsfirelens
log driver, see Send Amazon ECS logs to an service or Partner. If you have a custom driver that isn't listed, you can fork the Amazon ECS container agent project that's available on GitHub and customize it to work with that driver. We encourage you to submit pull requests for changes that you would like to have included. However, we don't currently provide support for running modified copies of this software. - options Mapping[str, str]
- The configuration options to send to the log driver.
The options you can specify depend on the log driver. Some of the options you can specify when you use the
awslogs
log driver to route logs to Amazon CloudWatch include the following:- awslogs-create-group Required: No Specify whether you want the log group to be created automatically. If this option isn't specified, it defaults to false. Your IAM policy must include the logs:CreateLogGroup permission before you attempt to use awslogs-create-group. + awslogs-region Required: Yes Specify the Region that the awslogs log driver is to send your Docker logs to. You can choose to send all of your logs from clusters in different Regions to a single region in CloudWatch Logs. This is so that they're all visible in one location. Otherwise, you can separate them by Region for more granularity. Make sure that the specified log group exists in the Region that you specify with this option. + awslogs-group Required: Yes Make sure to specify a log group that the awslogs log driver sends its log streams to. + awslogs-stream-prefix Required: Yes, when using the Fargate launch type.Optional for the EC2 launch type, required for the Fargate launch type. Use the awslogs-stream-prefix option to associate a log stream with the specified prefix, the container name, and the ID of the Amazon ECS task that the container belongs to. If you specify a prefix with this option, then the log stream takes the format prefix-name/container-name/ecs-task-id. If you don't specify a prefix with this option, then the log stream is named after the container ID that's assigned by the Docker daemon on the container instance. Because it's difficult to trace logs back to the container that sent them with just the Docker container ID (which is only available on the container instance), we recommend that you specify a prefix with this option. For Amazon ECS services, you can use the service name as the prefix. Doing so, you can trace log streams to the service that the container belongs to, the name of the container that sent them, and the ID of the task that the container belongs to. You must specify a stream-prefix for your logs to have your logs appear in the Log pane when using the Amazon ECS console. + awslogs-datetime-format Required: No This option defines a multiline start pattern in Python strftime format. A log message consists of a line that matches the pattern and any following lines that don’t match the pattern. The matched line is the delimiter between log messages. One example of a use case for using this format is for parsing output such as a stack dump, which might otherwise be logged in multiple entries. The correct pattern allows it to be captured in a single entry. For more information, see awslogs-datetime-format. You cannot configure both the awslogs-datetime-format and awslogs-multiline-pattern options. Multiline logging performs regular expression parsing and matching of all log messages. This might have a negative impact on logging performance. + awslogs-multiline-pattern Required: No This option defines a multiline start pattern that uses a regular expression. A log message consists of a line that matches the pattern and any following lines that don’t match the pattern. The matched line is the delimiter between log messages. For more information, see awslogs-multiline-pattern. This option is ignored if awslogs-datetime-format is also configured. You cannot configure both the awslogs-datetime-format and awslogs-multiline-pattern options. Multiline logging performs regular expression parsing and matching of all log messages. This might have a negative impact on logging performance. + mode Required: No Valid values: non-blocking | blocking This option defines the delivery mode of log messages from the container to CloudWatch Logs. The delivery mode you choose affects application availability when the flow of logs from container to CloudWatch is interrupted. If you use the blocking mode and the flow of logs to CloudWatch is interrupted, calls from container code to write to the stdout and stderr streams will block. The logging thread of the application will block as a result. This may cause the application to become unresponsive and lead to container healthcheck failure. If you use the non-blocking mode, the container's logs are instead stored in an in-memory intermediate buffer configured with the max-buffer-size option. This prevents the application from becoming unresponsive when logs cannot be sent to CloudWatch. We recommend using this mode if you want to ensure service availability and are okay with some log loss. For more information, see Preventing log loss with non-blocking mode in the awslogs container log driver. + max-buffer-size Required: No Default value: 1m When non-blocking mode is used, the max-buffer-size log option controls the size of the buffer that's used for intermediate message storage. Make sure to specify an adequate buffer size based on your application. When the buffer fills up, further logs cannot be stored. Logs that cannot be stored are lost.
To route logs using the
splunk
log router, you need to specify asplunk-token
and asplunk-url
. When you use theawsfirelens
log router to route logs to an AWS Service or AWS Partner Network destination for log storage and analytics, you can set thelog-driver-buffer-limit
option to limit the number of events that are buffered in memory, before being sent to the log router container. It can help to resolve potential log loss issue because high throughput might result in memory running out for the buffer inside of Docker. Other options you can specify when usingawsfirelens
to route logs depend on the destination. When you export logs to Amazon Data Firehose, you can specify the AWS Region withregion
and a name for the log stream withdelivery_stream
. When you export logs to Amazon Kinesis Data Streams, you can specify an AWS Region withregion
and a data stream name withstream
. When you export logs to Amazon OpenSearch Service, you can specify options likeName
,Host
(OpenSearch Service endpoint without protocol),Port
,Index
,Type
,Aws_auth
,Aws_region
,Suppress_Type_Name
, andtls
. When you export logs to Amazon S3, you can specify the bucket using thebucket
option. You can also specifyregion
,total_file_size
,upload_timeout
, anduse_put_object
as options. This parameter requires version 1.19 of the Docker Remote API or greater on your container instance. To check the Docker Remote API version on your container instance, log in to your container instance and run the following command:sudo docker version --format '{{.Server.APIVersion}}'
- awslogs-create-group Required: No Specify whether you want the log group to be created automatically. If this option isn't specified, it defaults to false. Your IAM policy must include the logs:CreateLogGroup permission before you attempt to use awslogs-create-group. + awslogs-region Required: Yes Specify the Region that the awslogs log driver is to send your Docker logs to. You can choose to send all of your logs from clusters in different Regions to a single region in CloudWatch Logs. This is so that they're all visible in one location. Otherwise, you can separate them by Region for more granularity. Make sure that the specified log group exists in the Region that you specify with this option. + awslogs-group Required: Yes Make sure to specify a log group that the awslogs log driver sends its log streams to. + awslogs-stream-prefix Required: Yes, when using the Fargate launch type.Optional for the EC2 launch type, required for the Fargate launch type. Use the awslogs-stream-prefix option to associate a log stream with the specified prefix, the container name, and the ID of the Amazon ECS task that the container belongs to. If you specify a prefix with this option, then the log stream takes the format prefix-name/container-name/ecs-task-id. If you don't specify a prefix with this option, then the log stream is named after the container ID that's assigned by the Docker daemon on the container instance. Because it's difficult to trace logs back to the container that sent them with just the Docker container ID (which is only available on the container instance), we recommend that you specify a prefix with this option. For Amazon ECS services, you can use the service name as the prefix. Doing so, you can trace log streams to the service that the container belongs to, the name of the container that sent them, and the ID of the task that the container belongs to. You must specify a stream-prefix for your logs to have your logs appear in the Log pane when using the Amazon ECS console. + awslogs-datetime-format Required: No This option defines a multiline start pattern in Python strftime format. A log message consists of a line that matches the pattern and any following lines that don’t match the pattern. The matched line is the delimiter between log messages. One example of a use case for using this format is for parsing output such as a stack dump, which might otherwise be logged in multiple entries. The correct pattern allows it to be captured in a single entry. For more information, see awslogs-datetime-format. You cannot configure both the awslogs-datetime-format and awslogs-multiline-pattern options. Multiline logging performs regular expression parsing and matching of all log messages. This might have a negative impact on logging performance. + awslogs-multiline-pattern Required: No This option defines a multiline start pattern that uses a regular expression. A log message consists of a line that matches the pattern and any following lines that don’t match the pattern. The matched line is the delimiter between log messages. For more information, see awslogs-multiline-pattern. This option is ignored if awslogs-datetime-format is also configured. You cannot configure both the awslogs-datetime-format and awslogs-multiline-pattern options. Multiline logging performs regular expression parsing and matching of all log messages. This might have a negative impact on logging performance. + mode Required: No Valid values: non-blocking | blocking This option defines the delivery mode of log messages from the container to CloudWatch Logs. The delivery mode you choose affects application availability when the flow of logs from container to CloudWatch is interrupted. If you use the blocking mode and the flow of logs to CloudWatch is interrupted, calls from container code to write to the stdout and stderr streams will block. The logging thread of the application will block as a result. This may cause the application to become unresponsive and lead to container healthcheck failure. If you use the non-blocking mode, the container's logs are instead stored in an in-memory intermediate buffer configured with the max-buffer-size option. This prevents the application from becoming unresponsive when logs cannot be sent to CloudWatch. We recommend using this mode if you want to ensure service availability and are okay with some log loss. For more information, see Preventing log loss with non-blocking mode in the awslogs container log driver. + max-buffer-size Required: No Default value: 1m When non-blocking mode is used, the max-buffer-size log option controls the size of the buffer that's used for intermediate message storage. Make sure to specify an adequate buffer size based on your application. When the buffer fills up, further logs cannot be stored. Logs that cannot be stored are lost.
To route logs using the
- secret_
options Sequence[ServiceSecret] - The secrets to pass to the log configuration. For more information, see Specifying sensitive data in the Amazon Elastic Container Service Developer Guide.
- log
Driver String - The log driver to use for the container.
For tasks on FARGATElong, the supported log drivers are
awslogs
,splunk
, andawsfirelens
. For tasks hosted on Amazon EC2 instances, the supported log drivers areawslogs
,fluentd
,gelf
,json-file
,journald
,syslog
,splunk
, andawsfirelens
. For more information about using theawslogs
log driver, see Send Amazon ECS logs to CloudWatch in the Amazon Elastic Container Service Developer Guide. For more information about using theawsfirelens
log driver, see Send Amazon ECS logs to an service or Partner. If you have a custom driver that isn't listed, you can fork the Amazon ECS container agent project that's available on GitHub and customize it to work with that driver. We encourage you to submit pull requests for changes that you would like to have included. However, we don't currently provide support for running modified copies of this software. - options Map<String>
- The configuration options to send to the log driver.
The options you can specify depend on the log driver. Some of the options you can specify when you use the
awslogs
log driver to route logs to Amazon CloudWatch include the following:- awslogs-create-group Required: No Specify whether you want the log group to be created automatically. If this option isn't specified, it defaults to false. Your IAM policy must include the logs:CreateLogGroup permission before you attempt to use awslogs-create-group. + awslogs-region Required: Yes Specify the Region that the awslogs log driver is to send your Docker logs to. You can choose to send all of your logs from clusters in different Regions to a single region in CloudWatch Logs. This is so that they're all visible in one location. Otherwise, you can separate them by Region for more granularity. Make sure that the specified log group exists in the Region that you specify with this option. + awslogs-group Required: Yes Make sure to specify a log group that the awslogs log driver sends its log streams to. + awslogs-stream-prefix Required: Yes, when using the Fargate launch type.Optional for the EC2 launch type, required for the Fargate launch type. Use the awslogs-stream-prefix option to associate a log stream with the specified prefix, the container name, and the ID of the Amazon ECS task that the container belongs to. If you specify a prefix with this option, then the log stream takes the format prefix-name/container-name/ecs-task-id. If you don't specify a prefix with this option, then the log stream is named after the container ID that's assigned by the Docker daemon on the container instance. Because it's difficult to trace logs back to the container that sent them with just the Docker container ID (which is only available on the container instance), we recommend that you specify a prefix with this option. For Amazon ECS services, you can use the service name as the prefix. Doing so, you can trace log streams to the service that the container belongs to, the name of the container that sent them, and the ID of the task that the container belongs to. You must specify a stream-prefix for your logs to have your logs appear in the Log pane when using the Amazon ECS console. + awslogs-datetime-format Required: No This option defines a multiline start pattern in Python strftime format. A log message consists of a line that matches the pattern and any following lines that don’t match the pattern. The matched line is the delimiter between log messages. One example of a use case for using this format is for parsing output such as a stack dump, which might otherwise be logged in multiple entries. The correct pattern allows it to be captured in a single entry. For more information, see awslogs-datetime-format. You cannot configure both the awslogs-datetime-format and awslogs-multiline-pattern options. Multiline logging performs regular expression parsing and matching of all log messages. This might have a negative impact on logging performance. + awslogs-multiline-pattern Required: No This option defines a multiline start pattern that uses a regular expression. A log message consists of a line that matches the pattern and any following lines that don’t match the pattern. The matched line is the delimiter between log messages. For more information, see awslogs-multiline-pattern. This option is ignored if awslogs-datetime-format is also configured. You cannot configure both the awslogs-datetime-format and awslogs-multiline-pattern options. Multiline logging performs regular expression parsing and matching of all log messages. This might have a negative impact on logging performance. + mode Required: No Valid values: non-blocking | blocking This option defines the delivery mode of log messages from the container to CloudWatch Logs. The delivery mode you choose affects application availability when the flow of logs from container to CloudWatch is interrupted. If you use the blocking mode and the flow of logs to CloudWatch is interrupted, calls from container code to write to the stdout and stderr streams will block. The logging thread of the application will block as a result. This may cause the application to become unresponsive and lead to container healthcheck failure. If you use the non-blocking mode, the container's logs are instead stored in an in-memory intermediate buffer configured with the max-buffer-size option. This prevents the application from becoming unresponsive when logs cannot be sent to CloudWatch. We recommend using this mode if you want to ensure service availability and are okay with some log loss. For more information, see Preventing log loss with non-blocking mode in the awslogs container log driver. + max-buffer-size Required: No Default value: 1m When non-blocking mode is used, the max-buffer-size log option controls the size of the buffer that's used for intermediate message storage. Make sure to specify an adequate buffer size based on your application. When the buffer fills up, further logs cannot be stored. Logs that cannot be stored are lost.
To route logs using the
splunk
log router, you need to specify asplunk-token
and asplunk-url
. When you use theawsfirelens
log router to route logs to an AWS Service or AWS Partner Network destination for log storage and analytics, you can set thelog-driver-buffer-limit
option to limit the number of events that are buffered in memory, before being sent to the log router container. It can help to resolve potential log loss issue because high throughput might result in memory running out for the buffer inside of Docker. Other options you can specify when usingawsfirelens
to route logs depend on the destination. When you export logs to Amazon Data Firehose, you can specify the AWS Region withregion
and a name for the log stream withdelivery_stream
. When you export logs to Amazon Kinesis Data Streams, you can specify an AWS Region withregion
and a data stream name withstream
. When you export logs to Amazon OpenSearch Service, you can specify options likeName
,Host
(OpenSearch Service endpoint without protocol),Port
,Index
,Type
,Aws_auth
,Aws_region
,Suppress_Type_Name
, andtls
. When you export logs to Amazon S3, you can specify the bucket using thebucket
option. You can also specifyregion
,total_file_size
,upload_timeout
, anduse_put_object
as options. This parameter requires version 1.19 of the Docker Remote API or greater on your container instance. To check the Docker Remote API version on your container instance, log in to your container instance and run the following command:sudo docker version --format '{{.Server.APIVersion}}'
- awslogs-create-group Required: No Specify whether you want the log group to be created automatically. If this option isn't specified, it defaults to false. Your IAM policy must include the logs:CreateLogGroup permission before you attempt to use awslogs-create-group. + awslogs-region Required: Yes Specify the Region that the awslogs log driver is to send your Docker logs to. You can choose to send all of your logs from clusters in different Regions to a single region in CloudWatch Logs. This is so that they're all visible in one location. Otherwise, you can separate them by Region for more granularity. Make sure that the specified log group exists in the Region that you specify with this option. + awslogs-group Required: Yes Make sure to specify a log group that the awslogs log driver sends its log streams to. + awslogs-stream-prefix Required: Yes, when using the Fargate launch type.Optional for the EC2 launch type, required for the Fargate launch type. Use the awslogs-stream-prefix option to associate a log stream with the specified prefix, the container name, and the ID of the Amazon ECS task that the container belongs to. If you specify a prefix with this option, then the log stream takes the format prefix-name/container-name/ecs-task-id. If you don't specify a prefix with this option, then the log stream is named after the container ID that's assigned by the Docker daemon on the container instance. Because it's difficult to trace logs back to the container that sent them with just the Docker container ID (which is only available on the container instance), we recommend that you specify a prefix with this option. For Amazon ECS services, you can use the service name as the prefix. Doing so, you can trace log streams to the service that the container belongs to, the name of the container that sent them, and the ID of the task that the container belongs to. You must specify a stream-prefix for your logs to have your logs appear in the Log pane when using the Amazon ECS console. + awslogs-datetime-format Required: No This option defines a multiline start pattern in Python strftime format. A log message consists of a line that matches the pattern and any following lines that don’t match the pattern. The matched line is the delimiter between log messages. One example of a use case for using this format is for parsing output such as a stack dump, which might otherwise be logged in multiple entries. The correct pattern allows it to be captured in a single entry. For more information, see awslogs-datetime-format. You cannot configure both the awslogs-datetime-format and awslogs-multiline-pattern options. Multiline logging performs regular expression parsing and matching of all log messages. This might have a negative impact on logging performance. + awslogs-multiline-pattern Required: No This option defines a multiline start pattern that uses a regular expression. A log message consists of a line that matches the pattern and any following lines that don’t match the pattern. The matched line is the delimiter between log messages. For more information, see awslogs-multiline-pattern. This option is ignored if awslogs-datetime-format is also configured. You cannot configure both the awslogs-datetime-format and awslogs-multiline-pattern options. Multiline logging performs regular expression parsing and matching of all log messages. This might have a negative impact on logging performance. + mode Required: No Valid values: non-blocking | blocking This option defines the delivery mode of log messages from the container to CloudWatch Logs. The delivery mode you choose affects application availability when the flow of logs from container to CloudWatch is interrupted. If you use the blocking mode and the flow of logs to CloudWatch is interrupted, calls from container code to write to the stdout and stderr streams will block. The logging thread of the application will block as a result. This may cause the application to become unresponsive and lead to container healthcheck failure. If you use the non-blocking mode, the container's logs are instead stored in an in-memory intermediate buffer configured with the max-buffer-size option. This prevents the application from becoming unresponsive when logs cannot be sent to CloudWatch. We recommend using this mode if you want to ensure service availability and are okay with some log loss. For more information, see Preventing log loss with non-blocking mode in the awslogs container log driver. + max-buffer-size Required: No Default value: 1m When non-blocking mode is used, the max-buffer-size log option controls the size of the buffer that's used for intermediate message storage. Make sure to specify an adequate buffer size based on your application. When the buffer fills up, further logs cannot be stored. Logs that cannot be stored are lost.
To route logs using the
- secret
Options List<Property Map> - The secrets to pass to the log configuration. For more information, see Specifying sensitive data in the Amazon Elastic Container Service Developer Guide.
ServiceManagedEbsVolumeConfiguration, ServiceManagedEbsVolumeConfigurationArgs
- Role
Arn string - The ARN of the IAM role to associate with this volume. This is the Amazon ECS infrastructure IAM role that is used to manage your AWS infrastructure. We recommend using the Amazon ECS-managed
AmazonECSInfrastructureRolePolicyForVolumes
IAM policy with this role. For more information, see Amazon ECS infrastructure IAM role in the Amazon ECS Developer Guide. - Encrypted bool
- Indicates whether the volume should be encrypted. If no value is specified, encryption is turned on by default. This parameter maps 1:1 with the
Encrypted
parameter of the CreateVolume API in the Amazon EC2 API Reference. - Filesystem
Type string - The filesystem type for the volume. For volumes created from a snapshot, you must specify the same filesystem type that the volume was using when the snapshot was created. If there is a filesystem type mismatch, the task will fail to start.
The available Linux filesystem types are
ext3
,ext4
, andxfs
. If no value is specified, thexfs
filesystem type is used by default. The available Windows filesystem types areNTFS
. - Iops int
The number of I/O operations per second (IOPS). For
gp3
,io1
, andio2
volumes, this represents the number of IOPS that are provisioned for the volume. Forgp2
volumes, this represents the baseline performance of the volume and the rate at which the volume accumulates I/O credits for bursting. The following are the supported values for each volume type.gp3
: 3,000 - 16,000 IOPSio1
: 100 - 64,000 IOPSio2
: 100 - 256,000 IOPS
This parameter is required for
io1
andio2
volume types. The default forgp3
volumes is3,000 IOPS
. This parameter is not supported forst1
,sc1
, orstandard
volume types. This parameter maps 1:1 with theIops
parameter of the CreateVolume API in the Amazon EC2 API Reference.- Kms
Key stringId - The Amazon Resource Name (ARN) identifier of the AWS Key Management Service key to use for Amazon EBS encryption. When encryption is turned on and no AWS Key Management Service key is specified, the default AWS managed key for Amazon EBS volumes is used. This parameter maps 1:1 with the
KmsKeyId
parameter of the CreateVolume API in the Amazon EC2 API Reference. AWS authenticates the AWS Key Management Service key asynchronously. Therefore, if you specify an ID, alias, or ARN that is invalid, the action can appear to complete, but eventually fails. - Size
In intGi B - The size of the volume in GiB. You must specify either a volume size or a snapshot ID. If you specify a snapshot ID, the snapshot size is used for the volume size by default. You can optionally specify a volume size greater than or equal to the snapshot size. This parameter maps 1:1 with the
Size
parameter of the CreateVolume API in the Amazon EC2 API Reference. The following are the supported volume size values for each volume type.gp2
andgp3
: 1-16,384io1
andio2
: 4-16,384st1
andsc1
: 125-16,384standard
: 1-1,024
- Snapshot
Id string - The snapshot that Amazon ECS uses to create the volume. You must specify either a snapshot ID or a volume size. This parameter maps 1:1 with the
SnapshotId
parameter of the CreateVolume API in the Amazon EC2 API Reference. - List<Pulumi.
Aws Native. Ecs. Inputs. Service Ebs Tag Specification> - The tags to apply to the volume. Amazon ECS applies service-managed tags by default. This parameter maps 1:1 with the
TagSpecifications.N
parameter of the CreateVolume API in the Amazon EC2 API Reference. - Throughput int
- The throughput to provision for a volume, in MiB/s, with a maximum of 1,000 MiB/s. This parameter maps 1:1 with the
Throughput
parameter of the CreateVolume API in the Amazon EC2 API Reference. This parameter is only supported for thegp3
volume type. - Volume
Type string - The volume type. This parameter maps 1:1 with the
VolumeType
parameter of the CreateVolume API in the Amazon EC2 API Reference. For more information, see Amazon EBS volume types in the Amazon EC2 User Guide. The following are the supported volume types.- General Purpose SSD:
gp2
|gp3
- Provisioned IOPS SSD:
io1
|io2
- Throughput Optimized HDD:
st1
- Cold HDD:
sc1
- Magnetic:
standard
The magnetic volume type is not supported on Fargate.
- General Purpose SSD:
- Role
Arn string - The ARN of the IAM role to associate with this volume. This is the Amazon ECS infrastructure IAM role that is used to manage your AWS infrastructure. We recommend using the Amazon ECS-managed
AmazonECSInfrastructureRolePolicyForVolumes
IAM policy with this role. For more information, see Amazon ECS infrastructure IAM role in the Amazon ECS Developer Guide. - Encrypted bool
- Indicates whether the volume should be encrypted. If no value is specified, encryption is turned on by default. This parameter maps 1:1 with the
Encrypted
parameter of the CreateVolume API in the Amazon EC2 API Reference. - Filesystem
Type string - The filesystem type for the volume. For volumes created from a snapshot, you must specify the same filesystem type that the volume was using when the snapshot was created. If there is a filesystem type mismatch, the task will fail to start.
The available Linux filesystem types are
ext3
,ext4
, andxfs
. If no value is specified, thexfs
filesystem type is used by default. The available Windows filesystem types areNTFS
. - Iops int
The number of I/O operations per second (IOPS). For
gp3
,io1
, andio2
volumes, this represents the number of IOPS that are provisioned for the volume. Forgp2
volumes, this represents the baseline performance of the volume and the rate at which the volume accumulates I/O credits for bursting. The following are the supported values for each volume type.gp3
: 3,000 - 16,000 IOPSio1
: 100 - 64,000 IOPSio2
: 100 - 256,000 IOPS
This parameter is required for
io1
andio2
volume types. The default forgp3
volumes is3,000 IOPS
. This parameter is not supported forst1
,sc1
, orstandard
volume types. This parameter maps 1:1 with theIops
parameter of the CreateVolume API in the Amazon EC2 API Reference.- Kms
Key stringId - The Amazon Resource Name (ARN) identifier of the AWS Key Management Service key to use for Amazon EBS encryption. When encryption is turned on and no AWS Key Management Service key is specified, the default AWS managed key for Amazon EBS volumes is used. This parameter maps 1:1 with the
KmsKeyId
parameter of the CreateVolume API in the Amazon EC2 API Reference. AWS authenticates the AWS Key Management Service key asynchronously. Therefore, if you specify an ID, alias, or ARN that is invalid, the action can appear to complete, but eventually fails. - Size
In intGi B - The size of the volume in GiB. You must specify either a volume size or a snapshot ID. If you specify a snapshot ID, the snapshot size is used for the volume size by default. You can optionally specify a volume size greater than or equal to the snapshot size. This parameter maps 1:1 with the
Size
parameter of the CreateVolume API in the Amazon EC2 API Reference. The following are the supported volume size values for each volume type.gp2
andgp3
: 1-16,384io1
andio2
: 4-16,384st1
andsc1
: 125-16,384standard
: 1-1,024
- Snapshot
Id string - The snapshot that Amazon ECS uses to create the volume. You must specify either a snapshot ID or a volume size. This parameter maps 1:1 with the
SnapshotId
parameter of the CreateVolume API in the Amazon EC2 API Reference. - []Service
Ebs Tag Specification - The tags to apply to the volume. Amazon ECS applies service-managed tags by default. This parameter maps 1:1 with the
TagSpecifications.N
parameter of the CreateVolume API in the Amazon EC2 API Reference. - Throughput int
- The throughput to provision for a volume, in MiB/s, with a maximum of 1,000 MiB/s. This parameter maps 1:1 with the
Throughput
parameter of the CreateVolume API in the Amazon EC2 API Reference. This parameter is only supported for thegp3
volume type. - Volume
Type string - The volume type. This parameter maps 1:1 with the
VolumeType
parameter of the CreateVolume API in the Amazon EC2 API Reference. For more information, see Amazon EBS volume types in the Amazon EC2 User Guide. The following are the supported volume types.- General Purpose SSD:
gp2
|gp3
- Provisioned IOPS SSD:
io1
|io2
- Throughput Optimized HDD:
st1
- Cold HDD:
sc1
- Magnetic:
standard
The magnetic volume type is not supported on Fargate.
- General Purpose SSD:
- role
Arn String - The ARN of the IAM role to associate with this volume. This is the Amazon ECS infrastructure IAM role that is used to manage your AWS infrastructure. We recommend using the Amazon ECS-managed
AmazonECSInfrastructureRolePolicyForVolumes
IAM policy with this role. For more information, see Amazon ECS infrastructure IAM role in the Amazon ECS Developer Guide. - encrypted Boolean
- Indicates whether the volume should be encrypted. If no value is specified, encryption is turned on by default. This parameter maps 1:1 with the
Encrypted
parameter of the CreateVolume API in the Amazon EC2 API Reference. - filesystem
Type String - The filesystem type for the volume. For volumes created from a snapshot, you must specify the same filesystem type that the volume was using when the snapshot was created. If there is a filesystem type mismatch, the task will fail to start.
The available Linux filesystem types are
ext3
,ext4
, andxfs
. If no value is specified, thexfs
filesystem type is used by default. The available Windows filesystem types areNTFS
. - iops Integer
The number of I/O operations per second (IOPS). For
gp3
,io1
, andio2
volumes, this represents the number of IOPS that are provisioned for the volume. Forgp2
volumes, this represents the baseline performance of the volume and the rate at which the volume accumulates I/O credits for bursting. The following are the supported values for each volume type.gp3
: 3,000 - 16,000 IOPSio1
: 100 - 64,000 IOPSio2
: 100 - 256,000 IOPS
This parameter is required for
io1
andio2
volume types. The default forgp3
volumes is3,000 IOPS
. This parameter is not supported forst1
,sc1
, orstandard
volume types. This parameter maps 1:1 with theIops
parameter of the CreateVolume API in the Amazon EC2 API Reference.- kms
Key StringId - The Amazon Resource Name (ARN) identifier of the AWS Key Management Service key to use for Amazon EBS encryption. When encryption is turned on and no AWS Key Management Service key is specified, the default AWS managed key for Amazon EBS volumes is used. This parameter maps 1:1 with the
KmsKeyId
parameter of the CreateVolume API in the Amazon EC2 API Reference. AWS authenticates the AWS Key Management Service key asynchronously. Therefore, if you specify an ID, alias, or ARN that is invalid, the action can appear to complete, but eventually fails. - size
In IntegerGi B - The size of the volume in GiB. You must specify either a volume size or a snapshot ID. If you specify a snapshot ID, the snapshot size is used for the volume size by default. You can optionally specify a volume size greater than or equal to the snapshot size. This parameter maps 1:1 with the
Size
parameter of the CreateVolume API in the Amazon EC2 API Reference. The following are the supported volume size values for each volume type.gp2
andgp3
: 1-16,384io1
andio2
: 4-16,384st1
andsc1
: 125-16,384standard
: 1-1,024
- snapshot
Id String - The snapshot that Amazon ECS uses to create the volume. You must specify either a snapshot ID or a volume size. This parameter maps 1:1 with the
SnapshotId
parameter of the CreateVolume API in the Amazon EC2 API Reference. - List<Service
Ebs Tag Specification> - The tags to apply to the volume. Amazon ECS applies service-managed tags by default. This parameter maps 1:1 with the
TagSpecifications.N
parameter of the CreateVolume API in the Amazon EC2 API Reference. - throughput Integer
- The throughput to provision for a volume, in MiB/s, with a maximum of 1,000 MiB/s. This parameter maps 1:1 with the
Throughput
parameter of the CreateVolume API in the Amazon EC2 API Reference. This parameter is only supported for thegp3
volume type. - volume
Type String - The volume type. This parameter maps 1:1 with the
VolumeType
parameter of the CreateVolume API in the Amazon EC2 API Reference. For more information, see Amazon EBS volume types in the Amazon EC2 User Guide. The following are the supported volume types.- General Purpose SSD:
gp2
|gp3
- Provisioned IOPS SSD:
io1
|io2
- Throughput Optimized HDD:
st1
- Cold HDD:
sc1
- Magnetic:
standard
The magnetic volume type is not supported on Fargate.
- General Purpose SSD:
- role
Arn string - The ARN of the IAM role to associate with this volume. This is the Amazon ECS infrastructure IAM role that is used to manage your AWS infrastructure. We recommend using the Amazon ECS-managed
AmazonECSInfrastructureRolePolicyForVolumes
IAM policy with this role. For more information, see Amazon ECS infrastructure IAM role in the Amazon ECS Developer Guide. - encrypted boolean
- Indicates whether the volume should be encrypted. If no value is specified, encryption is turned on by default. This parameter maps 1:1 with the
Encrypted
parameter of the CreateVolume API in the Amazon EC2 API Reference. - filesystem
Type string - The filesystem type for the volume. For volumes created from a snapshot, you must specify the same filesystem type that the volume was using when the snapshot was created. If there is a filesystem type mismatch, the task will fail to start.
The available Linux filesystem types are
ext3
,ext4
, andxfs
. If no value is specified, thexfs
filesystem type is used by default. The available Windows filesystem types areNTFS
. - iops number
The number of I/O operations per second (IOPS). For
gp3
,io1
, andio2
volumes, this represents the number of IOPS that are provisioned for the volume. Forgp2
volumes, this represents the baseline performance of the volume and the rate at which the volume accumulates I/O credits for bursting. The following are the supported values for each volume type.gp3
: 3,000 - 16,000 IOPSio1
: 100 - 64,000 IOPSio2
: 100 - 256,000 IOPS
This parameter is required for
io1
andio2
volume types. The default forgp3
volumes is3,000 IOPS
. This parameter is not supported forst1
,sc1
, orstandard
volume types. This parameter maps 1:1 with theIops
parameter of the CreateVolume API in the Amazon EC2 API Reference.- kms
Key stringId - The Amazon Resource Name (ARN) identifier of the AWS Key Management Service key to use for Amazon EBS encryption. When encryption is turned on and no AWS Key Management Service key is specified, the default AWS managed key for Amazon EBS volumes is used. This parameter maps 1:1 with the
KmsKeyId
parameter of the CreateVolume API in the Amazon EC2 API Reference. AWS authenticates the AWS Key Management Service key asynchronously. Therefore, if you specify an ID, alias, or ARN that is invalid, the action can appear to complete, but eventually fails. - size
In numberGi B - The size of the volume in GiB. You must specify either a volume size or a snapshot ID. If you specify a snapshot ID, the snapshot size is used for the volume size by default. You can optionally specify a volume size greater than or equal to the snapshot size. This parameter maps 1:1 with the
Size
parameter of the CreateVolume API in the Amazon EC2 API Reference. The following are the supported volume size values for each volume type.gp2
andgp3
: 1-16,384io1
andio2
: 4-16,384st1
andsc1
: 125-16,384standard
: 1-1,024
- snapshot
Id string - The snapshot that Amazon ECS uses to create the volume. You must specify either a snapshot ID or a volume size. This parameter maps 1:1 with the
SnapshotId
parameter of the CreateVolume API in the Amazon EC2 API Reference. - Service
Ebs Tag Specification[] - The tags to apply to the volume. Amazon ECS applies service-managed tags by default. This parameter maps 1:1 with the
TagSpecifications.N
parameter of the CreateVolume API in the Amazon EC2 API Reference. - throughput number
- The throughput to provision for a volume, in MiB/s, with a maximum of 1,000 MiB/s. This parameter maps 1:1 with the
Throughput
parameter of the CreateVolume API in the Amazon EC2 API Reference. This parameter is only supported for thegp3
volume type. - volume
Type string - The volume type. This parameter maps 1:1 with the
VolumeType
parameter of the CreateVolume API in the Amazon EC2 API Reference. For more information, see Amazon EBS volume types in the Amazon EC2 User Guide. The following are the supported volume types.- General Purpose SSD:
gp2
|gp3
- Provisioned IOPS SSD:
io1
|io2
- Throughput Optimized HDD:
st1
- Cold HDD:
sc1
- Magnetic:
standard
The magnetic volume type is not supported on Fargate.
- General Purpose SSD:
- role_
arn str - The ARN of the IAM role to associate with this volume. This is the Amazon ECS infrastructure IAM role that is used to manage your AWS infrastructure. We recommend using the Amazon ECS-managed
AmazonECSInfrastructureRolePolicyForVolumes
IAM policy with this role. For more information, see Amazon ECS infrastructure IAM role in the Amazon ECS Developer Guide. - encrypted bool
- Indicates whether the volume should be encrypted. If no value is specified, encryption is turned on by default. This parameter maps 1:1 with the
Encrypted
parameter of the CreateVolume API in the Amazon EC2 API Reference. - filesystem_
type str - The filesystem type for the volume. For volumes created from a snapshot, you must specify the same filesystem type that the volume was using when the snapshot was created. If there is a filesystem type mismatch, the task will fail to start.
The available Linux filesystem types are
ext3
,ext4
, andxfs
. If no value is specified, thexfs
filesystem type is used by default. The available Windows filesystem types areNTFS
. - iops int
The number of I/O operations per second (IOPS). For
gp3
,io1
, andio2
volumes, this represents the number of IOPS that are provisioned for the volume. Forgp2
volumes, this represents the baseline performance of the volume and the rate at which the volume accumulates I/O credits for bursting. The following are the supported values for each volume type.gp3
: 3,000 - 16,000 IOPSio1
: 100 - 64,000 IOPSio2
: 100 - 256,000 IOPS
This parameter is required for
io1
andio2
volume types. The default forgp3
volumes is3,000 IOPS
. This parameter is not supported forst1
,sc1
, orstandard
volume types. This parameter maps 1:1 with theIops
parameter of the CreateVolume API in the Amazon EC2 API Reference.- kms_
key_ strid - The Amazon Resource Name (ARN) identifier of the AWS Key Management Service key to use for Amazon EBS encryption. When encryption is turned on and no AWS Key Management Service key is specified, the default AWS managed key for Amazon EBS volumes is used. This parameter maps 1:1 with the
KmsKeyId
parameter of the CreateVolume API in the Amazon EC2 API Reference. AWS authenticates the AWS Key Management Service key asynchronously. Therefore, if you specify an ID, alias, or ARN that is invalid, the action can appear to complete, but eventually fails. - size_
in_ intgi_ b - The size of the volume in GiB. You must specify either a volume size or a snapshot ID. If you specify a snapshot ID, the snapshot size is used for the volume size by default. You can optionally specify a volume size greater than or equal to the snapshot size. This parameter maps 1:1 with the
Size
parameter of the CreateVolume API in the Amazon EC2 API Reference. The following are the supported volume size values for each volume type.gp2
andgp3
: 1-16,384io1
andio2
: 4-16,384st1
andsc1
: 125-16,384standard
: 1-1,024
- snapshot_
id str - The snapshot that Amazon ECS uses to create the volume. You must specify either a snapshot ID or a volume size. This parameter maps 1:1 with the
SnapshotId
parameter of the CreateVolume API in the Amazon EC2 API Reference. - tag_
specifications Sequence[ServiceEbs Tag Specification] - The tags to apply to the volume. Amazon ECS applies service-managed tags by default. This parameter maps 1:1 with the
TagSpecifications.N
parameter of the CreateVolume API in the Amazon EC2 API Reference. - throughput int
- The throughput to provision for a volume, in MiB/s, with a maximum of 1,000 MiB/s. This parameter maps 1:1 with the
Throughput
parameter of the CreateVolume API in the Amazon EC2 API Reference. This parameter is only supported for thegp3
volume type. - volume_
type str - The volume type. This parameter maps 1:1 with the
VolumeType
parameter of the CreateVolume API in the Amazon EC2 API Reference. For more information, see Amazon EBS volume types in the Amazon EC2 User Guide. The following are the supported volume types.- General Purpose SSD:
gp2
|gp3
- Provisioned IOPS SSD:
io1
|io2
- Throughput Optimized HDD:
st1
- Cold HDD:
sc1
- Magnetic:
standard
The magnetic volume type is not supported on Fargate.
- General Purpose SSD:
- role
Arn String - The ARN of the IAM role to associate with this volume. This is the Amazon ECS infrastructure IAM role that is used to manage your AWS infrastructure. We recommend using the Amazon ECS-managed
AmazonECSInfrastructureRolePolicyForVolumes
IAM policy with this role. For more information, see Amazon ECS infrastructure IAM role in the Amazon ECS Developer Guide. - encrypted Boolean
- Indicates whether the volume should be encrypted. If no value is specified, encryption is turned on by default. This parameter maps 1:1 with the
Encrypted
parameter of the CreateVolume API in the Amazon EC2 API Reference. - filesystem
Type String - The filesystem type for the volume. For volumes created from a snapshot, you must specify the same filesystem type that the volume was using when the snapshot was created. If there is a filesystem type mismatch, the task will fail to start.
The available Linux filesystem types are
ext3
,ext4
, andxfs
. If no value is specified, thexfs
filesystem type is used by default. The available Windows filesystem types areNTFS
. - iops Number
The number of I/O operations per second (IOPS). For
gp3
,io1
, andio2
volumes, this represents the number of IOPS that are provisioned for the volume. Forgp2
volumes, this represents the baseline performance of the volume and the rate at which the volume accumulates I/O credits for bursting. The following are the supported values for each volume type.gp3
: 3,000 - 16,000 IOPSio1
: 100 - 64,000 IOPSio2
: 100 - 256,000 IOPS
This parameter is required for
io1
andio2
volume types. The default forgp3
volumes is3,000 IOPS
. This parameter is not supported forst1
,sc1
, orstandard
volume types. This parameter maps 1:1 with theIops
parameter of the CreateVolume API in the Amazon EC2 API Reference.- kms
Key StringId - The Amazon Resource Name (ARN) identifier of the AWS Key Management Service key to use for Amazon EBS encryption. When encryption is turned on and no AWS Key Management Service key is specified, the default AWS managed key for Amazon EBS volumes is used. This parameter maps 1:1 with the
KmsKeyId
parameter of the CreateVolume API in the Amazon EC2 API Reference. AWS authenticates the AWS Key Management Service key asynchronously. Therefore, if you specify an ID, alias, or ARN that is invalid, the action can appear to complete, but eventually fails. - size
In NumberGi B - The size of the volume in GiB. You must specify either a volume size or a snapshot ID. If you specify a snapshot ID, the snapshot size is used for the volume size by default. You can optionally specify a volume size greater than or equal to the snapshot size. This parameter maps 1:1 with the
Size
parameter of the CreateVolume API in the Amazon EC2 API Reference. The following are the supported volume size values for each volume type.gp2
andgp3
: 1-16,384io1
andio2
: 4-16,384st1
andsc1
: 125-16,384standard
: 1-1,024
- snapshot
Id String - The snapshot that Amazon ECS uses to create the volume. You must specify either a snapshot ID or a volume size. This parameter maps 1:1 with the
SnapshotId
parameter of the CreateVolume API in the Amazon EC2 API Reference. - List<Property Map>
- The tags to apply to the volume. Amazon ECS applies service-managed tags by default. This parameter maps 1:1 with the
TagSpecifications.N
parameter of the CreateVolume API in the Amazon EC2 API Reference. - throughput Number
- The throughput to provision for a volume, in MiB/s, with a maximum of 1,000 MiB/s. This parameter maps 1:1 with the
Throughput
parameter of the CreateVolume API in the Amazon EC2 API Reference. This parameter is only supported for thegp3
volume type. - volume
Type String - The volume type. This parameter maps 1:1 with the
VolumeType
parameter of the CreateVolume API in the Amazon EC2 API Reference. For more information, see Amazon EBS volume types in the Amazon EC2 User Guide. The following are the supported volume types.- General Purpose SSD:
gp2
|gp3
- Provisioned IOPS SSD:
io1
|io2
- Throughput Optimized HDD:
st1
- Cold HDD:
sc1
- Magnetic:
standard
The magnetic volume type is not supported on Fargate.
- General Purpose SSD:
ServiceNetworkConfiguration, ServiceNetworkConfigurationArgs
- Awsvpc
Configuration Pulumi.Aws Native. Ecs. Inputs. Service Aws Vpc Configuration - The VPC subnets and security groups that are associated with a task. All specified subnets and security groups must be from the same VPC.
- Awsvpc
Configuration ServiceAws Vpc Configuration - The VPC subnets and security groups that are associated with a task. All specified subnets and security groups must be from the same VPC.
- awsvpc
Configuration ServiceAws Vpc Configuration - The VPC subnets and security groups that are associated with a task. All specified subnets and security groups must be from the same VPC.
- awsvpc
Configuration ServiceAws Vpc Configuration - The VPC subnets and security groups that are associated with a task. All specified subnets and security groups must be from the same VPC.
- awsvpc_
configuration ServiceAws Vpc Configuration - The VPC subnets and security groups that are associated with a task. All specified subnets and security groups must be from the same VPC.
- awsvpc
Configuration Property Map - The VPC subnets and security groups that are associated with a task. All specified subnets and security groups must be from the same VPC.
ServicePlacementConstraint, ServicePlacementConstraintArgs
- Type
Pulumi.
Aws Native. Ecs. Service Placement Constraint Type - The type of constraint. Use
distinctInstance
to ensure that each task in a particular group is running on a different container instance. UsememberOf
to restrict the selection to a group of valid candidates. - Expression string
- A cluster query language expression to apply to the constraint. The expression can have a maximum length of 2000 characters. You can't specify an expression if the constraint type is
distinctInstance
. For more information, see Cluster query language in the Amazon Elastic Container Service Developer Guide.
- Type
Service
Placement Constraint Type - The type of constraint. Use
distinctInstance
to ensure that each task in a particular group is running on a different container instance. UsememberOf
to restrict the selection to a group of valid candidates. - Expression string
- A cluster query language expression to apply to the constraint. The expression can have a maximum length of 2000 characters. You can't specify an expression if the constraint type is
distinctInstance
. For more information, see Cluster query language in the Amazon Elastic Container Service Developer Guide.
- type
Service
Placement Constraint Type - The type of constraint. Use
distinctInstance
to ensure that each task in a particular group is running on a different container instance. UsememberOf
to restrict the selection to a group of valid candidates. - expression String
- A cluster query language expression to apply to the constraint. The expression can have a maximum length of 2000 characters. You can't specify an expression if the constraint type is
distinctInstance
. For more information, see Cluster query language in the Amazon Elastic Container Service Developer Guide.
- type
Service
Placement Constraint Type - The type of constraint. Use
distinctInstance
to ensure that each task in a particular group is running on a different container instance. UsememberOf
to restrict the selection to a group of valid candidates. - expression string
- A cluster query language expression to apply to the constraint. The expression can have a maximum length of 2000 characters. You can't specify an expression if the constraint type is
distinctInstance
. For more information, see Cluster query language in the Amazon Elastic Container Service Developer Guide.
- type
Service
Placement Constraint Type - The type of constraint. Use
distinctInstance
to ensure that each task in a particular group is running on a different container instance. UsememberOf
to restrict the selection to a group of valid candidates. - expression str
- A cluster query language expression to apply to the constraint. The expression can have a maximum length of 2000 characters. You can't specify an expression if the constraint type is
distinctInstance
. For more information, see Cluster query language in the Amazon Elastic Container Service Developer Guide.
- type
"distinct
Instance" | "member Of" - The type of constraint. Use
distinctInstance
to ensure that each task in a particular group is running on a different container instance. UsememberOf
to restrict the selection to a group of valid candidates. - expression String
- A cluster query language expression to apply to the constraint. The expression can have a maximum length of 2000 characters. You can't specify an expression if the constraint type is
distinctInstance
. For more information, see Cluster query language in the Amazon Elastic Container Service Developer Guide.
ServicePlacementConstraintType, ServicePlacementConstraintTypeArgs
- Distinct
Instance - distinctInstance
- Member
Of - memberOf
- Service
Placement Constraint Type Distinct Instance - distinctInstance
- Service
Placement Constraint Type Member Of - memberOf
- Distinct
Instance - distinctInstance
- Member
Of - memberOf
- Distinct
Instance - distinctInstance
- Member
Of - memberOf
- DISTINCT_INSTANCE
- distinctInstance
- MEMBER_OF
- memberOf
- "distinct
Instance" - distinctInstance
- "member
Of" - memberOf
ServicePlacementStrategy, ServicePlacementStrategyArgs
- Type
Pulumi.
Aws Native. Ecs. Service Placement Strategy Type - The type of placement strategy. The
random
placement strategy randomly places tasks on available candidates. Thespread
placement strategy spreads placement across available candidates evenly based on thefield
parameter. Thebinpack
strategy places tasks on available candidates that have the least available amount of the resource that's specified with thefield
parameter. For example, if you binpack on memory, a task is placed on the instance with the least amount of remaining memory but still enough to run the task. - Field string
- The field to apply the placement strategy against. For the
spread
placement strategy, valid values areinstanceId
(orhost
, which has the same effect), or any platform or custom attribute that's applied to a container instance, such asattribute:ecs.availability-zone
. For thebinpack
placement strategy, valid values arecpu
andmemory
. For therandom
placement strategy, this field is not used.
- Type
Service
Placement Strategy Type - The type of placement strategy. The
random
placement strategy randomly places tasks on available candidates. Thespread
placement strategy spreads placement across available candidates evenly based on thefield
parameter. Thebinpack
strategy places tasks on available candidates that have the least available amount of the resource that's specified with thefield
parameter. For example, if you binpack on memory, a task is placed on the instance with the least amount of remaining memory but still enough to run the task. - Field string
- The field to apply the placement strategy against. For the
spread
placement strategy, valid values areinstanceId
(orhost
, which has the same effect), or any platform or custom attribute that's applied to a container instance, such asattribute:ecs.availability-zone
. For thebinpack
placement strategy, valid values arecpu
andmemory
. For therandom
placement strategy, this field is not used.
- type
Service
Placement Strategy Type - The type of placement strategy. The
random
placement strategy randomly places tasks on available candidates. Thespread
placement strategy spreads placement across available candidates evenly based on thefield
parameter. Thebinpack
strategy places tasks on available candidates that have the least available amount of the resource that's specified with thefield
parameter. For example, if you binpack on memory, a task is placed on the instance with the least amount of remaining memory but still enough to run the task. - field String
- The field to apply the placement strategy against. For the
spread
placement strategy, valid values areinstanceId
(orhost
, which has the same effect), or any platform or custom attribute that's applied to a container instance, such asattribute:ecs.availability-zone
. For thebinpack
placement strategy, valid values arecpu
andmemory
. For therandom
placement strategy, this field is not used.
- type
Service
Placement Strategy Type - The type of placement strategy. The
random
placement strategy randomly places tasks on available candidates. Thespread
placement strategy spreads placement across available candidates evenly based on thefield
parameter. Thebinpack
strategy places tasks on available candidates that have the least available amount of the resource that's specified with thefield
parameter. For example, if you binpack on memory, a task is placed on the instance with the least amount of remaining memory but still enough to run the task. - field string
- The field to apply the placement strategy against. For the
spread
placement strategy, valid values areinstanceId
(orhost
, which has the same effect), or any platform or custom attribute that's applied to a container instance, such asattribute:ecs.availability-zone
. For thebinpack
placement strategy, valid values arecpu
andmemory
. For therandom
placement strategy, this field is not used.
- type
Service
Placement Strategy Type - The type of placement strategy. The
random
placement strategy randomly places tasks on available candidates. Thespread
placement strategy spreads placement across available candidates evenly based on thefield
parameter. Thebinpack
strategy places tasks on available candidates that have the least available amount of the resource that's specified with thefield
parameter. For example, if you binpack on memory, a task is placed on the instance with the least amount of remaining memory but still enough to run the task. - field str
- The field to apply the placement strategy against. For the
spread
placement strategy, valid values areinstanceId
(orhost
, which has the same effect), or any platform or custom attribute that's applied to a container instance, such asattribute:ecs.availability-zone
. For thebinpack
placement strategy, valid values arecpu
andmemory
. For therandom
placement strategy, this field is not used.
- type "binpack" | "random" | "spread"
- The type of placement strategy. The
random
placement strategy randomly places tasks on available candidates. Thespread
placement strategy spreads placement across available candidates evenly based on thefield
parameter. Thebinpack
strategy places tasks on available candidates that have the least available amount of the resource that's specified with thefield
parameter. For example, if you binpack on memory, a task is placed on the instance with the least amount of remaining memory but still enough to run the task. - field String
- The field to apply the placement strategy against. For the
spread
placement strategy, valid values areinstanceId
(orhost
, which has the same effect), or any platform or custom attribute that's applied to a container instance, such asattribute:ecs.availability-zone
. For thebinpack
placement strategy, valid values arecpu
andmemory
. For therandom
placement strategy, this field is not used.
ServicePlacementStrategyType, ServicePlacementStrategyTypeArgs
- Binpack
- binpack
- Random
- random
- Spread
- spread
- Service
Placement Strategy Type Binpack - binpack
- Service
Placement Strategy Type Random - random
- Service
Placement Strategy Type Spread - spread
- Binpack
- binpack
- Random
- random
- Spread
- spread
- Binpack
- binpack
- Random
- random
- Spread
- spread
- BINPACK
- binpack
- RANDOM
- random
- SPREAD
- spread
- "binpack"
- binpack
- "random"
- random
- "spread"
- spread
ServicePropagateTags, ServicePropagateTagsArgs
- Service
- SERVICE
- Task
Definition - TASK_DEFINITION
- Service
Propagate Tags Service - SERVICE
- Service
Propagate Tags Task Definition - TASK_DEFINITION
- Service
- SERVICE
- Task
Definition - TASK_DEFINITION
- Service
- SERVICE
- Task
Definition - TASK_DEFINITION
- SERVICE
- SERVICE
- TASK_DEFINITION
- TASK_DEFINITION
- "SERVICE"
- SERVICE
- "TASK_DEFINITION"
- TASK_DEFINITION
ServiceRegistry, ServiceRegistryArgs
- Container
Name string - The container name value to be used for your service discovery service. It's already specified in the task definition. If the task definition that your service task specifies uses the
bridge
orhost
network mode, you must specify acontainerName
andcontainerPort
combination from the task definition. If the task definition that your service task specifies uses theawsvpc
network mode and a type SRV DNS record is used, you must specify either acontainerName
andcontainerPort
combination or aport
value. However, you can't specify both. - Container
Port int - The port value to be used for your service discovery service. It's already specified in the task definition. If the task definition your service task specifies uses the
bridge
orhost
network mode, you must specify acontainerName
andcontainerPort
combination from the task definition. If the task definition your service task specifies uses theawsvpc
network mode and a type SRV DNS record is used, you must specify either acontainerName
andcontainerPort
combination or aport
value. However, you can't specify both. - Port int
- The port value used if your service discovery service specified an SRV record. This field might be used if both the
awsvpc
network mode and SRV records are used. - Registry
Arn string - The Amazon Resource Name (ARN) of the service registry. The currently supported service registry is CMAP. For more information, see CreateService.
- Container
Name string - The container name value to be used for your service discovery service. It's already specified in the task definition. If the task definition that your service task specifies uses the
bridge
orhost
network mode, you must specify acontainerName
andcontainerPort
combination from the task definition. If the task definition that your service task specifies uses theawsvpc
network mode and a type SRV DNS record is used, you must specify either acontainerName
andcontainerPort
combination or aport
value. However, you can't specify both. - Container
Port int - The port value to be used for your service discovery service. It's already specified in the task definition. If the task definition your service task specifies uses the
bridge
orhost
network mode, you must specify acontainerName
andcontainerPort
combination from the task definition. If the task definition your service task specifies uses theawsvpc
network mode and a type SRV DNS record is used, you must specify either acontainerName
andcontainerPort
combination or aport
value. However, you can't specify both. - Port int
- The port value used if your service discovery service specified an SRV record. This field might be used if both the
awsvpc
network mode and SRV records are used. - Registry
Arn string - The Amazon Resource Name (ARN) of the service registry. The currently supported service registry is CMAP. For more information, see CreateService.
- container
Name String - The container name value to be used for your service discovery service. It's already specified in the task definition. If the task definition that your service task specifies uses the
bridge
orhost
network mode, you must specify acontainerName
andcontainerPort
combination from the task definition. If the task definition that your service task specifies uses theawsvpc
network mode and a type SRV DNS record is used, you must specify either acontainerName
andcontainerPort
combination or aport
value. However, you can't specify both. - container
Port Integer - The port value to be used for your service discovery service. It's already specified in the task definition. If the task definition your service task specifies uses the
bridge
orhost
network mode, you must specify acontainerName
andcontainerPort
combination from the task definition. If the task definition your service task specifies uses theawsvpc
network mode and a type SRV DNS record is used, you must specify either acontainerName
andcontainerPort
combination or aport
value. However, you can't specify both. - port Integer
- The port value used if your service discovery service specified an SRV record. This field might be used if both the
awsvpc
network mode and SRV records are used. - registry
Arn String - The Amazon Resource Name (ARN) of the service registry. The currently supported service registry is CMAP. For more information, see CreateService.
- container
Name string - The container name value to be used for your service discovery service. It's already specified in the task definition. If the task definition that your service task specifies uses the
bridge
orhost
network mode, you must specify acontainerName
andcontainerPort
combination from the task definition. If the task definition that your service task specifies uses theawsvpc
network mode and a type SRV DNS record is used, you must specify either acontainerName
andcontainerPort
combination or aport
value. However, you can't specify both. - container
Port number - The port value to be used for your service discovery service. It's already specified in the task definition. If the task definition your service task specifies uses the
bridge
orhost
network mode, you must specify acontainerName
andcontainerPort
combination from the task definition. If the task definition your service task specifies uses theawsvpc
network mode and a type SRV DNS record is used, you must specify either acontainerName
andcontainerPort
combination or aport
value. However, you can't specify both. - port number
- The port value used if your service discovery service specified an SRV record. This field might be used if both the
awsvpc
network mode and SRV records are used. - registry
Arn string - The Amazon Resource Name (ARN) of the service registry. The currently supported service registry is CMAP. For more information, see CreateService.
- container_
name str - The container name value to be used for your service discovery service. It's already specified in the task definition. If the task definition that your service task specifies uses the
bridge
orhost
network mode, you must specify acontainerName
andcontainerPort
combination from the task definition. If the task definition that your service task specifies uses theawsvpc
network mode and a type SRV DNS record is used, you must specify either acontainerName
andcontainerPort
combination or aport
value. However, you can't specify both. - container_
port int - The port value to be used for your service discovery service. It's already specified in the task definition. If the task definition your service task specifies uses the
bridge
orhost
network mode, you must specify acontainerName
andcontainerPort
combination from the task definition. If the task definition your service task specifies uses theawsvpc
network mode and a type SRV DNS record is used, you must specify either acontainerName
andcontainerPort
combination or aport
value. However, you can't specify both. - port int
- The port value used if your service discovery service specified an SRV record. This field might be used if both the
awsvpc
network mode and SRV records are used. - registry_
arn str - The Amazon Resource Name (ARN) of the service registry. The currently supported service registry is CMAP. For more information, see CreateService.
- container
Name String - The container name value to be used for your service discovery service. It's already specified in the task definition. If the task definition that your service task specifies uses the
bridge
orhost
network mode, you must specify acontainerName
andcontainerPort
combination from the task definition. If the task definition that your service task specifies uses theawsvpc
network mode and a type SRV DNS record is used, you must specify either acontainerName
andcontainerPort
combination or aport
value. However, you can't specify both. - container
Port Number - The port value to be used for your service discovery service. It's already specified in the task definition. If the task definition your service task specifies uses the
bridge
orhost
network mode, you must specify acontainerName
andcontainerPort
combination from the task definition. If the task definition your service task specifies uses theawsvpc
network mode and a type SRV DNS record is used, you must specify either acontainerName
andcontainerPort
combination or aport
value. However, you can't specify both. - port Number
- The port value used if your service discovery service specified an SRV record. This field might be used if both the
awsvpc
network mode and SRV records are used. - registry
Arn String - The Amazon Resource Name (ARN) of the service registry. The currently supported service registry is CMAP. For more information, see CreateService.
ServiceSchedulingStrategy, ServiceSchedulingStrategyArgs
- Daemon
- DAEMON
- Replica
- REPLICA
- Service
Scheduling Strategy Daemon - DAEMON
- Service
Scheduling Strategy Replica - REPLICA
- Daemon
- DAEMON
- Replica
- REPLICA
- Daemon
- DAEMON
- Replica
- REPLICA
- DAEMON
- DAEMON
- REPLICA
- REPLICA
- "DAEMON"
- DAEMON
- "REPLICA"
- REPLICA
ServiceSecret, ServiceSecretArgs
- Name string
- The name of the secret.
- Value
From string - The secret to expose to the container. The supported values are either the full ARN of the ASMlong secret or the full ARN of the parameter in the SSM Parameter Store. For information about the require IAMlong permissions, see Required IAM permissions for Amazon ECS secrets (for Secrets Manager) or Required IAM permissions for Amazon ECS secrets (for Systems Manager Parameter store) in the Amazon Elastic Container Service Developer Guide. If the SSM Parameter Store parameter exists in the same Region as the task you're launching, then you can use either the full ARN or name of the parameter. If the parameter exists in a different Region, then the full ARN must be specified.
- Name string
- The name of the secret.
- Value
From string - The secret to expose to the container. The supported values are either the full ARN of the ASMlong secret or the full ARN of the parameter in the SSM Parameter Store. For information about the require IAMlong permissions, see Required IAM permissions for Amazon ECS secrets (for Secrets Manager) or Required IAM permissions for Amazon ECS secrets (for Systems Manager Parameter store) in the Amazon Elastic Container Service Developer Guide. If the SSM Parameter Store parameter exists in the same Region as the task you're launching, then you can use either the full ARN or name of the parameter. If the parameter exists in a different Region, then the full ARN must be specified.
- name String
- The name of the secret.
- value
From String - The secret to expose to the container. The supported values are either the full ARN of the ASMlong secret or the full ARN of the parameter in the SSM Parameter Store. For information about the require IAMlong permissions, see Required IAM permissions for Amazon ECS secrets (for Secrets Manager) or Required IAM permissions for Amazon ECS secrets (for Systems Manager Parameter store) in the Amazon Elastic Container Service Developer Guide. If the SSM Parameter Store parameter exists in the same Region as the task you're launching, then you can use either the full ARN or name of the parameter. If the parameter exists in a different Region, then the full ARN must be specified.
- name string
- The name of the secret.
- value
From string - The secret to expose to the container. The supported values are either the full ARN of the ASMlong secret or the full ARN of the parameter in the SSM Parameter Store. For information about the require IAMlong permissions, see Required IAM permissions for Amazon ECS secrets (for Secrets Manager) or Required IAM permissions for Amazon ECS secrets (for Systems Manager Parameter store) in the Amazon Elastic Container Service Developer Guide. If the SSM Parameter Store parameter exists in the same Region as the task you're launching, then you can use either the full ARN or name of the parameter. If the parameter exists in a different Region, then the full ARN must be specified.
- name str
- The name of the secret.
- value_
from str - The secret to expose to the container. The supported values are either the full ARN of the ASMlong secret or the full ARN of the parameter in the SSM Parameter Store. For information about the require IAMlong permissions, see Required IAM permissions for Amazon ECS secrets (for Secrets Manager) or Required IAM permissions for Amazon ECS secrets (for Systems Manager Parameter store) in the Amazon Elastic Container Service Developer Guide. If the SSM Parameter Store parameter exists in the same Region as the task you're launching, then you can use either the full ARN or name of the parameter. If the parameter exists in a different Region, then the full ARN must be specified.
- name String
- The name of the secret.
- value
From String - The secret to expose to the container. The supported values are either the full ARN of the ASMlong secret or the full ARN of the parameter in the SSM Parameter Store. For information about the require IAMlong permissions, see Required IAM permissions for Amazon ECS secrets (for Secrets Manager) or Required IAM permissions for Amazon ECS secrets (for Systems Manager Parameter store) in the Amazon Elastic Container Service Developer Guide. If the SSM Parameter Store parameter exists in the same Region as the task you're launching, then you can use either the full ARN or name of the parameter. If the parameter exists in a different Region, then the full ARN must be specified.
ServiceTag, ServiceTagArgs
ServiceTimeoutConfiguration, ServiceTimeoutConfigurationArgs
- Idle
Timeout intSeconds - The amount of time in seconds a connection will stay active while idle. A value of
0
can be set to disableidleTimeout
. TheidleTimeout
default forHTTP
/HTTP2
/GRPC
is 5 minutes. TheidleTimeout
default forTCP
is 1 hour. - Per
Request intTimeout Seconds - The amount of time waiting for the upstream to respond with a complete response per request. A value of
0
can be set to disableperRequestTimeout
.perRequestTimeout
can only be set if Service ConnectappProtocol
isn'tTCP
. OnlyidleTimeout
is allowed forTCP
appProtocol
.
- Idle
Timeout intSeconds - The amount of time in seconds a connection will stay active while idle. A value of
0
can be set to disableidleTimeout
. TheidleTimeout
default forHTTP
/HTTP2
/GRPC
is 5 minutes. TheidleTimeout
default forTCP
is 1 hour. - Per
Request intTimeout Seconds - The amount of time waiting for the upstream to respond with a complete response per request. A value of
0
can be set to disableperRequestTimeout
.perRequestTimeout
can only be set if Service ConnectappProtocol
isn'tTCP
. OnlyidleTimeout
is allowed forTCP
appProtocol
.
- idle
Timeout IntegerSeconds - The amount of time in seconds a connection will stay active while idle. A value of
0
can be set to disableidleTimeout
. TheidleTimeout
default forHTTP
/HTTP2
/GRPC
is 5 minutes. TheidleTimeout
default forTCP
is 1 hour. - per
Request IntegerTimeout Seconds - The amount of time waiting for the upstream to respond with a complete response per request. A value of
0
can be set to disableperRequestTimeout
.perRequestTimeout
can only be set if Service ConnectappProtocol
isn'tTCP
. OnlyidleTimeout
is allowed forTCP
appProtocol
.
- idle
Timeout numberSeconds - The amount of time in seconds a connection will stay active while idle. A value of
0
can be set to disableidleTimeout
. TheidleTimeout
default forHTTP
/HTTP2
/GRPC
is 5 minutes. TheidleTimeout
default forTCP
is 1 hour. - per
Request numberTimeout Seconds - The amount of time waiting for the upstream to respond with a complete response per request. A value of
0
can be set to disableperRequestTimeout
.perRequestTimeout
can only be set if Service ConnectappProtocol
isn'tTCP
. OnlyidleTimeout
is allowed forTCP
appProtocol
.
- idle_
timeout_ intseconds - The amount of time in seconds a connection will stay active while idle. A value of
0
can be set to disableidleTimeout
. TheidleTimeout
default forHTTP
/HTTP2
/GRPC
is 5 minutes. TheidleTimeout
default forTCP
is 1 hour. - per_
request_ inttimeout_ seconds - The amount of time waiting for the upstream to respond with a complete response per request. A value of
0
can be set to disableperRequestTimeout
.perRequestTimeout
can only be set if Service ConnectappProtocol
isn'tTCP
. OnlyidleTimeout
is allowed forTCP
appProtocol
.
- idle
Timeout NumberSeconds - The amount of time in seconds a connection will stay active while idle. A value of
0
can be set to disableidleTimeout
. TheidleTimeout
default forHTTP
/HTTP2
/GRPC
is 5 minutes. TheidleTimeout
default forTCP
is 1 hour. - per
Request NumberTimeout Seconds - The amount of time waiting for the upstream to respond with a complete response per request. A value of
0
can be set to disableperRequestTimeout
.perRequestTimeout
can only be set if Service ConnectappProtocol
isn'tTCP
. OnlyidleTimeout
is allowed forTCP
appProtocol
.
ServiceVolumeConfiguration, ServiceVolumeConfigurationArgs
- Name string
- The name of the volume. This value must match the volume name from the
Volume
object in the task definition. - Managed
Ebs Pulumi.Volume Aws Native. Ecs. Inputs. Service Managed Ebs Volume Configuration - The configuration for the Amazon EBS volume that Amazon ECS creates and manages on your behalf. These settings are used to create each Amazon EBS volume, with one volume created for each task in the service. The Amazon EBS volumes are visible in your account in the Amazon EC2 console once they are created.
- Name string
- The name of the volume. This value must match the volume name from the
Volume
object in the task definition. - Managed
Ebs ServiceVolume Managed Ebs Volume Configuration - The configuration for the Amazon EBS volume that Amazon ECS creates and manages on your behalf. These settings are used to create each Amazon EBS volume, with one volume created for each task in the service. The Amazon EBS volumes are visible in your account in the Amazon EC2 console once they are created.
- name String
- The name of the volume. This value must match the volume name from the
Volume
object in the task definition. - managed
Ebs ServiceVolume Managed Ebs Volume Configuration - The configuration for the Amazon EBS volume that Amazon ECS creates and manages on your behalf. These settings are used to create each Amazon EBS volume, with one volume created for each task in the service. The Amazon EBS volumes are visible in your account in the Amazon EC2 console once they are created.
- name string
- The name of the volume. This value must match the volume name from the
Volume
object in the task definition. - managed
Ebs ServiceVolume Managed Ebs Volume Configuration - The configuration for the Amazon EBS volume that Amazon ECS creates and manages on your behalf. These settings are used to create each Amazon EBS volume, with one volume created for each task in the service. The Amazon EBS volumes are visible in your account in the Amazon EC2 console once they are created.
- name str
- The name of the volume. This value must match the volume name from the
Volume
object in the task definition. - managed_
ebs_ Servicevolume Managed Ebs Volume Configuration - The configuration for the Amazon EBS volume that Amazon ECS creates and manages on your behalf. These settings are used to create each Amazon EBS volume, with one volume created for each task in the service. The Amazon EBS volumes are visible in your account in the Amazon EC2 console once they are created.
- name String
- The name of the volume. This value must match the volume name from the
Volume
object in the task definition. - managed
Ebs Property MapVolume - The configuration for the Amazon EBS volume that Amazon ECS creates and manages on your behalf. These settings are used to create each Amazon EBS volume, with one volume created for each task in the service. The Amazon EBS volumes are visible in your account in the Amazon EC2 console once they are created.
ServiceVpcLatticeConfiguration, ServiceVpcLatticeConfigurationArgs
- Port
Name string - Role
Arn string - Target
Group stringArn
- Port
Name string - Role
Arn string - Target
Group stringArn
- port
Name String - role
Arn String - target
Group StringArn
- port
Name string - role
Arn string - target
Group stringArn
- port_
name str - role_
arn str - target_
group_ strarn
- port
Name String - role
Arn String - target
Group StringArn
Tag, TagArgs
Package Details
- Repository
- AWS Native pulumi/pulumi-aws-native
- License
- Apache-2.0
We recommend new projects start with resources from the AWS provider.