diff --git a/Cargo.toml b/Cargo.toml index 9e841d6fafd6..179b3547ba2d 100644 --- a/Cargo.toml +++ b/Cargo.toml @@ -2,10 +2,10 @@ resolver = "2" exclude = [ "examples/webassembly", - "examples/lambda", - "examples/test-utils", - "examples/cross_service", "examples/examples", + "examples/cross_service", + "examples/test-utils", + "examples/lambda", "tests/webassembly", "tests/no-default-features" ] @@ -151,6 +151,7 @@ members = [ "sdk/devopsguru", "sdk/directconnect", "sdk/directory", + "sdk/directoryservicedata", "sdk/dlm", "sdk/docdb", "sdk/docdbelastic", diff --git a/aws-models/cost-explorer.json b/aws-models/cost-explorer.json index 2800a9bc06c0..a70595c61bb0 100644 --- a/aws-models/cost-explorer.json +++ b/aws-models/cost-explorer.json @@ -1020,7 +1020,7 @@ "DimensionValue": { "target": "com.amazonaws.costexplorer#GenericString", "traits": { - "smithy.api#documentation": "
The dimension for the anomaly (for example, an Amazon Web Service in a service\n monitor).
" + "smithy.api#documentation": "The dimension for the anomaly (for example, an Amazon Web Servicesservice in a service\n monitor).
" } }, "RootCauses": { @@ -3021,6 +3021,26 @@ "smithy.api#documentation": "The field that contains a list of disk (local storage) metrics that are associated\n with the current instance.
" } }, + "com.amazonaws.costexplorer#DynamoDBCapacityDetails": { + "type": "structure", + "members": { + "CapacityUnits": { + "target": "com.amazonaws.costexplorer#GenericString", + "traits": { + "smithy.api#documentation": "The capacity unit of the recommended reservation.
" + } + }, + "Region": { + "target": "com.amazonaws.costexplorer#GenericString", + "traits": { + "smithy.api#documentation": "The Amazon Web Services Region of the recommended reservation.
" + } + } + }, + "traits": { + "smithy.api#documentation": "The DynamoDB reservations that Amazon Web Services recommends that you purchase.
" + } + }, "com.amazonaws.costexplorer#EBSResourceUtilization": { "type": "structure", "members": { @@ -4398,7 +4418,7 @@ "Context": { "target": "com.amazonaws.costexplorer#Context", "traits": { - "smithy.api#documentation": "The context for the call to GetDimensionValues
. This can be\n RESERVATIONS
or COST_AND_USAGE
. The default value is\n COST_AND_USAGE
. If the context is set to RESERVATIONS
, the\n resulting dimension values can be used in the GetReservationUtilization
\n operation. If the context is set to COST_AND_USAGE
, the resulting dimension\n values can be used in the GetCostAndUsage
operation.
If you set the context to COST_AND_USAGE
, you can use the following\n dimensions for searching:
AZ - The Availability Zone. An example is us-east-1a
.
BILLING_ENTITY - The Amazon Web Services seller that your account is with. Possible\n values are the following:
\n- Amazon Web Services(Amazon Web Services): The entity that sells Amazon Web Services.
\n- AISPL (Amazon Internet Services Pvt. Ltd.): The local Indian entity that's an acting\n reseller for Amazon Web Services in India.
\n- Amazon Web Services Marketplace: The entity that supports the sale of solutions that are built on\n Amazon Web Services by third-party software providers.
\nCACHE_ENGINE - The Amazon ElastiCache operating system. Examples are Windows or\n Linux.
\nDEPLOYMENT_OPTION - The scope of Amazon Relational Database Service deployments.\n Valid values are SingleAZ
and MultiAZ
.
DATABASE_ENGINE - The Amazon Relational Database Service database. Examples are\n Aurora or MySQL.
\nINSTANCE_TYPE - The type of Amazon EC2 instance. An example is\n m4.xlarge
.
INSTANCE_TYPE_FAMILY - A family of instance types optimized to fit different use\n cases. Examples are Compute Optimized
(for example, C4
,\n C5
, C6g
, and C7g
), Memory\n Optimization
(for example, R4
, R5n
, R5b
,\n and R6g
).
INVOICING_ENTITY - The name of the entity that issues the Amazon Web Services\n invoice.
\nLEGAL_ENTITY_NAME - The name of the organization that sells you Amazon Web Services\n services, such as Amazon Web Services.
\nLINKED_ACCOUNT - The description in the attribute map that includes the full name\n of the member account. The value field contains the Amazon Web Services ID of the member\n account.
\nOPERATING_SYSTEM - The operating system. Examples are Windows or Linux.
\nOPERATION - The action performed. Examples include RunInstance
and\n CreateBucket
.
PLATFORM - The Amazon EC2 operating system. Examples are Windows or\n Linux.
\nPURCHASE_TYPE - The reservation type of the purchase that this usage is related to.\n Examples include On-Demand Instances and Standard Reserved Instances.
\nRESERVATION_ID - The unique identifier for an Amazon Web Services Reservation\n Instance.
\nSAVINGS_PLAN_ARN - The unique identifier for your Savings Plans.
\nSAVINGS_PLANS_TYPE - Type of Savings Plans (EC2 Instance or Compute).
\nSERVICE - The Amazon Web Services service such as Amazon DynamoDB.
\nTENANCY - The tenancy of a resource. Examples are shared or dedicated.
\nUSAGE_TYPE - The type of usage. An example is DataTransfer-In-Bytes. The response\n for the GetDimensionValues
operation includes a unit attribute. Examples\n include GB and Hrs.
USAGE_TYPE_GROUP - The grouping of common usage types. An example is Amazon EC2:\n CloudWatch – Alarms. The response for this operation includes a unit attribute.
\nREGION - The Amazon Web Services Region.
\nRECORD_TYPE - The different types of charges such as Reserved Instance (RI) fees,\n usage costs, tax refunds, and credits.
\nRESOURCE_ID - The unique identifier of the resource. ResourceId is an opt-in\n feature only available for last 14 days for EC2-Compute Service.
\nIf you set the context to RESERVATIONS
, you can use the following\n dimensions for searching:
AZ - The Availability Zone. An example is us-east-1a
.
CACHE_ENGINE - The Amazon ElastiCache operating system. Examples are Windows or\n Linux.
\nDEPLOYMENT_OPTION - The scope of Amazon Relational Database Service deployments.\n Valid values are SingleAZ
and MultiAZ
.
INSTANCE_TYPE - The type of Amazon EC2 instance. An example is\n m4.xlarge
.
LINKED_ACCOUNT - The description in the attribute map that includes the full name\n of the member account. The value field contains the Amazon Web Services ID of the member\n account.
\nPLATFORM - The Amazon EC2 operating system. Examples are Windows or\n Linux.
\nREGION - The Amazon Web Services Region.
\nSCOPE (Utilization only) - The scope of a Reserved Instance (RI). Values are\n regional or a single Availability Zone.
\nTAG (Coverage only) - The tags that are associated with a Reserved Instance\n (RI).
\nTENANCY - The tenancy of a resource. Examples are shared or dedicated.
\nIf you set the context to SAVINGS_PLANS
, you can use the following\n dimensions for searching:
SAVINGS_PLANS_TYPE - Type of Savings Plans (EC2 Instance or Compute)
\nPAYMENT_OPTION - The payment option for the given Savings Plans (for example, All\n Upfront)
\nREGION - The Amazon Web Services Region.
\nINSTANCE_TYPE_FAMILY - The family of instances (For example,\n m5
)
LINKED_ACCOUNT - The description in the attribute map that includes the full name\n of the member account. The value field contains the Amazon Web Services ID of the member\n account.
\nSAVINGS_PLAN_ARN - The unique identifier for your Savings Plans.
\nThe context for the call to GetDimensionValues
. This can be\n RESERVATIONS
or COST_AND_USAGE
. The default value is\n COST_AND_USAGE
. If the context is set to RESERVATIONS
, the\n resulting dimension values can be used in the GetReservationUtilization
\n operation. If the context is set to COST_AND_USAGE
, the resulting dimension\n values can be used in the GetCostAndUsage
operation.
If you set the context to COST_AND_USAGE
, you can use the following\n dimensions for searching:
AZ - The Availability Zone. An example is us-east-1a
.
BILLING_ENTITY - The Amazon Web Services seller that your account is with. Possible\n values are the following:
\n- Amazon Web Services(Amazon Web Services): The entity that sells Amazon Web Servicesservices.
\n- AISPL (Amazon Internet Services Pvt. Ltd.): The local Indian entity that's an acting\n reseller for Amazon Web Servicesservices in India.
\n- Amazon Web Services Marketplace: The entity that supports the sale of solutions that are built on\n Amazon Web Services by third-party software providers.
\nCACHE_ENGINE - The Amazon ElastiCache operating system. Examples are Windows or\n Linux.
\nDEPLOYMENT_OPTION - The scope of Amazon Relational Database Service deployments.\n Valid values are SingleAZ
and MultiAZ
.
DATABASE_ENGINE - The Amazon Relational Database Service database. Examples are\n Aurora or MySQL.
\nINSTANCE_TYPE - The type of Amazon EC2 instance. An example is\n m4.xlarge
.
INSTANCE_TYPE_FAMILY - A family of instance types optimized to fit different use\n cases. Examples are Compute Optimized
(for example, C4
,\n C5
, C6g
, and C7g
), Memory\n Optimization
(for example, R4
, R5n
, R5b
,\n and R6g
).
INVOICING_ENTITY - The name of the entity that issues the Amazon Web Services\n invoice.
\nLEGAL_ENTITY_NAME - The name of the organization that sells you Amazon Web Services\n services, such as Amazon Web Services.
\nLINKED_ACCOUNT - The description in the attribute map that includes the full name\n of the member account. The value field contains the Amazon Web Services ID of the member\n account.
\nOPERATING_SYSTEM - The operating system. Examples are Windows or Linux.
\nOPERATION - The action performed. Examples include RunInstance
and\n CreateBucket
.
PLATFORM - The Amazon EC2 operating system. Examples are Windows or\n Linux.
\nPURCHASE_TYPE - The reservation type of the purchase that this usage is related to.\n Examples include On-Demand Instances and Standard Reserved Instances.
\nRESERVATION_ID - The unique identifier for an Amazon Web Services Reservation\n Instance.
\nSAVINGS_PLAN_ARN - The unique identifier for your Savings Plans.
\nSAVINGS_PLANS_TYPE - Type of Savings Plans (EC2 Instance or Compute).
\nSERVICE - The Amazon Web Services service such as Amazon DynamoDB.
\nTENANCY - The tenancy of a resource. Examples are shared or dedicated.
\nUSAGE_TYPE - The type of usage. An example is DataTransfer-In-Bytes. The response\n for the GetDimensionValues
operation includes a unit attribute. Examples\n include GB and Hrs.
USAGE_TYPE_GROUP - The grouping of common usage types. An example is Amazon EC2:\n CloudWatch – Alarms. The response for this operation includes a unit attribute.
\nREGION - The Amazon Web Services Region.
\nRECORD_TYPE - The different types of charges such as Reserved Instance (RI) fees,\n usage costs, tax refunds, and credits.
\nRESOURCE_ID - The unique identifier of the resource. ResourceId is an opt-in\n feature only available for last 14 days for EC2-Compute Service.
\nIf you set the context to RESERVATIONS
, you can use the following\n dimensions for searching:
AZ - The Availability Zone. An example is us-east-1a
.
CACHE_ENGINE - The Amazon ElastiCache operating system. Examples are Windows or\n Linux.
\nDEPLOYMENT_OPTION - The scope of Amazon Relational Database Service deployments.\n Valid values are SingleAZ
and MultiAZ
.
INSTANCE_TYPE - The type of Amazon EC2 instance. An example is\n m4.xlarge
.
LINKED_ACCOUNT - The description in the attribute map that includes the full name\n of the member account. The value field contains the Amazon Web Services ID of the member\n account.
\nPLATFORM - The Amazon EC2 operating system. Examples are Windows or\n Linux.
\nREGION - The Amazon Web Services Region.
\nSCOPE (Utilization only) - The scope of a Reserved Instance (RI). Values are\n regional or a single Availability Zone.
\nTAG (Coverage only) - The tags that are associated with a Reserved Instance\n (RI).
\nTENANCY - The tenancy of a resource. Examples are shared or dedicated.
\nIf you set the context to SAVINGS_PLANS
, you can use the following\n dimensions for searching:
SAVINGS_PLANS_TYPE - Type of Savings Plans (EC2 Instance or Compute)
\nPAYMENT_OPTION - The payment option for the given Savings Plans (for example, All\n Upfront)
\nREGION - The Amazon Web Services Region.
\nINSTANCE_TYPE_FAMILY - The family of instances (For example,\n m5
)
LINKED_ACCOUNT - The description in the attribute map that includes the full name\n of the member account. The value field contains the Amazon Web Services ID of the member\n account.
\nSAVINGS_PLAN_ARN - The unique identifier for your Savings Plans.
\nThe average utilization of your instances. Amazon Web Services uses this to calculate\n your recommended reservation purchases.
" + "smithy.api#documentation": "The average utilization of your recommendations. Amazon Web Services uses this to\n calculate your recommended reservation purchases.
" } }, "EstimatedBreakEvenInMonths": { "target": "com.amazonaws.costexplorer#GenericString", "traits": { - "smithy.api#documentation": "How long Amazon Web Services estimates that it takes for this instance to start saving\n you money, in months.
" + "smithy.api#documentation": "How long Amazon Web Services estimates that it takes for this recommendation to start\n saving you money, in months.
" } }, "CurrencyCode": { "target": "com.amazonaws.costexplorer#GenericString", "traits": { - "smithy.api#documentation": "The currency code that Amazon Web Services used to calculate the costs for this\n instance.
" + "smithy.api#documentation": "The currency code that Amazon Web Services used to calculate the costs for this\n recommendation.
" } }, "EstimatedMonthlySavingsAmount": { @@ -7346,13 +7366,43 @@ "UpfrontCost": { "target": "com.amazonaws.costexplorer#GenericString", "traits": { - "smithy.api#documentation": "How much purchasing this instance costs you upfront.
" + "smithy.api#documentation": "How much purchasing this recommendation costs you upfront.
" } }, "RecurringStandardMonthlyCost": { "target": "com.amazonaws.costexplorer#GenericString", "traits": { - "smithy.api#documentation": "How much purchasing this instance costs you on a monthly basis.
" + "smithy.api#documentation": "How much purchasing this recommendation costs you on a monthly basis.
" + } + }, + "ReservedCapacityDetails": { + "target": "com.amazonaws.costexplorer#ReservedCapacityDetails", + "traits": { + "smithy.api#documentation": "Details about the reservations that Amazon Web Services recommends that you\n purchase.
" + } + }, + "RecommendedNumberOfCapacityUnitsToPurchase": { + "target": "com.amazonaws.costexplorer#GenericString", + "traits": { + "smithy.api#documentation": "The number of reserved capacity units that Amazon Web Services recommends that you\n purchase.
" + } + }, + "MinimumNumberOfCapacityUnitsUsedPerHour": { + "target": "com.amazonaws.costexplorer#GenericString", + "traits": { + "smithy.api#documentation": "The minimum number of provisioned capacity units that you used in an hour during the\n historical period. Amazon Web Services uses this to calculate your recommended\n reservation purchases.
" + } + }, + "MaximumNumberOfCapacityUnitsUsedPerHour": { + "target": "com.amazonaws.costexplorer#GenericString", + "traits": { + "smithy.api#documentation": "The maximum number of provisioned capacity units that you used in an hour during the\n historical period. Amazon Web Services uses this to calculate your recommended\n reservation purchases.
" + } + }, + "AverageNumberOfCapacityUnitsUsedPerHour": { + "target": "com.amazonaws.costexplorer#GenericString", + "traits": { + "smithy.api#documentation": "The average number of provisioned capacity units that you used in an hour during the\n historical period. Amazon Web Services uses this to calculate your recommended\n reservation purchases.
" } } }, @@ -7462,6 +7512,20 @@ "target": "com.amazonaws.costexplorer#ReservationUtilizationGroup" } }, + "com.amazonaws.costexplorer#ReservedCapacityDetails": { + "type": "structure", + "members": { + "DynamoDBCapacityDetails": { + "target": "com.amazonaws.costexplorer#DynamoDBCapacityDetails", + "traits": { + "smithy.api#documentation": "The DynamoDB reservations that Amazon Web Services recommends that you purchase.
" + } + } + }, + "traits": { + "smithy.api#documentation": "Details about the reservations that Amazon Web Services recommends that you\n purchase.
" + } + }, "com.amazonaws.costexplorer#ReservedHours": { "type": "string" }, @@ -7777,7 +7841,7 @@ "Service": { "target": "com.amazonaws.costexplorer#GenericString", "traits": { - "smithy.api#documentation": "The Amazon Web Service name that's associated with the cost anomaly.
" + "smithy.api#documentation": "The Amazon Web Servicesservice name that's associated with the cost anomaly.
" } }, "Region": { @@ -7806,7 +7870,7 @@ } }, "traits": { - "smithy.api#documentation": "The combination of Amazon Web Service, linked account, linked account name,\n Region, and usage type where a cost anomaly is observed. The linked account name will\n only be available when the account name can be identified.
" + "smithy.api#documentation": "The combination of Amazon Web Servicesservice, linked account, linked account name,\n Region, and usage type where a cost anomaly is observed. The linked account name will\n only be available when the account name can be identified.
" } }, "com.amazonaws.costexplorer#RootCauses": { diff --git a/aws-models/directory-service-data.json b/aws-models/directory-service-data.json new file mode 100644 index 000000000000..04f9162bda5b --- /dev/null +++ b/aws-models/directory-service-data.json @@ -0,0 +1,3464 @@ +{ + "smithy": "2.0", + "shapes": { + "com.amazonaws.directoryservicedata#AccessDeniedException": { + "type": "structure", + "members": { + "Message": { + "target": "com.amazonaws.directoryservicedata#ExceptionMessage" + }, + "Reason": { + "target": "com.amazonaws.directoryservicedata#AccessDeniedReason", + "traits": { + "smithy.api#documentation": "Reason the request was unauthorized.
" + } + } + }, + "traits": { + "smithy.api#documentation": " You don't have permission to perform the request or access the directory. It can also\n occur when the DirectoryId
doesn't exist or the user, member, or group might be\n outside of your organizational unit (OU).
Make sure that you have the authentication and authorization to perform the action.\n Review the directory information in the request, and make sure that the object isn't outside\n of your OU.
", + "smithy.api#error": "client", + "smithy.api#httpError": 403 + } + }, + "com.amazonaws.directoryservicedata#AccessDeniedReason": { + "type": "enum", + "members": { + "IAM_AUTH": { + "target": "smithy.api#Unit", + "traits": { + "smithy.api#enumValue": "IAM_AUTH" + } + }, + "DIRECTORY_AUTH": { + "target": "smithy.api#Unit", + "traits": { + "smithy.api#enumValue": "DIRECTORY_AUTH" + } + }, + "DATA_DISABLED": { + "target": "smithy.api#Unit", + "traits": { + "smithy.api#enumValue": "DATA_DISABLED" + } + } + } + }, + "com.amazonaws.directoryservicedata#AddGroupMember": { + "type": "operation", + "input": { + "target": "com.amazonaws.directoryservicedata#AddGroupMemberRequest" + }, + "output": { + "target": "com.amazonaws.directoryservicedata#AddGroupMemberResult" + }, + "errors": [ + { + "target": "com.amazonaws.directoryservicedata#AccessDeniedException" + }, + { + "target": "com.amazonaws.directoryservicedata#ConflictException" + }, + { + "target": "com.amazonaws.directoryservicedata#DirectoryUnavailableException" + }, + { + "target": "com.amazonaws.directoryservicedata#InternalServerException" + }, + { + "target": "com.amazonaws.directoryservicedata#ResourceNotFoundException" + }, + { + "target": "com.amazonaws.directoryservicedata#ThrottlingException" + }, + { + "target": "com.amazonaws.directoryservicedata#ValidationException" + } + ], + "traits": { + "smithy.api#documentation": "Adds an existing user, group, or computer as a group member.
", + "smithy.api#http": { + "uri": "/GroupMemberships/AddGroupMember", + "method": "POST" + } + } + }, + "com.amazonaws.directoryservicedata#AddGroupMemberRequest": { + "type": "structure", + "members": { + "DirectoryId": { + "target": "com.amazonaws.directoryservicedata#DirectoryId", + "traits": { + "smithy.api#documentation": "The identifier (ID) of the directory that's associated with the group.
", + "smithy.api#httpQuery": "DirectoryId", + "smithy.api#required": {} + } + }, + "GroupName": { + "target": "com.amazonaws.directoryservicedata#GroupName", + "traits": { + "smithy.api#documentation": "The name of the group.
", + "smithy.api#required": {} + } + }, + "MemberName": { + "target": "com.amazonaws.directoryservicedata#MemberName", + "traits": { + "smithy.api#documentation": " The SAMAccountName
of the user, group, or computer to add as a group member.\n
The domain name that's associated with the group member. This parameter is required only\n when adding a member outside of your Managed Microsoft AD domain to a group inside of your\n Managed Microsoft AD domain. This parameter defaults to the Managed Microsoft AD domain.
\nThis parameter is case insensitive.
\nA unique and case-sensitive identifier that you provide to make sure the idempotency of\n the request, so multiple identical calls have the same effect as one single call.
\nA client token is valid for 8 hours after the first request that uses it completes. After\n 8 hours, any request with the same client token is treated as a new request. If the request\n succeeds, any future uses of that token will be idempotent for another 8 hours.
\n If you submit a request with the same client token but change one of the other parameters\n within the 8-hour idempotency window, Directory Service Data returns an ConflictException
.
This parameter is optional when using the CLI or SDK.
\nIndicates that the attribute type value is a string. For example:
\n\n \"S\": \"S Group\"
\n
Indicates that the attribute type value is a number. For example:
\n\n \"N\": \"16\"
\n
Indicates that the attribute type value is a boolean. For example:
\n\n \"BOOL\": true
\n
Indicates that the attribute type value is a string set. For example:
\n\n \"SS\": [\"sample_service_class/host.sample.com:1234/sample_service_name_1\",\n \"sample_service_class/host.sample.com:1234/sample_service_name_2\"]
\n
The data type for an attribute. Each attribute value is described as a name-value pair.\n The name is the AD schema name, and the value is the data itself. For a list of supported\n attributes, see Directory Service Data Attributes.\n
" + } + }, + "com.amazonaws.directoryservicedata#Attributes": { + "type": "map", + "key": { + "target": "com.amazonaws.directoryservicedata#LdapDisplayName" + }, + "value": { + "target": "com.amazonaws.directoryservicedata#AttributeValue" + }, + "traits": { + "smithy.api#length": { + "min": 1, + "max": 25 + } + } + }, + "com.amazonaws.directoryservicedata#BooleanAttributeValue": { + "type": "boolean", + "traits": { + "smithy.api#sensitive": {} + } + }, + "com.amazonaws.directoryservicedata#ClientToken": { + "type": "string", + "traits": { + "smithy.api#length": { + "min": 1, + "max": 128 + }, + "smithy.api#pattern": "^[\\x00-\\x7F]+$" + } + }, + "com.amazonaws.directoryservicedata#ConflictException": { + "type": "structure", + "members": { + "Message": { + "target": "com.amazonaws.directoryservicedata#ExceptionMessage" + } + }, + "traits": { + "smithy.api#documentation": "This error will occur when you try to create a resource that conflicts with an existing\n object. It can also occur when adding a member to a group that the member is already\n in.
\nThis error can be caused by a request sent within the 8-hour idempotency window with the\n same client token but different input parameters. Client tokens should not be re-used across\n different requests. After 8 hours, any request with the same client token is treated as a new\n request.
", + "smithy.api#error": "client", + "smithy.api#httpError": 409 + } + }, + "com.amazonaws.directoryservicedata#CreateGroup": { + "type": "operation", + "input": { + "target": "com.amazonaws.directoryservicedata#CreateGroupRequest" + }, + "output": { + "target": "com.amazonaws.directoryservicedata#CreateGroupResult" + }, + "errors": [ + { + "target": "com.amazonaws.directoryservicedata#AccessDeniedException" + }, + { + "target": "com.amazonaws.directoryservicedata#ConflictException" + }, + { + "target": "com.amazonaws.directoryservicedata#DirectoryUnavailableException" + }, + { + "target": "com.amazonaws.directoryservicedata#InternalServerException" + }, + { + "target": "com.amazonaws.directoryservicedata#ThrottlingException" + }, + { + "target": "com.amazonaws.directoryservicedata#ValidationException" + } + ], + "traits": { + "smithy.api#documentation": "Creates a new group.
", + "smithy.api#http": { + "uri": "/Groups/CreateGroup", + "method": "POST" + } + } + }, + "com.amazonaws.directoryservicedata#CreateGroupRequest": { + "type": "structure", + "members": { + "DirectoryId": { + "target": "com.amazonaws.directoryservicedata#DirectoryId", + "traits": { + "smithy.api#documentation": "The identifier (ID) of the directory that's associated with the group.
", + "smithy.api#httpQuery": "DirectoryId", + "smithy.api#required": {} + } + }, + "SAMAccountName": { + "target": "com.amazonaws.directoryservicedata#GroupName", + "traits": { + "smithy.api#documentation": "The name of the group.
", + "smithy.api#required": {} + } + }, + "GroupType": { + "target": "com.amazonaws.directoryservicedata#GroupType", + "traits": { + "smithy.api#documentation": "The AD group type. For details, see Active Directory security group type.
" + } + }, + "GroupScope": { + "target": "com.amazonaws.directoryservicedata#GroupScope", + "traits": { + "smithy.api#documentation": "The scope of the AD group. For details, see Active Directory security group scope.
" + } + }, + "OtherAttributes": { + "target": "com.amazonaws.directoryservicedata#Attributes", + "traits": { + "smithy.api#documentation": "An expression that defines one or more attributes with the data type and value of each\n attribute.
" + } + }, + "ClientToken": { + "target": "com.amazonaws.directoryservicedata#ClientToken", + "traits": { + "smithy.api#documentation": "A unique and case-sensitive identifier that you provide to make sure the idempotency of\n the request, so multiple identical calls have the same effect as one single call.
\nA client token is valid for 8 hours after the first request that uses it completes. After\n 8 hours, any request with the same client token is treated as a new request. If the request\n succeeds, any future uses of that token will be idempotent for another 8 hours.
\n If you submit a request with the same client token but change one of the other parameters\n within the 8-hour idempotency window, Directory Service Data returns an ConflictException
.
This parameter is optional when using the CLI or SDK.
\nThe identifier (ID) of the directory that's associated with the group.
" + } + }, + "SAMAccountName": { + "target": "com.amazonaws.directoryservicedata#GroupName", + "traits": { + "smithy.api#documentation": "The name of the group.
" + } + }, + "SID": { + "target": "com.amazonaws.directoryservicedata#SID", + "traits": { + "smithy.api#documentation": "The unique security identifier (SID) of the group.
" + } + } + }, + "traits": { + "smithy.api#output": {} + } + }, + "com.amazonaws.directoryservicedata#CreateUser": { + "type": "operation", + "input": { + "target": "com.amazonaws.directoryservicedata#CreateUserRequest" + }, + "output": { + "target": "com.amazonaws.directoryservicedata#CreateUserResult" + }, + "errors": [ + { + "target": "com.amazonaws.directoryservicedata#AccessDeniedException" + }, + { + "target": "com.amazonaws.directoryservicedata#ConflictException" + }, + { + "target": "com.amazonaws.directoryservicedata#DirectoryUnavailableException" + }, + { + "target": "com.amazonaws.directoryservicedata#InternalServerException" + }, + { + "target": "com.amazonaws.directoryservicedata#ThrottlingException" + }, + { + "target": "com.amazonaws.directoryservicedata#ValidationException" + } + ], + "traits": { + "smithy.api#documentation": "Creates a new user.
", + "smithy.api#http": { + "uri": "/Users/CreateUser", + "method": "POST" + } + } + }, + "com.amazonaws.directoryservicedata#CreateUserRequest": { + "type": "structure", + "members": { + "DirectoryId": { + "target": "com.amazonaws.directoryservicedata#DirectoryId", + "traits": { + "smithy.api#documentation": "The identifier (ID) of the directory that’s associated with the user.
", + "smithy.api#httpQuery": "DirectoryId", + "smithy.api#required": {} + } + }, + "SAMAccountName": { + "target": "com.amazonaws.directoryservicedata#UserName", + "traits": { + "smithy.api#documentation": "The name of the user.
", + "smithy.api#required": {} + } + }, + "EmailAddress": { + "target": "com.amazonaws.directoryservicedata#EmailAddress", + "traits": { + "smithy.api#documentation": "The email address of the user.
" + } + }, + "GivenName": { + "target": "com.amazonaws.directoryservicedata#GivenName", + "traits": { + "smithy.api#documentation": "The first name of the user.
" + } + }, + "Surname": { + "target": "com.amazonaws.directoryservicedata#Surname", + "traits": { + "smithy.api#documentation": "The last name of the user.
" + } + }, + "OtherAttributes": { + "target": "com.amazonaws.directoryservicedata#Attributes", + "traits": { + "smithy.api#documentation": "An expression that defines one or more attribute names with the data type and value of\n each attribute. A key is an attribute name, and the value is a list of maps. For a list of\n supported attributes, see Directory Service Data Attributes.
\nAttribute names are case insensitive.
\nA unique and case-sensitive identifier that you provide to make sure the idempotency of\n the request, so multiple identical calls have the same effect as one single call.
\nA client token is valid for 8 hours after the first request that uses it completes. After\n 8 hours, any request with the same client token is treated as a new request. If the request\n succeeds, any future uses of that token will be idempotent for another 8 hours.
\n If you submit a request with the same client token but change one of the other parameters\n within the 8-hour idempotency window, Directory Service Data returns an ConflictException
.
This parameter is optional when using the CLI or SDK.
\nThe identifier (ID) of the directory where the address block is added.
" + } + }, + "SID": { + "target": "com.amazonaws.directoryservicedata#SID", + "traits": { + "smithy.api#documentation": "The unique security identifier (SID) of the user.
" + } + }, + "SAMAccountName": { + "target": "com.amazonaws.directoryservicedata#UserName", + "traits": { + "smithy.api#documentation": "The name of the user.
" + } + } + }, + "traits": { + "smithy.api#output": {} + } + }, + "com.amazonaws.directoryservicedata#DeleteGroup": { + "type": "operation", + "input": { + "target": "com.amazonaws.directoryservicedata#DeleteGroupRequest" + }, + "output": { + "target": "com.amazonaws.directoryservicedata#DeleteGroupResult" + }, + "errors": [ + { + "target": "com.amazonaws.directoryservicedata#AccessDeniedException" + }, + { + "target": "com.amazonaws.directoryservicedata#ConflictException" + }, + { + "target": "com.amazonaws.directoryservicedata#DirectoryUnavailableException" + }, + { + "target": "com.amazonaws.directoryservicedata#InternalServerException" + }, + { + "target": "com.amazonaws.directoryservicedata#ResourceNotFoundException" + }, + { + "target": "com.amazonaws.directoryservicedata#ThrottlingException" + }, + { + "target": "com.amazonaws.directoryservicedata#ValidationException" + } + ], + "traits": { + "smithy.api#documentation": "Deletes a group.
", + "smithy.api#http": { + "uri": "/Groups/DeleteGroup", + "method": "POST" + } + } + }, + "com.amazonaws.directoryservicedata#DeleteGroupRequest": { + "type": "structure", + "members": { + "DirectoryId": { + "target": "com.amazonaws.directoryservicedata#DirectoryId", + "traits": { + "smithy.api#documentation": "The identifier (ID) of the directory that's associated with the group.
", + "smithy.api#httpQuery": "DirectoryId", + "smithy.api#required": {} + } + }, + "SAMAccountName": { + "target": "com.amazonaws.directoryservicedata#GroupName", + "traits": { + "smithy.api#documentation": "The name of the group.
", + "smithy.api#required": {} + } + }, + "ClientToken": { + "target": "com.amazonaws.directoryservicedata#ClientToken", + "traits": { + "smithy.api#documentation": "A unique and case-sensitive identifier that you provide to make sure the idempotency of\n the request, so multiple identical calls have the same effect as one single call.
\nA client token is valid for 8 hours after the first request that uses it completes. After\n 8 hours, any request with the same client token is treated as a new request. If the request\n succeeds, any future uses of that token will be idempotent for another 8 hours.
\n If you submit a request with the same client token but change one of the other parameters\n within the 8-hour idempotency window, Directory Service Data returns an ConflictException
.
This parameter is optional when using the CLI or SDK.
\nDeletes a user.
", + "smithy.api#http": { + "uri": "/Users/DeleteUser", + "method": "POST" + } + } + }, + "com.amazonaws.directoryservicedata#DeleteUserRequest": { + "type": "structure", + "members": { + "DirectoryId": { + "target": "com.amazonaws.directoryservicedata#DirectoryId", + "traits": { + "smithy.api#documentation": "The identifier (ID) of the directory that's associated with the user.
", + "smithy.api#httpQuery": "DirectoryId", + "smithy.api#required": {} + } + }, + "SAMAccountName": { + "target": "com.amazonaws.directoryservicedata#UserName", + "traits": { + "smithy.api#documentation": "The name of the user.
", + "smithy.api#required": {} + } + }, + "ClientToken": { + "target": "com.amazonaws.directoryservicedata#ClientToken", + "traits": { + "smithy.api#documentation": "A unique and case-sensitive identifier that you provide to make sure the idempotency of\n the request, so multiple identical calls have the same effect as one single call.
\nA client token is valid for 8 hours after the first request that uses it completes. After\n 8 hours, any request with the same client token is treated as a new request. If the request\n succeeds, any future uses of that token will be idempotent for another 8 hours.
\n If you submit a request with the same client token but change one of the other parameters\n within the 8-hour idempotency window, Directory Service Data returns an ConflictException
.
This parameter is optional when using the CLI or SDK.
\nReturns information about a specific group.
", + "smithy.api#http": { + "uri": "/Groups/DescribeGroup", + "method": "POST" + }, + "smithy.api#readonly": {}, + "smithy.test#smokeTests": [ + { + "id": "DescribeGroupFailure", + "params": { + "DirectoryId": "d-1111111111", + "SAMAccountName": "test-group" + }, + "expect": { + "failure": { + "errorId": "com.amazonaws.directoryservicedata#AccessDeniedException" + } + }, + "vendorParamsShape": "aws.test#AwsVendorParams", + "vendorParams": { + "region": "us-west-2" + } + } + ] + } + }, + "com.amazonaws.directoryservicedata#DescribeGroupRequest": { + "type": "structure", + "members": { + "DirectoryId": { + "target": "com.amazonaws.directoryservicedata#DirectoryId", + "traits": { + "smithy.api#documentation": "The Identifier (ID) of the directory associated with the group.
", + "smithy.api#httpQuery": "DirectoryId", + "smithy.api#required": {} + } + }, + "Realm": { + "target": "com.amazonaws.directoryservicedata#Realm", + "traits": { + "smithy.api#documentation": "The domain name that's associated with the group.
\nThis parameter is optional, so you can return groups outside of your Managed Microsoft AD\n domain. When no value is defined, only your Managed Microsoft AD groups are returned.
\nThis value is case insensitive.
\nThe name of the group.
", + "smithy.api#required": {} + } + }, + "OtherAttributes": { + "target": "com.amazonaws.directoryservicedata#LdapDisplayNameList", + "traits": { + "smithy.api#documentation": "One or more attributes to be returned for the group. For a list of supported attributes,\n see Directory Service Data Attributes.\n
" + } + } + }, + "traits": { + "smithy.api#input": {} + } + }, + "com.amazonaws.directoryservicedata#DescribeGroupResult": { + "type": "structure", + "members": { + "DirectoryId": { + "target": "com.amazonaws.directoryservicedata#DirectoryId", + "traits": { + "smithy.api#documentation": "The identifier (ID) of the directory that's associated with the group.
" + } + }, + "Realm": { + "target": "com.amazonaws.directoryservicedata#Realm", + "traits": { + "smithy.api#documentation": "The domain name that's associated with the group.
" + } + }, + "SID": { + "target": "com.amazonaws.directoryservicedata#SID", + "traits": { + "smithy.api#documentation": "The unique security identifier (SID) of the group.
" + } + }, + "SAMAccountName": { + "target": "com.amazonaws.directoryservicedata#GroupName", + "traits": { + "smithy.api#documentation": "The name of the group.
" + } + }, + "DistinguishedName": { + "target": "com.amazonaws.directoryservicedata#DistinguishedName", + "traits": { + "smithy.api#documentation": "The distinguished name of the object.
" + } + }, + "GroupType": { + "target": "com.amazonaws.directoryservicedata#GroupType", + "traits": { + "smithy.api#documentation": "The AD group type. For details, see Active Directory security group type.
" + } + }, + "GroupScope": { + "target": "com.amazonaws.directoryservicedata#GroupScope", + "traits": { + "smithy.api#documentation": "The scope of the AD group. For details, see Active Directory security groups.
" + } + }, + "OtherAttributes": { + "target": "com.amazonaws.directoryservicedata#Attributes", + "traits": { + "smithy.api#documentation": "The attribute values that are returned for the attribute names that are included in the\n request.
" + } + } + }, + "traits": { + "smithy.api#output": {} + } + }, + "com.amazonaws.directoryservicedata#DescribeUser": { + "type": "operation", + "input": { + "target": "com.amazonaws.directoryservicedata#DescribeUserRequest" + }, + "output": { + "target": "com.amazonaws.directoryservicedata#DescribeUserResult" + }, + "errors": [ + { + "target": "com.amazonaws.directoryservicedata#AccessDeniedException" + }, + { + "target": "com.amazonaws.directoryservicedata#DirectoryUnavailableException" + }, + { + "target": "com.amazonaws.directoryservicedata#InternalServerException" + }, + { + "target": "com.amazonaws.directoryservicedata#ResourceNotFoundException" + }, + { + "target": "com.amazonaws.directoryservicedata#ThrottlingException" + }, + { + "target": "com.amazonaws.directoryservicedata#ValidationException" + } + ], + "traits": { + "smithy.api#documentation": "Returns information about a specific user.
", + "smithy.api#http": { + "uri": "/Users/DescribeUser", + "method": "POST" + }, + "smithy.api#readonly": {}, + "smithy.test#smokeTests": [ + { + "id": "DescribeUserFailure", + "params": { + "DirectoryId": "d-1111111111", + "SAMAccountName": "test-user" + }, + "expect": { + "failure": { + "errorId": "com.amazonaws.directoryservicedata#AccessDeniedException" + } + }, + "vendorParamsShape": "aws.test#AwsVendorParams", + "vendorParams": { + "region": "us-west-2" + } + } + ] + } + }, + "com.amazonaws.directoryservicedata#DescribeUserRequest": { + "type": "structure", + "members": { + "DirectoryId": { + "target": "com.amazonaws.directoryservicedata#DirectoryId", + "traits": { + "smithy.api#documentation": "The identifier (ID) of the directory that's associated with the user.
", + "smithy.api#httpQuery": "DirectoryId", + "smithy.api#required": {} + } + }, + "SAMAccountName": { + "target": "com.amazonaws.directoryservicedata#UserName", + "traits": { + "smithy.api#documentation": "The name of the user.
", + "smithy.api#required": {} + } + }, + "OtherAttributes": { + "target": "com.amazonaws.directoryservicedata#LdapDisplayNameList", + "traits": { + "smithy.api#documentation": "One or more attribute names to be returned for the user. A key is an attribute name, and\n the value is a list of maps. For a list of supported attributes, see Directory Service Data Attributes.
" + } + }, + "Realm": { + "target": "com.amazonaws.directoryservicedata#Realm", + "traits": { + "smithy.api#documentation": "The domain name that's associated with the user.
\nThis parameter is optional, so you can return users outside your Managed Microsoft AD domain.\n When no value is defined, only your Managed Microsoft AD users are returned.
\nThis value is case insensitive.
\nThe identifier (ID) of the directory that's associated with the user.
" + } + }, + "Realm": { + "target": "com.amazonaws.directoryservicedata#Realm", + "traits": { + "smithy.api#documentation": "The domain name that's associated with the user.
" + } + }, + "SID": { + "target": "com.amazonaws.directoryservicedata#SID", + "traits": { + "smithy.api#documentation": "The unique security identifier (SID) of the user.
" + } + }, + "SAMAccountName": { + "target": "com.amazonaws.directoryservicedata#UserName", + "traits": { + "smithy.api#documentation": "The name of the user.
" + } + }, + "DistinguishedName": { + "target": "com.amazonaws.directoryservicedata#DistinguishedName", + "traits": { + "smithy.api#documentation": "The distinguished name of the object.
" + } + }, + "UserPrincipalName": { + "target": "com.amazonaws.directoryservicedata#UserPrincipalName", + "traits": { + "smithy.api#documentation": "The UPN that is an Internet-style login name for a user and is based on the Internet\n standard RFC 822. The UPN is shorter\n than the distinguished name and easier to remember.
" + } + }, + "EmailAddress": { + "target": "com.amazonaws.directoryservicedata#EmailAddress", + "traits": { + "smithy.api#documentation": "The email address of the user.
" + } + }, + "GivenName": { + "target": "com.amazonaws.directoryservicedata#GivenName", + "traits": { + "smithy.api#documentation": "The first name of the user.
" + } + }, + "Surname": { + "target": "com.amazonaws.directoryservicedata#Surname", + "traits": { + "smithy.api#documentation": "The last name of the user.
" + } + }, + "Enabled": { + "target": "smithy.api#Boolean", + "traits": { + "smithy.api#documentation": "Indicates whether the user account is active.
" + } + }, + "OtherAttributes": { + "target": "com.amazonaws.directoryservicedata#Attributes", + "traits": { + "smithy.api#documentation": "The attribute values that are returned for the attribute names that are included in the\n request.
\nAttribute names are case insensitive.
\nAmazon Web Services Directory Service Data is an extension of Directory Service. This API reference provides detailed information\n about Directory Service Data operations and object types.
\n With Directory Service Data, you can create, read, update, and delete users, groups, and memberships from\n your Managed Microsoft AD without additional costs and without deploying dedicated management\n instances. You can also perform built-in object management tasks across directories without\n direct network connectivity, which simplifies provisioning and access management to achieve\n fully automated deployments. Directory Service Data supports user and group write operations, such as\n CreateUser
and CreateGroup
, within the organizational unit (OU) of\n your Managed Microsoft AD. Directory Service Data supports read operations, such as ListUsers
and\n ListGroups
, on all users, groups, and group memberships within your\n Managed Microsoft AD and across trusted realms. Directory Service Data supports adding and removing group members in\n your OU and the Amazon Web Services Delegated Groups OU, so you can grant and deny access to specific roles\n and permissions. For more information, see Manage users and\n groups in the Directory Service Administration Guide.
Directory management operations and configuration changes made against the Directory Service\n API will also reflect in Directory Service Data API with eventual consistency. You can expect a short delay\n between management changes, such as adding a new directory trust and calling the Directory Service Data API\n for the newly created trusted realm.
\nDirectory Service Data connects to your Managed Microsoft AD domain controllers and performs operations on\n underlying directory objects. When you create your Managed Microsoft AD, you choose subnets for domain\n controllers that Directory Service creates on your behalf. If a domain controller is unavailable, Directory Service Data\n uses an available domain controller. As a result, you might notice eventual consistency while\n objects replicate from one domain controller to another domain controller. For more\n information, see What\n gets created in the Directory Service Administration Guide.\n Directory limits vary by Managed Microsoft AD edition:
\n\n Standard edition – Supports 8 transactions per\n second (TPS) for read operations and 4 TPS for write operations per directory. There's a\n concurrency limit of 10 concurrent requests.
\n\n Enterprise edition – Supports 16 transactions per\n second (TPS) for read operations and 8 TPS for write operations per directory. There's a\n concurrency limit of 10 concurrent requests.
\n\n Amazon Web Services Account - Supports a total of 100 TPS for\n Directory Service Data operations across all directories.
\nDirectory Service Data only supports the Managed Microsoft AD directory type and is only available in the primary\n Amazon Web Services Region. For more information, see Managed Microsoft AD\n and Primary vs additional Regions in the Directory Service Administration\n Guide.
", + "smithy.api#title": "AWS Directory Service Data", + "smithy.api#xmlNamespace": { + "uri": "http://directoryservicedata.amazonaws.com/doc/2023-05-31/" + }, + "smithy.rules#endpointRuleSet": { + "version": "1.0", + "parameters": { + "Region": { + "builtIn": "AWS::Region", + "required": false, + "documentation": "The AWS region used to dispatch the request.", + "type": "String" + }, + "UseDualStack": { + "builtIn": "AWS::UseDualStack", + "required": true, + "default": false, + "documentation": "When true, use the dual-stack endpoint. If the configured endpoint does not support dual-stack, dispatching the request MAY return an error.", + "type": "Boolean" + }, + "UseFIPS": { + "builtIn": "AWS::UseFIPS", + "required": true, + "default": false, + "documentation": "When true, send this request to the FIPS-compliant regional endpoint. If the configured endpoint does not have a FIPS compliant endpoint, dispatching the request will return an error.", + "type": "Boolean" + }, + "Endpoint": { + "builtIn": "SDK::Endpoint", + "required": false, + "documentation": "Override the endpoint used to send this request", + "type": "String" + } + }, + "rules": [ + { + "conditions": [ + { + "fn": "isSet", + "argv": [ + { + "ref": "Endpoint" + } + ] + } + ], + "rules": [ + { + "conditions": [ + { + "fn": "booleanEquals", + "argv": [ + { + "ref": "UseFIPS" + }, + true + ] + } + ], + "error": "Invalid Configuration: FIPS and custom endpoint are not supported", + "type": "error" + }, + { + "conditions": [], + "rules": [ + { + "conditions": [ + { + "fn": "booleanEquals", + "argv": [ + { + "ref": "UseDualStack" + }, + true + ] + } + ], + "error": "Invalid Configuration: Dualstack and custom endpoint are not supported", + "type": "error" + }, + { + "conditions": [], + "endpoint": { + "url": { + "ref": "Endpoint" + }, + "properties": {}, + "headers": {} + }, + "type": "endpoint" + } + ], + "type": "tree" + } + ], + "type": "tree" + }, + { + "conditions": [], + "rules": [ + { + "conditions": [ + { + "fn": "isSet", + "argv": [ + { + "ref": "Region" + } + ] + } + ], + "rules": [ + { + "conditions": [ + { + "fn": "aws.partition", + "argv": [ + { + "ref": "Region" + } + ], + "assign": "PartitionResult" + } + ], + "rules": [ + { + "conditions": [ + { + "fn": "booleanEquals", + "argv": [ + { + "ref": "UseFIPS" + }, + true + ] + }, + { + "fn": "booleanEquals", + "argv": [ + { + "ref": "UseDualStack" + }, + true + ] + } + ], + "rules": [ + { + "conditions": [ + { + "fn": "booleanEquals", + "argv": [ + true, + { + "fn": "getAttr", + "argv": [ + { + "ref": "PartitionResult" + }, + "supportsFIPS" + ] + } + ] + }, + { + "fn": "booleanEquals", + "argv": [ + true, + { + "fn": "getAttr", + "argv": [ + { + "ref": "PartitionResult" + }, + "supportsDualStack" + ] + } + ] + } + ], + "rules": [ + { + "conditions": [], + "rules": [ + { + "conditions": [], + "endpoint": { + "url": "https://ds-data-fips.{Region}.{PartitionResult#dualStackDnsSuffix}", + "properties": {}, + "headers": {} + }, + "type": "endpoint" + } + ], + "type": "tree" + } + ], + "type": "tree" + }, + { + "conditions": [], + "error": "FIPS and DualStack are enabled, but this partition does not support one or both", + "type": "error" + } + ], + "type": "tree" + }, + { + "conditions": [ + { + "fn": "booleanEquals", + "argv": [ + { + "ref": "UseFIPS" + }, + true + ] + } + ], + "rules": [ + { + "conditions": [ + { + "fn": "booleanEquals", + "argv": [ + { + "fn": "getAttr", + "argv": [ + { + "ref": "PartitionResult" + }, + "supportsFIPS" + ] + }, + true + ] + } + ], + "rules": [ + { + "conditions": [], + "rules": [ + { + "conditions": [], + "endpoint": { + "url": "https://ds-data-fips.{Region}.{PartitionResult#dnsSuffix}", + "properties": {}, + "headers": {} + }, + "type": "endpoint" + } + ], + "type": "tree" + } + ], + "type": "tree" + }, + { + "conditions": [], + "error": "FIPS is enabled but this partition does not support FIPS", + "type": "error" + } + ], + "type": "tree" + }, + { + "conditions": [ + { + "fn": "booleanEquals", + "argv": [ + { + "ref": "UseDualStack" + }, + true + ] + } + ], + "rules": [ + { + "conditions": [ + { + "fn": "booleanEquals", + "argv": [ + true, + { + "fn": "getAttr", + "argv": [ + { + "ref": "PartitionResult" + }, + "supportsDualStack" + ] + } + ] + } + ], + "rules": [ + { + "conditions": [], + "rules": [ + { + "conditions": [], + "endpoint": { + "url": "https://ds-data.{Region}.{PartitionResult#dualStackDnsSuffix}", + "properties": {}, + "headers": {} + }, + "type": "endpoint" + } + ], + "type": "tree" + } + ], + "type": "tree" + }, + { + "conditions": [], + "error": "DualStack is enabled but this partition does not support DualStack", + "type": "error" + } + ], + "type": "tree" + }, + { + "conditions": [], + "rules": [ + { + "conditions": [], + "endpoint": { + "url": "https://ds-data.{Region}.{PartitionResult#dnsSuffix}", + "properties": {}, + "headers": {} + }, + "type": "endpoint" + } + ], + "type": "tree" + } + ], + "type": "tree" + } + ], + "type": "tree" + }, + { + "conditions": [], + "error": "Invalid Configuration: Missing Region", + "type": "error" + } + ], + "type": "tree" + } + ] + }, + "smithy.rules#endpointTests": { + "testCases": [ + { + "documentation": "For region us-east-1 with FIPS enabled and DualStack enabled", + "expect": { + "endpoint": { + "url": "https://ds-data-fips.us-east-1.api.aws" + } + }, + "params": { + "Region": "us-east-1", + "UseFIPS": true, + "UseDualStack": true + } + }, + { + "documentation": "For region us-east-1 with FIPS enabled and DualStack disabled", + "expect": { + "endpoint": { + "url": "https://ds-data-fips.us-east-1.amazonaws.com" + } + }, + "params": { + "Region": "us-east-1", + "UseFIPS": true, + "UseDualStack": false + } + }, + { + "documentation": "For region us-east-1 with FIPS disabled and DualStack enabled", + "expect": { + "endpoint": { + "url": "https://ds-data.us-east-1.api.aws" + } + }, + "params": { + "Region": "us-east-1", + "UseFIPS": false, + "UseDualStack": true + } + }, + { + "documentation": "For region us-east-1 with FIPS disabled and DualStack disabled", + "expect": { + "endpoint": { + "url": "https://ds-data.us-east-1.amazonaws.com" + } + }, + "params": { + "Region": "us-east-1", + "UseFIPS": false, + "UseDualStack": false + } + }, + { + "documentation": "For region cn-north-1 with FIPS enabled and DualStack enabled", + "expect": { + "endpoint": { + "url": "https://ds-data-fips.cn-north-1.api.amazonwebservices.com.cn" + } + }, + "params": { + "Region": "cn-north-1", + "UseFIPS": true, + "UseDualStack": true + } + }, + { + "documentation": "For region cn-north-1 with FIPS enabled and DualStack disabled", + "expect": { + "endpoint": { + "url": "https://ds-data-fips.cn-north-1.amazonaws.com.cn" + } + }, + "params": { + "Region": "cn-north-1", + "UseFIPS": true, + "UseDualStack": false + } + }, + { + "documentation": "For region cn-north-1 with FIPS disabled and DualStack enabled", + "expect": { + "endpoint": { + "url": "https://ds-data.cn-north-1.api.amazonwebservices.com.cn" + } + }, + "params": { + "Region": "cn-north-1", + "UseFIPS": false, + "UseDualStack": true + } + }, + { + "documentation": "For region cn-north-1 with FIPS disabled and DualStack disabled", + "expect": { + "endpoint": { + "url": "https://ds-data.cn-north-1.amazonaws.com.cn" + } + }, + "params": { + "Region": "cn-north-1", + "UseFIPS": false, + "UseDualStack": false + } + }, + { + "documentation": "For region us-gov-east-1 with FIPS enabled and DualStack enabled", + "expect": { + "endpoint": { + "url": "https://ds-data-fips.us-gov-east-1.api.aws" + } + }, + "params": { + "Region": "us-gov-east-1", + "UseFIPS": true, + "UseDualStack": true + } + }, + { + "documentation": "For region us-gov-east-1 with FIPS enabled and DualStack disabled", + "expect": { + "endpoint": { + "url": "https://ds-data-fips.us-gov-east-1.amazonaws.com" + } + }, + "params": { + "Region": "us-gov-east-1", + "UseFIPS": true, + "UseDualStack": false + } + }, + { + "documentation": "For region us-gov-east-1 with FIPS disabled and DualStack enabled", + "expect": { + "endpoint": { + "url": "https://ds-data.us-gov-east-1.api.aws" + } + }, + "params": { + "Region": "us-gov-east-1", + "UseFIPS": false, + "UseDualStack": true + } + }, + { + "documentation": "For region us-gov-east-1 with FIPS disabled and DualStack disabled", + "expect": { + "endpoint": { + "url": "https://ds-data.us-gov-east-1.amazonaws.com" + } + }, + "params": { + "Region": "us-gov-east-1", + "UseFIPS": false, + "UseDualStack": false + } + }, + { + "documentation": "For region us-iso-east-1 with FIPS enabled and DualStack enabled", + "expect": { + "error": "FIPS and DualStack are enabled, but this partition does not support one or both" + }, + "params": { + "Region": "us-iso-east-1", + "UseFIPS": true, + "UseDualStack": true + } + }, + { + "documentation": "For region us-iso-east-1 with FIPS enabled and DualStack disabled", + "expect": { + "endpoint": { + "url": "https://ds-data-fips.us-iso-east-1.c2s.ic.gov" + } + }, + "params": { + "Region": "us-iso-east-1", + "UseFIPS": true, + "UseDualStack": false + } + }, + { + "documentation": "For region us-iso-east-1 with FIPS disabled and DualStack enabled", + "expect": { + "error": "DualStack is enabled but this partition does not support DualStack" + }, + "params": { + "Region": "us-iso-east-1", + "UseFIPS": false, + "UseDualStack": true + } + }, + { + "documentation": "For region us-iso-east-1 with FIPS disabled and DualStack disabled", + "expect": { + "endpoint": { + "url": "https://ds-data.us-iso-east-1.c2s.ic.gov" + } + }, + "params": { + "Region": "us-iso-east-1", + "UseFIPS": false, + "UseDualStack": false + } + }, + { + "documentation": "For region us-isob-east-1 with FIPS enabled and DualStack enabled", + "expect": { + "error": "FIPS and DualStack are enabled, but this partition does not support one or both" + }, + "params": { + "Region": "us-isob-east-1", + "UseFIPS": true, + "UseDualStack": true + } + }, + { + "documentation": "For region us-isob-east-1 with FIPS enabled and DualStack disabled", + "expect": { + "endpoint": { + "url": "https://ds-data-fips.us-isob-east-1.sc2s.sgov.gov" + } + }, + "params": { + "Region": "us-isob-east-1", + "UseFIPS": true, + "UseDualStack": false + } + }, + { + "documentation": "For region us-isob-east-1 with FIPS disabled and DualStack enabled", + "expect": { + "error": "DualStack is enabled but this partition does not support DualStack" + }, + "params": { + "Region": "us-isob-east-1", + "UseFIPS": false, + "UseDualStack": true + } + }, + { + "documentation": "For region us-isob-east-1 with FIPS disabled and DualStack disabled", + "expect": { + "endpoint": { + "url": "https://ds-data.us-isob-east-1.sc2s.sgov.gov" + } + }, + "params": { + "Region": "us-isob-east-1", + "UseFIPS": false, + "UseDualStack": false + } + }, + { + "documentation": "For custom endpoint with region set and fips disabled and dualstack disabled", + "expect": { + "endpoint": { + "url": "https://example.com" + } + }, + "params": { + "Region": "us-east-1", + "UseFIPS": false, + "UseDualStack": false, + "Endpoint": "https://example.com" + } + }, + { + "documentation": "For custom endpoint with region not set and fips disabled and dualstack disabled", + "expect": { + "endpoint": { + "url": "https://example.com" + } + }, + "params": { + "UseFIPS": false, + "UseDualStack": false, + "Endpoint": "https://example.com" + } + }, + { + "documentation": "For custom endpoint with fips enabled and dualstack disabled", + "expect": { + "error": "Invalid Configuration: FIPS and custom endpoint are not supported" + }, + "params": { + "Region": "us-east-1", + "UseFIPS": true, + "UseDualStack": false, + "Endpoint": "https://example.com" + } + }, + { + "documentation": "For custom endpoint with fips disabled and dualstack enabled", + "expect": { + "error": "Invalid Configuration: Dualstack and custom endpoint are not supported" + }, + "params": { + "Region": "us-east-1", + "UseFIPS": false, + "UseDualStack": true, + "Endpoint": "https://example.com" + } + }, + { + "documentation": "Missing region", + "expect": { + "error": "Invalid Configuration: Missing Region" + } + } + ], + "version": "1.0" + } + } + }, + "com.amazonaws.directoryservicedata#DirectoryUnavailableException": { + "type": "structure", + "members": { + "Message": { + "target": "com.amazonaws.directoryservicedata#ExceptionMessage" + }, + "Reason": { + "target": "com.amazonaws.directoryservicedata#DirectoryUnavailableReason", + "traits": { + "smithy.api#documentation": "Reason the request failed for the specified directory.
" + } + } + }, + "traits": { + "smithy.api#documentation": "The request could not be completed due to a problem in the configuration or current state\n of the specified directory.
", + "smithy.api#error": "client", + "smithy.api#httpError": 400, + "smithy.api#retryable": {} + } + }, + "com.amazonaws.directoryservicedata#DirectoryUnavailableReason": { + "type": "enum", + "members": { + "INVALID_DIRECTORY_STATE": { + "target": "smithy.api#Unit", + "traits": { + "smithy.api#enumValue": "INVALID_DIRECTORY_STATE" + } + }, + "DIRECTORY_TIMEOUT": { + "target": "smithy.api#Unit", + "traits": { + "smithy.api#enumValue": "DIRECTORY_TIMEOUT" + } + }, + "DIRECTORY_RESOURCES_EXCEEDED": { + "target": "smithy.api#Unit", + "traits": { + "smithy.api#enumValue": "DIRECTORY_RESOURCES_EXCEEDED" + } + }, + "NO_DISK_SPACE": { + "target": "smithy.api#Unit", + "traits": { + "smithy.api#enumValue": "NO_DISK_SPACE" + } + }, + "TRUST_AUTH_FAILURE": { + "target": "smithy.api#Unit", + "traits": { + "smithy.api#enumValue": "TRUST_AUTH_FAILURE" + } + } + } + }, + "com.amazonaws.directoryservicedata#DisableUser": { + "type": "operation", + "input": { + "target": "com.amazonaws.directoryservicedata#DisableUserRequest" + }, + "output": { + "target": "com.amazonaws.directoryservicedata#DisableUserResult" + }, + "errors": [ + { + "target": "com.amazonaws.directoryservicedata#AccessDeniedException" + }, + { + "target": "com.amazonaws.directoryservicedata#ConflictException" + }, + { + "target": "com.amazonaws.directoryservicedata#DirectoryUnavailableException" + }, + { + "target": "com.amazonaws.directoryservicedata#InternalServerException" + }, + { + "target": "com.amazonaws.directoryservicedata#ResourceNotFoundException" + }, + { + "target": "com.amazonaws.directoryservicedata#ThrottlingException" + }, + { + "target": "com.amazonaws.directoryservicedata#ValidationException" + } + ], + "traits": { + "smithy.api#documentation": "Deactivates an active user account. For information about how to enable an inactive user\n account, see ResetUserPassword\n in the Directory Service API Reference.
", + "smithy.api#http": { + "uri": "/Users/DisableUser", + "method": "POST" + } + } + }, + "com.amazonaws.directoryservicedata#DisableUserRequest": { + "type": "structure", + "members": { + "DirectoryId": { + "target": "com.amazonaws.directoryservicedata#DirectoryId", + "traits": { + "smithy.api#documentation": "The identifier (ID) of the directory that's associated with the user.
", + "smithy.api#httpQuery": "DirectoryId", + "smithy.api#required": {} + } + }, + "SAMAccountName": { + "target": "com.amazonaws.directoryservicedata#UserName", + "traits": { + "smithy.api#documentation": "The name of the user.
", + "smithy.api#required": {} + } + }, + "ClientToken": { + "target": "com.amazonaws.directoryservicedata#ClientToken", + "traits": { + "smithy.api#documentation": "A unique and case-sensitive identifier that you provide to make sure the idempotency of\n the request, so multiple identical calls have the same effect as one single call.
\nA client token is valid for 8 hours after the first request that uses it completes. After\n 8 hours, any request with the same client token is treated as a new request. If the request\n succeeds, any future uses of that token will be idempotent for another 8 hours.
\n If you submit a request with the same client token but change one of the other parameters\n within the 8-hour idempotency window, Directory Service Data returns an ConflictException
.
This parameter is optional when using the CLI or SDK.
\nThe unique security identifier (SID) of the group.
" + } + }, + "SAMAccountName": { + "target": "com.amazonaws.directoryservicedata#GroupName", + "traits": { + "smithy.api#documentation": "The name of the group.
", + "smithy.api#required": {} + } + }, + "DistinguishedName": { + "target": "com.amazonaws.directoryservicedata#DistinguishedName", + "traits": { + "smithy.api#documentation": "The distinguished name of the object.
" + } + }, + "GroupType": { + "target": "com.amazonaws.directoryservicedata#GroupType", + "traits": { + "smithy.api#documentation": "The AD group type. For details, see Active Directory security group type.
" + } + }, + "GroupScope": { + "target": "com.amazonaws.directoryservicedata#GroupScope", + "traits": { + "smithy.api#documentation": "The scope of the AD group. For details, see Active Directory security groups\n
" + } + }, + "OtherAttributes": { + "target": "com.amazonaws.directoryservicedata#Attributes", + "traits": { + "smithy.api#documentation": "An expression of one or more attributes, data types, and the values of a group.
" + } + } + }, + "traits": { + "smithy.api#documentation": "A group object that contains identifying information and attributes for a specified\n group.
" + } + }, + "com.amazonaws.directoryservicedata#GroupList": { + "type": "list", + "member": { + "target": "com.amazonaws.directoryservicedata#Group" + } + }, + "com.amazonaws.directoryservicedata#GroupName": { + "type": "string", + "traits": { + "smithy.api#length": { + "min": 1, + "max": 64 + }, + "smithy.api#pattern": "^[^:;|=+\"*?<>/\\\\,\\[\\]@]+$" + } + }, + "com.amazonaws.directoryservicedata#GroupScope": { + "type": "enum", + "members": { + "DOMAIN_LOCAL": { + "target": "smithy.api#Unit", + "traits": { + "smithy.api#enumValue": "DomainLocal" + } + }, + "GLOBAL": { + "target": "smithy.api#Unit", + "traits": { + "smithy.api#enumValue": "Global" + } + }, + "UNIVERSAL": { + "target": "smithy.api#Unit", + "traits": { + "smithy.api#enumValue": "Universal" + } + }, + "BUILTIN_LOCAL": { + "target": "smithy.api#Unit", + "traits": { + "smithy.api#enumValue": "BuiltinLocal" + } + } + } + }, + "com.amazonaws.directoryservicedata#GroupSummary": { + "type": "structure", + "members": { + "SID": { + "target": "com.amazonaws.directoryservicedata#SID", + "traits": { + "smithy.api#documentation": "The unique security identifier (SID) of the group.
", + "smithy.api#required": {} + } + }, + "SAMAccountName": { + "target": "com.amazonaws.directoryservicedata#GroupName", + "traits": { + "smithy.api#documentation": "The name of the group.
", + "smithy.api#required": {} + } + }, + "GroupType": { + "target": "com.amazonaws.directoryservicedata#GroupType", + "traits": { + "smithy.api#documentation": "The AD group type. For details, see Active Directory security group type.
", + "smithy.api#required": {} + } + }, + "GroupScope": { + "target": "com.amazonaws.directoryservicedata#GroupScope", + "traits": { + "smithy.api#documentation": "The scope of the AD group. For details, see Active Directory security groups.
", + "smithy.api#required": {} + } + } + }, + "traits": { + "smithy.api#documentation": "A structure containing a subset of fields of a group object from a directory.
" + } + }, + "com.amazonaws.directoryservicedata#GroupSummaryList": { + "type": "list", + "member": { + "target": "com.amazonaws.directoryservicedata#GroupSummary" + } + }, + "com.amazonaws.directoryservicedata#GroupType": { + "type": "enum", + "members": { + "DISTRIBUTION": { + "target": "smithy.api#Unit", + "traits": { + "smithy.api#enumValue": "Distribution" + } + }, + "SECURITY": { + "target": "smithy.api#Unit", + "traits": { + "smithy.api#enumValue": "Security" + } + } + } + }, + "com.amazonaws.directoryservicedata#InternalServerException": { + "type": "structure", + "members": { + "Message": { + "target": "com.amazonaws.directoryservicedata#ExceptionMessage" + } + }, + "traits": { + "smithy.api#documentation": "The operation didn't succeed because an internal error occurred. Try again later.
", + "smithy.api#error": "server", + "smithy.api#httpError": 500, + "smithy.api#retryable": {} + } + }, + "com.amazonaws.directoryservicedata#LdapDisplayName": { + "type": "string", + "traits": { + "smithy.api#length": { + "min": 1, + "max": 63 + }, + "smithy.api#pattern": "^[A-Za-z*][A-Za-z-*]*$" + } + }, + "com.amazonaws.directoryservicedata#LdapDisplayNameList": { + "type": "list", + "member": { + "target": "com.amazonaws.directoryservicedata#LdapDisplayName" + }, + "traits": { + "smithy.api#length": { + "min": 1, + "max": 25 + } + } + }, + "com.amazonaws.directoryservicedata#ListGroupMembers": { + "type": "operation", + "input": { + "target": "com.amazonaws.directoryservicedata#ListGroupMembersRequest" + }, + "output": { + "target": "com.amazonaws.directoryservicedata#ListGroupMembersResult" + }, + "errors": [ + { + "target": "com.amazonaws.directoryservicedata#AccessDeniedException" + }, + { + "target": "com.amazonaws.directoryservicedata#DirectoryUnavailableException" + }, + { + "target": "com.amazonaws.directoryservicedata#InternalServerException" + }, + { + "target": "com.amazonaws.directoryservicedata#ResourceNotFoundException" + }, + { + "target": "com.amazonaws.directoryservicedata#ThrottlingException" + }, + { + "target": "com.amazonaws.directoryservicedata#ValidationException" + } + ], + "traits": { + "smithy.api#documentation": "Returns member information for the specified group.
\n This operation supports pagination with the use of the NextToken
request and\n response parameters. If more results are available, the\n ListGroupMembers.NextToken
member contains a token that you pass in the next\n call to ListGroupMembers
. This retrieves the next set of items.
You can also specify a maximum number of return results with the MaxResults
\n parameter.
The identifier (ID) of the directory that's associated with the group.
", + "smithy.api#httpQuery": "DirectoryId", + "smithy.api#required": {} + } + }, + "Realm": { + "target": "com.amazonaws.directoryservicedata#Realm", + "traits": { + "smithy.api#documentation": "The domain name that's associated with the group.
\nThis parameter is optional, so you can return members from a group outside of your\n Managed Microsoft AD domain. When no value is defined, only members of your Managed Microsoft AD groups are\n returned.
\nThis value is case insensitive.
\nThe domain name that's associated with the group member. This parameter defaults to the\n Managed Microsoft AD domain.
\nThis parameter is optional and case insensitive.
\nThe name of the group.
", + "smithy.api#required": {} + } + }, + "NextToken": { + "target": "com.amazonaws.directoryservicedata#NextToken", + "traits": { + "smithy.api#documentation": "An encoded paging token for paginated calls that can be passed back to retrieve the next\n page.
" + } + }, + "MaxResults": { + "target": "com.amazonaws.directoryservicedata#MaxResults", + "traits": { + "smithy.api#documentation": "The maximum number of results to be returned per request.
" + } + } + }, + "traits": { + "smithy.api#input": {} + } + }, + "com.amazonaws.directoryservicedata#ListGroupMembersResult": { + "type": "structure", + "members": { + "DirectoryId": { + "target": "com.amazonaws.directoryservicedata#DirectoryId", + "traits": { + "smithy.api#documentation": "Identifier (ID) of the directory associated with the group.
" + } + }, + "Realm": { + "target": "com.amazonaws.directoryservicedata#Realm", + "traits": { + "smithy.api#documentation": "The domain name that's associated with the group.
" + } + }, + "MemberRealm": { + "target": "com.amazonaws.directoryservicedata#Realm", + "traits": { + "smithy.api#documentation": "The domain name that's associated with the member.
" + } + }, + "Members": { + "target": "com.amazonaws.directoryservicedata#MemberList", + "traits": { + "smithy.api#documentation": "The member information that the request returns.
" + } + }, + "NextToken": { + "target": "com.amazonaws.directoryservicedata#NextToken", + "traits": { + "smithy.api#documentation": "An encoded paging token for paginated calls that can be passed back to retrieve the next\n page.
" + } + } + }, + "traits": { + "smithy.api#output": {} + } + }, + "com.amazonaws.directoryservicedata#ListGroups": { + "type": "operation", + "input": { + "target": "com.amazonaws.directoryservicedata#ListGroupsRequest" + }, + "output": { + "target": "com.amazonaws.directoryservicedata#ListGroupsResult" + }, + "errors": [ + { + "target": "com.amazonaws.directoryservicedata#AccessDeniedException" + }, + { + "target": "com.amazonaws.directoryservicedata#DirectoryUnavailableException" + }, + { + "target": "com.amazonaws.directoryservicedata#InternalServerException" + }, + { + "target": "com.amazonaws.directoryservicedata#ThrottlingException" + }, + { + "target": "com.amazonaws.directoryservicedata#ValidationException" + } + ], + "traits": { + "smithy.api#documentation": "Returns group information for the specified directory.
\n This operation supports pagination with the use of the NextToken
request and\n response parameters. If more results are available, the ListGroups.NextToken
\n member contains a token that you pass in the next call to ListGroups
. This\n retrieves the next set of items.
You can also specify a maximum number of return results with the MaxResults
\n parameter.
Returns group information for the specified member.
\n This operation supports pagination with the use of the NextToken
request and\n response parameters. If more results are available, the\n ListGroupsForMember.NextToken
member contains a token that you pass in the next\n call to ListGroupsForMember
. This retrieves the next set of items.
You can also specify a maximum number of return results with the MaxResults
\n parameter.
The identifier (ID) of the directory that's associated with the member.
", + "smithy.api#httpQuery": "DirectoryId", + "smithy.api#required": {} + } + }, + "Realm": { + "target": "com.amazonaws.directoryservicedata#Realm", + "traits": { + "smithy.api#documentation": "The domain name that's associated with the group.
\nThis parameter is optional, so you can return groups outside of your Managed Microsoft AD\n domain. When no value is defined, only your Managed Microsoft AD groups are returned.
\nThis value is case insensitive and defaults to your Managed Microsoft AD domain.
\nThe domain name that's associated with the group member.
\nThis parameter is optional, so you can limit your results to the group members in a\n specific domain.
\n This parameter is case insensitive and defaults to Realm
\n
The SAMAccountName
of the user, group, or computer that's a member of the\n group.
An encoded paging token for paginated calls that can be passed back to retrieve the next\n page.
" + } + }, + "MaxResults": { + "target": "com.amazonaws.directoryservicedata#MaxResults", + "traits": { + "smithy.api#documentation": "The maximum number of results to be returned per request.
" + } + } + }, + "traits": { + "smithy.api#input": {} + } + }, + "com.amazonaws.directoryservicedata#ListGroupsForMemberResult": { + "type": "structure", + "members": { + "DirectoryId": { + "target": "com.amazonaws.directoryservicedata#DirectoryId", + "traits": { + "smithy.api#documentation": "The identifier (ID) of the directory that's associated with the member.
" + } + }, + "Realm": { + "target": "com.amazonaws.directoryservicedata#Realm", + "traits": { + "smithy.api#documentation": "The domain that's associated with the group.
" + } + }, + "MemberRealm": { + "target": "com.amazonaws.directoryservicedata#Realm", + "traits": { + "smithy.api#documentation": "The domain that's associated with the member.
" + } + }, + "Groups": { + "target": "com.amazonaws.directoryservicedata#GroupSummaryList", + "traits": { + "smithy.api#documentation": "The group information that the request returns.
" + } + }, + "NextToken": { + "target": "com.amazonaws.directoryservicedata#NextToken", + "traits": { + "smithy.api#documentation": "An encoded paging token for paginated calls that can be passed back to retrieve the next\n page.
" + } + } + }, + "traits": { + "smithy.api#output": {} + } + }, + "com.amazonaws.directoryservicedata#ListGroupsRequest": { + "type": "structure", + "members": { + "DirectoryId": { + "target": "com.amazonaws.directoryservicedata#DirectoryId", + "traits": { + "smithy.api#documentation": "The identifier (ID) of the directory that's associated with the group.
", + "smithy.api#httpQuery": "DirectoryId", + "smithy.api#required": {} + } + }, + "Realm": { + "target": "com.amazonaws.directoryservicedata#Realm", + "traits": { + "smithy.api#documentation": "The domain name associated with the directory.
\nThis parameter is optional, so you can return groups outside of your Managed Microsoft AD\n domain. When no value is defined, only your Managed Microsoft AD groups are returned.
\nThis value is case insensitive.
\nAn encoded paging token for paginated calls that can be passed back to retrieve the next\n page.
" + } + }, + "MaxResults": { + "target": "com.amazonaws.directoryservicedata#MaxResults", + "traits": { + "smithy.api#documentation": "The maximum number of results to be returned per request.
" + } + } + }, + "traits": { + "smithy.api#input": {} + } + }, + "com.amazonaws.directoryservicedata#ListGroupsResult": { + "type": "structure", + "members": { + "DirectoryId": { + "target": "com.amazonaws.directoryservicedata#DirectoryId", + "traits": { + "smithy.api#documentation": "The identifier (ID) of the directory that's associated with the group.
" + } + }, + "Realm": { + "target": "com.amazonaws.directoryservicedata#Realm", + "traits": { + "smithy.api#documentation": "The domain name associated with the group.
" + } + }, + "Groups": { + "target": "com.amazonaws.directoryservicedata#GroupSummaryList", + "traits": { + "smithy.api#documentation": "The group information that the request returns.
" + } + }, + "NextToken": { + "target": "com.amazonaws.directoryservicedata#NextToken", + "traits": { + "smithy.api#documentation": "An encoded paging token for paginated calls that can be passed back to retrieve the next\n page.
" + } + } + }, + "traits": { + "smithy.api#output": {} + } + }, + "com.amazonaws.directoryservicedata#ListUsers": { + "type": "operation", + "input": { + "target": "com.amazonaws.directoryservicedata#ListUsersRequest" + }, + "output": { + "target": "com.amazonaws.directoryservicedata#ListUsersResult" + }, + "errors": [ + { + "target": "com.amazonaws.directoryservicedata#AccessDeniedException" + }, + { + "target": "com.amazonaws.directoryservicedata#DirectoryUnavailableException" + }, + { + "target": "com.amazonaws.directoryservicedata#InternalServerException" + }, + { + "target": "com.amazonaws.directoryservicedata#ThrottlingException" + }, + { + "target": "com.amazonaws.directoryservicedata#ValidationException" + } + ], + "traits": { + "smithy.api#documentation": "Returns user information for the specified directory.
\n This operation supports pagination with the use of the NextToken
request and\n response parameters. If more results are available, the ListUsers.NextToken
\n member contains a token that you pass in the next call to ListUsers
. This\n retrieves the next set of items.
You can also specify a maximum number of return results with the MaxResults
\n parameter.
The identifier (ID) of the directory that's associated with the user.
", + "smithy.api#httpQuery": "DirectoryId", + "smithy.api#required": {} + } + }, + "Realm": { + "target": "com.amazonaws.directoryservicedata#Realm", + "traits": { + "smithy.api#documentation": "The domain name that's associated with the user.
\nThis parameter is optional, so you can return users outside of your Managed Microsoft AD\n domain. When no value is defined, only your Managed Microsoft AD users are returned.
\nThis value is case insensitive.
\nAn encoded paging token for paginated calls that can be passed back to retrieve the next\n page.
" + } + }, + "MaxResults": { + "target": "com.amazonaws.directoryservicedata#MaxResults", + "traits": { + "smithy.api#documentation": "The maximum number of results to be returned per request.
" + } + } + }, + "traits": { + "smithy.api#input": {} + } + }, + "com.amazonaws.directoryservicedata#ListUsersResult": { + "type": "structure", + "members": { + "DirectoryId": { + "target": "com.amazonaws.directoryservicedata#DirectoryId", + "traits": { + "smithy.api#documentation": "The identifier (ID) of the directory that's associated with the user.
" + } + }, + "Realm": { + "target": "com.amazonaws.directoryservicedata#Realm", + "traits": { + "smithy.api#documentation": "The domain that's associated with the user.
" + } + }, + "Users": { + "target": "com.amazonaws.directoryservicedata#UserSummaryList", + "traits": { + "smithy.api#documentation": "The user information that the request returns.
" + } + }, + "NextToken": { + "target": "com.amazonaws.directoryservicedata#NextToken", + "traits": { + "smithy.api#documentation": "An encoded paging token for paginated calls that can be passed back to retrieve the next\n page.
" + } + } + }, + "traits": { + "smithy.api#output": {} + } + }, + "com.amazonaws.directoryservicedata#MaxResults": { + "type": "integer", + "traits": { + "smithy.api#range": { + "min": 1, + "max": 250 + } + } + }, + "com.amazonaws.directoryservicedata#Member": { + "type": "structure", + "members": { + "SID": { + "target": "com.amazonaws.directoryservicedata#SID", + "traits": { + "smithy.api#documentation": "The unique security identifier (SID) of the group member.
", + "smithy.api#required": {} + } + }, + "SAMAccountName": { + "target": "com.amazonaws.directoryservicedata#MemberName", + "traits": { + "smithy.api#documentation": "The name of the group member.
", + "smithy.api#required": {} + } + }, + "MemberType": { + "target": "com.amazonaws.directoryservicedata#MemberType", + "traits": { + "smithy.api#documentation": "The AD type of the member object.
", + "smithy.api#required": {} + } + } + }, + "traits": { + "smithy.api#documentation": "A member object that contains identifying information for a specified member.
" + } + }, + "com.amazonaws.directoryservicedata#MemberList": { + "type": "list", + "member": { + "target": "com.amazonaws.directoryservicedata#Member" + } + }, + "com.amazonaws.directoryservicedata#MemberName": { + "type": "string", + "traits": { + "smithy.api#length": { + "min": 1, + "max": 63 + }, + "smithy.api#pattern": "^[^:;|=+\"*?<>/\\\\,\\[\\]@]+$" + } + }, + "com.amazonaws.directoryservicedata#MemberType": { + "type": "enum", + "members": { + "USER": { + "target": "smithy.api#Unit", + "traits": { + "smithy.api#enumValue": "USER" + } + }, + "GROUP": { + "target": "smithy.api#Unit", + "traits": { + "smithy.api#enumValue": "GROUP" + } + }, + "COMPUTER": { + "target": "smithy.api#Unit", + "traits": { + "smithy.api#enumValue": "COMPUTER" + } + } + } + }, + "com.amazonaws.directoryservicedata#NextToken": { + "type": "string", + "traits": { + "smithy.api#length": { + "min": 1, + "max": 6144 + }, + "smithy.api#sensitive": {} + } + }, + "com.amazonaws.directoryservicedata#NumberAttributeValue": { + "type": "long", + "traits": { + "smithy.api#sensitive": {} + } + }, + "com.amazonaws.directoryservicedata#Realm": { + "type": "string", + "traits": { + "smithy.api#length": { + "min": 1, + "max": 255 + }, + "smithy.api#pattern": "^([a-zA-Z0-9]+[\\\\.-])+([a-zA-Z0-9])+[.]?$" + } + }, + "com.amazonaws.directoryservicedata#RemoveGroupMember": { + "type": "operation", + "input": { + "target": "com.amazonaws.directoryservicedata#RemoveGroupMemberRequest" + }, + "output": { + "target": "com.amazonaws.directoryservicedata#RemoveGroupMemberResult" + }, + "errors": [ + { + "target": "com.amazonaws.directoryservicedata#AccessDeniedException" + }, + { + "target": "com.amazonaws.directoryservicedata#ConflictException" + }, + { + "target": "com.amazonaws.directoryservicedata#DirectoryUnavailableException" + }, + { + "target": "com.amazonaws.directoryservicedata#InternalServerException" + }, + { + "target": "com.amazonaws.directoryservicedata#ResourceNotFoundException" + }, + { + "target": "com.amazonaws.directoryservicedata#ThrottlingException" + }, + { + "target": "com.amazonaws.directoryservicedata#ValidationException" + } + ], + "traits": { + "smithy.api#documentation": "Removes a member from a group.
", + "smithy.api#http": { + "uri": "/GroupMemberships/RemoveGroupMember", + "method": "POST" + } + } + }, + "com.amazonaws.directoryservicedata#RemoveGroupMemberRequest": { + "type": "structure", + "members": { + "DirectoryId": { + "target": "com.amazonaws.directoryservicedata#DirectoryId", + "traits": { + "smithy.api#documentation": "The identifier (ID) of the directory that's associated with the member.
", + "smithy.api#httpQuery": "DirectoryId", + "smithy.api#required": {} + } + }, + "GroupName": { + "target": "com.amazonaws.directoryservicedata#GroupName", + "traits": { + "smithy.api#documentation": "The name of the group.
", + "smithy.api#required": {} + } + }, + "MemberName": { + "target": "com.amazonaws.directoryservicedata#MemberName", + "traits": { + "smithy.api#documentation": " The SAMAccountName
of the user, group, or computer to remove from the group.\n
The domain name that's associated with the group member. This parameter defaults to the\n Managed Microsoft AD domain.
\nThis parameter is optional and case insensitive.
\nA unique and case-sensitive identifier that you provide to make sure the idempotency of\n the request, so multiple identical calls have the same effect as one single call.
\nA client token is valid for 8 hours after the first request that uses it completes. After\n 8 hours, any request with the same client token is treated as a new request. If the request\n succeeds, any future uses of that token will be idempotent for another 8 hours.
\n If you submit a request with the same client token but change one of the other parameters\n within the 8-hour idempotency window, Directory Service Data returns an ConflictException
.
This parameter is optional when using the CLI or SDK.
\nThe resource couldn't be found.
", + "smithy.api#error": "client", + "smithy.api#httpError": 404 + } + }, + "com.amazonaws.directoryservicedata#SID": { + "type": "string", + "traits": { + "smithy.api#length": { + "min": 1, + "max": 256 + } + } + }, + "com.amazonaws.directoryservicedata#SearchGroups": { + "type": "operation", + "input": { + "target": "com.amazonaws.directoryservicedata#SearchGroupsRequest" + }, + "output": { + "target": "com.amazonaws.directoryservicedata#SearchGroupsResult" + }, + "errors": [ + { + "target": "com.amazonaws.directoryservicedata#AccessDeniedException" + }, + { + "target": "com.amazonaws.directoryservicedata#DirectoryUnavailableException" + }, + { + "target": "com.amazonaws.directoryservicedata#InternalServerException" + }, + { + "target": "com.amazonaws.directoryservicedata#ThrottlingException" + }, + { + "target": "com.amazonaws.directoryservicedata#ValidationException" + } + ], + "traits": { + "smithy.api#documentation": " Searches the specified directory for a group. You can find groups that match the\n SearchString
parameter with the value of their attributes included in the\n SearchString
parameter.
This operation supports pagination with the use of the NextToken
request and\n response parameters. If more results are available, the SearchGroups.NextToken
\n member contains a token that you pass in the next call to SearchGroups
. This\n retrieves the next set of items.
You can also specify a maximum number of return results with the MaxResults
\n parameter.
The identifier (ID) of the directory that's associated with the group.
", + "smithy.api#httpQuery": "DirectoryId", + "smithy.api#required": {} + } + }, + "SearchString": { + "target": "com.amazonaws.directoryservicedata#SearchString", + "traits": { + "smithy.api#documentation": "The attribute value that you want to search for.
\n Wildcard (*)
searches aren't supported. For a list of supported\n attributes, see Directory Service Data\n Attributes.
One or more data attributes that are used to search for a group. For a list of supported\n attributes, see Directory Service Data Attributes.\n
", + "smithy.api#required": {} + } + }, + "Realm": { + "target": "com.amazonaws.directoryservicedata#Realm", + "traits": { + "smithy.api#documentation": "The domain name that's associated with the group.
\nThis parameter is optional, so you can return groups outside of your Managed Microsoft AD\n domain. When no value is defined, only your Managed Microsoft AD groups are returned.
\nThis value is case insensitive.
\nAn encoded paging token for paginated calls that can be passed back to retrieve the next\n page.
" + } + }, + "MaxResults": { + "target": "com.amazonaws.directoryservicedata#MaxResults", + "traits": { + "smithy.api#documentation": "The maximum number of results to be returned per request.
" + } + } + }, + "traits": { + "smithy.api#input": {} + } + }, + "com.amazonaws.directoryservicedata#SearchGroupsResult": { + "type": "structure", + "members": { + "DirectoryId": { + "target": "com.amazonaws.directoryservicedata#DirectoryId", + "traits": { + "smithy.api#documentation": "The identifier (ID) of the directory that's associated with the group.
" + } + }, + "Realm": { + "target": "com.amazonaws.directoryservicedata#Realm", + "traits": { + "smithy.api#documentation": "The domain that's associated with the group.
" + } + }, + "Groups": { + "target": "com.amazonaws.directoryservicedata#GroupList", + "traits": { + "smithy.api#documentation": "The group information that the request returns.
" + } + }, + "NextToken": { + "target": "com.amazonaws.directoryservicedata#NextToken", + "traits": { + "smithy.api#documentation": "An encoded paging token for paginated calls that can be passed back to retrieve the next\n page.
" + } + } + }, + "traits": { + "smithy.api#output": {} + } + }, + "com.amazonaws.directoryservicedata#SearchString": { + "type": "string", + "traits": { + "smithy.api#length": { + "min": 1, + "max": 64 + }, + "smithy.api#sensitive": {} + } + }, + "com.amazonaws.directoryservicedata#SearchUsers": { + "type": "operation", + "input": { + "target": "com.amazonaws.directoryservicedata#SearchUsersRequest" + }, + "output": { + "target": "com.amazonaws.directoryservicedata#SearchUsersResult" + }, + "errors": [ + { + "target": "com.amazonaws.directoryservicedata#AccessDeniedException" + }, + { + "target": "com.amazonaws.directoryservicedata#DirectoryUnavailableException" + }, + { + "target": "com.amazonaws.directoryservicedata#InternalServerException" + }, + { + "target": "com.amazonaws.directoryservicedata#ThrottlingException" + }, + { + "target": "com.amazonaws.directoryservicedata#ValidationException" + } + ], + "traits": { + "smithy.api#documentation": " Searches the specified directory for a user. You can find users that match the\n SearchString
parameter with the value of their attributes included in the\n SearchString
parameter.
This operation supports pagination with the use of the NextToken
request and\n response parameters. If more results are available, the SearchUsers.NextToken
\n member contains a token that you pass in the next call to SearchUsers
. This\n retrieves the next set of items.
You can also specify a maximum number of return results with the MaxResults
\n parameter.
The identifier (ID) of the directory that's associated with the user.
", + "smithy.api#httpQuery": "DirectoryId", + "smithy.api#required": {} + } + }, + "Realm": { + "target": "com.amazonaws.directoryservicedata#Realm", + "traits": { + "smithy.api#documentation": "The domain name that's associated with the user.
\nThis parameter is optional, so you can return users outside of your Managed Microsoft AD\n domain. When no value is defined, only your Managed Microsoft AD users are returned.
\nThis value is case insensitive.
\nThe attribute value that you want to search for.
\n Wildcard (*)
searches aren't supported. For a list of supported\n attributes, see Directory Service Data\n Attributes.
One or more data attributes that are used to search for a user. For a list of supported\n attributes, see Directory Service Data Attributes.\n
", + "smithy.api#required": {} + } + }, + "NextToken": { + "target": "com.amazonaws.directoryservicedata#NextToken", + "traits": { + "smithy.api#documentation": "An encoded paging token for paginated calls that can be passed back to retrieve the next\n page.
" + } + }, + "MaxResults": { + "target": "com.amazonaws.directoryservicedata#MaxResults", + "traits": { + "smithy.api#documentation": "The maximum number of results to be returned per request.
" + } + } + }, + "traits": { + "smithy.api#input": {} + } + }, + "com.amazonaws.directoryservicedata#SearchUsersResult": { + "type": "structure", + "members": { + "DirectoryId": { + "target": "com.amazonaws.directoryservicedata#DirectoryId", + "traits": { + "smithy.api#documentation": "The identifier (ID) of the directory where the address block is added.
" + } + }, + "Realm": { + "target": "com.amazonaws.directoryservicedata#Realm", + "traits": { + "smithy.api#documentation": "The domain that's associated with the user.
" + } + }, + "Users": { + "target": "com.amazonaws.directoryservicedata#UserList", + "traits": { + "smithy.api#documentation": "The user information that the request returns.
" + } + }, + "NextToken": { + "target": "com.amazonaws.directoryservicedata#NextToken", + "traits": { + "smithy.api#documentation": "An encoded paging token for paginated calls that can be passed back to retrieve the next\n page.
" + } + } + }, + "traits": { + "smithy.api#output": {} + } + }, + "com.amazonaws.directoryservicedata#StringAttributeValue": { + "type": "string", + "traits": { + "smithy.api#length": { + "min": 1, + "max": 1024 + }, + "smithy.api#sensitive": {} + } + }, + "com.amazonaws.directoryservicedata#StringSetAttributeValue": { + "type": "list", + "member": { + "target": "com.amazonaws.directoryservicedata#StringAttributeValue" + }, + "traits": { + "smithy.api#length": { + "min": 0, + "max": 25 + }, + "smithy.api#sensitive": {} + } + }, + "com.amazonaws.directoryservicedata#Surname": { + "type": "string", + "traits": { + "smithy.api#length": { + "min": 1, + "max": 64 + }, + "smithy.api#sensitive": {} + } + }, + "com.amazonaws.directoryservicedata#ThrottlingException": { + "type": "structure", + "members": { + "Message": { + "target": "com.amazonaws.directoryservicedata#ExceptionMessage", + "traits": { + "smithy.api#required": {} + } + }, + "RetryAfterSeconds": { + "target": "smithy.api#Integer", + "traits": { + "smithy.api#documentation": "The recommended amount of seconds to retry after a throttling exception.
", + "smithy.api#httpHeader": "Retry-After" + } + } + }, + "traits": { + "smithy.api#documentation": "The limit on the number of requests per second has been exceeded.
", + "smithy.api#error": "client", + "smithy.api#httpError": 429, + "smithy.api#retryable": { + "throttling": true + } + } + }, + "com.amazonaws.directoryservicedata#UpdateGroup": { + "type": "operation", + "input": { + "target": "com.amazonaws.directoryservicedata#UpdateGroupRequest" + }, + "output": { + "target": "com.amazonaws.directoryservicedata#UpdateGroupResult" + }, + "errors": [ + { + "target": "com.amazonaws.directoryservicedata#AccessDeniedException" + }, + { + "target": "com.amazonaws.directoryservicedata#ConflictException" + }, + { + "target": "com.amazonaws.directoryservicedata#DirectoryUnavailableException" + }, + { + "target": "com.amazonaws.directoryservicedata#InternalServerException" + }, + { + "target": "com.amazonaws.directoryservicedata#ResourceNotFoundException" + }, + { + "target": "com.amazonaws.directoryservicedata#ThrottlingException" + }, + { + "target": "com.amazonaws.directoryservicedata#ValidationException" + } + ], + "traits": { + "smithy.api#documentation": "Updates group information.
", + "smithy.api#http": { + "uri": "/Groups/UpdateGroup", + "method": "POST" + } + } + }, + "com.amazonaws.directoryservicedata#UpdateGroupRequest": { + "type": "structure", + "members": { + "DirectoryId": { + "target": "com.amazonaws.directoryservicedata#DirectoryId", + "traits": { + "smithy.api#documentation": "The identifier (ID) of the directory that's associated with the group.
", + "smithy.api#httpQuery": "DirectoryId", + "smithy.api#required": {} + } + }, + "SAMAccountName": { + "target": "com.amazonaws.directoryservicedata#GroupName", + "traits": { + "smithy.api#documentation": "The name of the group.
", + "smithy.api#required": {} + } + }, + "GroupType": { + "target": "com.amazonaws.directoryservicedata#GroupType", + "traits": { + "smithy.api#documentation": "The AD group type. For details, see Active Directory security group type.
" + } + }, + "GroupScope": { + "target": "com.amazonaws.directoryservicedata#GroupScope", + "traits": { + "smithy.api#documentation": "The scope of the AD group. For details, see Active Directory security groups.
" + } + }, + "OtherAttributes": { + "target": "com.amazonaws.directoryservicedata#Attributes", + "traits": { + "smithy.api#documentation": "An expression that defines one or more attributes with the data type and the value of\n each attribute.
" + } + }, + "UpdateType": { + "target": "com.amazonaws.directoryservicedata#UpdateType", + "traits": { + "smithy.api#documentation": " The type of update to be performed. If no value exists for the attribute, use\n ADD
. Otherwise, use REPLACE
to change an attribute value or\n REMOVE
to clear the attribute value.
A unique and case-sensitive identifier that you provide to make sure the idempotency of\n the request, so multiple identical calls have the same effect as one single call.
\nA client token is valid for 8 hours after the first request that uses it completes. After\n 8 hours, any request with the same client token is treated as a new request. If the request\n succeeds, any future uses of that token will be idempotent for another 8 hours.
\n If you submit a request with the same client token but change one of the other parameters\n within the 8-hour idempotency window, Directory Service Data returns an ConflictException
.
This parameter is optional when using the CLI or SDK.
\nUpdates user information.
", + "smithy.api#http": { + "uri": "/Users/UpdateUser", + "method": "POST" + } + } + }, + "com.amazonaws.directoryservicedata#UpdateUserRequest": { + "type": "structure", + "members": { + "DirectoryId": { + "target": "com.amazonaws.directoryservicedata#DirectoryId", + "traits": { + "smithy.api#documentation": "The identifier (ID) of the directory that's associated with the user.
", + "smithy.api#httpQuery": "DirectoryId", + "smithy.api#required": {} + } + }, + "SAMAccountName": { + "target": "com.amazonaws.directoryservicedata#UserName", + "traits": { + "smithy.api#documentation": "The name of the user.
", + "smithy.api#required": {} + } + }, + "EmailAddress": { + "target": "com.amazonaws.directoryservicedata#EmailAddress", + "traits": { + "smithy.api#documentation": "The email address of the user.
" + } + }, + "GivenName": { + "target": "com.amazonaws.directoryservicedata#GivenName", + "traits": { + "smithy.api#documentation": "The first name of the user.
" + } + }, + "Surname": { + "target": "com.amazonaws.directoryservicedata#Surname", + "traits": { + "smithy.api#documentation": "The last name of the user.
" + } + }, + "OtherAttributes": { + "target": "com.amazonaws.directoryservicedata#Attributes", + "traits": { + "smithy.api#documentation": "An expression that defines one or more attribute names with the data type and value of\n each attribute. A key is an attribute name, and the value is a list of maps. For a list of\n supported attributes, see Directory Service Data Attributes.
\nAttribute names are case insensitive.
\n The type of update to be performed. If no value exists for the attribute, use\n ADD
. Otherwise, use REPLACE
to change an attribute value or\n REMOVE
to clear the attribute value.
A unique and case-sensitive identifier that you provide to make sure the idempotency of\n the request, so multiple identical calls have the same effect as one single call.
\nA client token is valid for 8 hours after the first request that uses it completes. After\n 8 hours, any request with the same client token is treated as a new request. If the request\n succeeds, any future uses of that token will be idempotent for another 8 hours.
\n If you submit a request with the same client token but change one of the other parameters\n within the 8-hour idempotency window, Directory Service Data returns an ConflictException
.
This parameter is optional when using the CLI or SDK.
\nThe unique security identifier (SID) of the user.
" + } + }, + "SAMAccountName": { + "target": "com.amazonaws.directoryservicedata#UserName", + "traits": { + "smithy.api#documentation": "The name of the user.
", + "smithy.api#required": {} + } + }, + "DistinguishedName": { + "target": "com.amazonaws.directoryservicedata#DistinguishedName", + "traits": { + "smithy.api#documentation": "The distinguished name of the object.
" + } + }, + "UserPrincipalName": { + "target": "com.amazonaws.directoryservicedata#UserPrincipalName", + "traits": { + "smithy.api#documentation": "The UPN that is an internet-style login name for a user and based on the internet\n standard RFC 822. The UPN is shorter\n than the distinguished name and easier to remember.
" + } + }, + "EmailAddress": { + "target": "com.amazonaws.directoryservicedata#EmailAddress", + "traits": { + "smithy.api#documentation": "The email address of the user.
" + } + }, + "GivenName": { + "target": "com.amazonaws.directoryservicedata#GivenName", + "traits": { + "smithy.api#documentation": "The first name of the user.
" + } + }, + "Surname": { + "target": "com.amazonaws.directoryservicedata#Surname", + "traits": { + "smithy.api#documentation": "The last name of the user.
" + } + }, + "Enabled": { + "target": "smithy.api#Boolean", + "traits": { + "smithy.api#documentation": "Indicates whether the user account is active.
" + } + }, + "OtherAttributes": { + "target": "com.amazonaws.directoryservicedata#Attributes", + "traits": { + "smithy.api#documentation": "An expression that includes one or more attributes, data types, and values of a\n user.
" + } + } + }, + "traits": { + "smithy.api#documentation": "A user object that contains identifying information and attributes for a specified user.\n
" + } + }, + "com.amazonaws.directoryservicedata#UserList": { + "type": "list", + "member": { + "target": "com.amazonaws.directoryservicedata#User" + } + }, + "com.amazonaws.directoryservicedata#UserName": { + "type": "string", + "traits": { + "smithy.api#length": { + "min": 1, + "max": 20 + }, + "smithy.api#pattern": "^[\\w\\-.]+$" + } + }, + "com.amazonaws.directoryservicedata#UserPrincipalName": { + "type": "string", + "traits": { + "smithy.api#length": { + "min": 1, + "max": 256 + }, + "smithy.api#sensitive": {} + } + }, + "com.amazonaws.directoryservicedata#UserSummary": { + "type": "structure", + "members": { + "SID": { + "target": "com.amazonaws.directoryservicedata#SID", + "traits": { + "smithy.api#documentation": "The unique security identifier (SID) of the user.
", + "smithy.api#required": {} + } + }, + "SAMAccountName": { + "target": "com.amazonaws.directoryservicedata#UserName", + "traits": { + "smithy.api#documentation": "The name of the user.
", + "smithy.api#required": {} + } + }, + "GivenName": { + "target": "com.amazonaws.directoryservicedata#GivenName", + "traits": { + "smithy.api#documentation": "The first name of the user.
" + } + }, + "Surname": { + "target": "com.amazonaws.directoryservicedata#Surname", + "traits": { + "smithy.api#documentation": "The last name of the user.
" + } + }, + "Enabled": { + "target": "smithy.api#Boolean", + "traits": { + "smithy.api#documentation": "Indicates whether the user account is active.
", + "smithy.api#required": {} + } + } + }, + "traits": { + "smithy.api#documentation": "A structure containing a subset of the fields of a user object from a directory.
" + } + }, + "com.amazonaws.directoryservicedata#UserSummaryList": { + "type": "list", + "member": { + "target": "com.amazonaws.directoryservicedata#UserSummary" + } + }, + "com.amazonaws.directoryservicedata#ValidationException": { + "type": "structure", + "members": { + "Message": { + "target": "com.amazonaws.directoryservicedata#ExceptionMessage" + }, + "Reason": { + "target": "com.amazonaws.directoryservicedata#ValidationExceptionReason", + "traits": { + "smithy.api#documentation": "Reason the request failed validation.
" + } + } + }, + "traits": { + "smithy.api#documentation": "The request isn't valid. Review the details in the error message to update the invalid\n parameters or values in your request.
", + "smithy.api#error": "client", + "smithy.api#httpError": 400 + } + }, + "com.amazonaws.directoryservicedata#ValidationExceptionReason": { + "type": "enum", + "members": { + "INVALID_REALM": { + "target": "smithy.api#Unit", + "traits": { + "smithy.api#enumValue": "INVALID_REALM" + } + }, + "INVALID_DIRECTORY_TYPE": { + "target": "smithy.api#Unit", + "traits": { + "smithy.api#enumValue": "INVALID_DIRECTORY_TYPE" + } + }, + "INVALID_SECONDARY_REGION": { + "target": "smithy.api#Unit", + "traits": { + "smithy.api#enumValue": "INVALID_SECONDARY_REGION" + } + }, + "INVALID_NEXT_TOKEN": { + "target": "smithy.api#Unit", + "traits": { + "smithy.api#enumValue": "INVALID_NEXT_TOKEN" + } + }, + "INVALID_ATTRIBUTE_VALUE": { + "target": "smithy.api#Unit", + "traits": { + "smithy.api#enumValue": "INVALID_ATTRIBUTE_VALUE" + } + }, + "INVALID_ATTRIBUTE_NAME": { + "target": "smithy.api#Unit", + "traits": { + "smithy.api#enumValue": "INVALID_ATTRIBUTE_NAME" + } + }, + "INVALID_ATTRIBUTE_FOR_USER": { + "target": "smithy.api#Unit", + "traits": { + "smithy.api#enumValue": "INVALID_ATTRIBUTE_FOR_USER" + } + }, + "INVALID_ATTRIBUTE_FOR_GROUP": { + "target": "smithy.api#Unit", + "traits": { + "smithy.api#enumValue": "INVALID_ATTRIBUTE_FOR_GROUP" + } + }, + "INVALID_ATTRIBUTE_FOR_SEARCH": { + "target": "smithy.api#Unit", + "traits": { + "smithy.api#enumValue": "INVALID_ATTRIBUTE_FOR_SEARCH" + } + }, + "INVALID_ATTRIBUTE_FOR_MODIFY": { + "target": "smithy.api#Unit", + "traits": { + "smithy.api#enumValue": "INVALID_ATTRIBUTE_FOR_MODIFY" + } + }, + "DUPLICATE_ATTRIBUTE": { + "target": "smithy.api#Unit", + "traits": { + "smithy.api#enumValue": "DUPLICATE_ATTRIBUTE" + } + }, + "MISSING_ATTRIBUTE": { + "target": "smithy.api#Unit", + "traits": { + "smithy.api#enumValue": "MISSING_ATTRIBUTE" + } + }, + "ATTRIBUTE_EXISTS": { + "target": "smithy.api#Unit", + "traits": { + "smithy.api#enumValue": "ATTRIBUTE_EXISTS" + } + }, + "LDAP_SIZE_LIMIT_EXCEEDED": { + "target": "smithy.api#Unit", + "traits": { + "smithy.api#enumValue": "LDAP_SIZE_LIMIT_EXCEEDED" + } + }, + "LDAP_UNSUPPORTED_OPERATION": { + "target": "smithy.api#Unit", + "traits": { + "smithy.api#enumValue": "LDAP_UNSUPPORTED_OPERATION" + } + } + } + } + } +} \ No newline at end of file diff --git a/aws-models/directory-service.json b/aws-models/directory-service.json index f7c9f940be47..abef873a0e03 100644 --- a/aws-models/directory-service.json +++ b/aws-models/directory-service.json @@ -98,7 +98,7 @@ } }, "traits": { - "smithy.api#documentation": "Client authentication is not available in this region at this time.
", + "smithy.api#documentation": "You do not have sufficient access to perform this action.
", "smithy.api#error": "client" } }, @@ -167,7 +167,7 @@ "target": "com.amazonaws.directoryservice#UpdateSecurityGroupForDirectoryControllers", "traits": { "smithy.api#default": false, - "smithy.api#documentation": "If set to true, updates the inbound and outbound rules of the security group that has\n the description: \"Amazon Web Services created security group for directory ID\n directory controllers.\" Following are the new rules:
\nInbound:
\nType: Custom UDP Rule, Protocol: UDP, Range: 88, Source: 0.0.0.0/0
\nType: Custom UDP Rule, Protocol: UDP, Range: 123, Source: 0.0.0.0/0
\nType: Custom UDP Rule, Protocol: UDP, Range: 138, Source: 0.0.0.0/0
\nType: Custom UDP Rule, Protocol: UDP, Range: 389, Source: 0.0.0.0/0
\nType: Custom UDP Rule, Protocol: UDP, Range: 464, Source: 0.0.0.0/0
\nType: Custom UDP Rule, Protocol: UDP, Range: 445, Source: 0.0.0.0/0
\nType: Custom TCP Rule, Protocol: TCP, Range: 88, Source: 0.0.0.0/0
\nType: Custom TCP Rule, Protocol: TCP, Range: 135, Source: 0.0.0.0/0
\nType: Custom TCP Rule, Protocol: TCP, Range: 445, Source: 0.0.0.0/0
\nType: Custom TCP Rule, Protocol: TCP, Range: 464, Source: 0.0.0.0/0
\nType: Custom TCP Rule, Protocol: TCP, Range: 636, Source: 0.0.0.0/0
\nType: Custom TCP Rule, Protocol: TCP, Range: 1024-65535, Source:\n 0.0.0.0/0
\nType: Custom TCP Rule, Protocol: TCP, Range: 3268-33269, Source:\n 0.0.0.0/0
\nType: DNS (UDP), Protocol: UDP, Range: 53, Source: 0.0.0.0/0
\nType: DNS (TCP), Protocol: TCP, Range: 53, Source: 0.0.0.0/0
\nType: LDAP, Protocol: TCP, Range: 389, Source: 0.0.0.0/0
\nType: All ICMP, Protocol: All, Range: N/A, Source: 0.0.0.0/0
\nOutbound:
\nType: All traffic, Protocol: All, Range: All, Destination: 0.0.0.0/0
\nThese security rules impact an internal network interface that is not exposed\n publicly.
" + "smithy.api#documentation": "If set to true, updates the inbound and outbound rules of the security group that has\n the description: \"Amazon Web Services created security group for directory ID\n directory controllers.\" Following are the new rules:
\nInbound:
\nType: Custom UDP Rule, Protocol: UDP, Range: 88, Source: Managed Microsoft AD VPC IPv4 CIDR
\nType: Custom UDP Rule, Protocol: UDP, Range: 123, Source: Managed Microsoft AD VPC IPv4 CIDR
\nType: Custom UDP Rule, Protocol: UDP, Range: 138, Source: Managed Microsoft AD VPC IPv4 CIDR
\nType: Custom UDP Rule, Protocol: UDP, Range: 389, Source: Managed Microsoft AD VPC IPv4 CIDR
\nType: Custom UDP Rule, Protocol: UDP, Range: 464, Source: Managed Microsoft AD VPC IPv4 CIDR
\nType: Custom UDP Rule, Protocol: UDP, Range: 445, Source: Managed Microsoft AD VPC IPv4 CIDR
\nType: Custom TCP Rule, Protocol: TCP, Range: 88, Source: Managed Microsoft AD VPC IPv4 CIDR
\nType: Custom TCP Rule, Protocol: TCP, Range: 135, Source: Managed Microsoft AD VPC IPv4 CIDR
\nType: Custom TCP Rule, Protocol: TCP, Range: 445, Source: Managed Microsoft AD VPC IPv4 CIDR
\nType: Custom TCP Rule, Protocol: TCP, Range: 464, Source: Managed Microsoft AD VPC IPv4 CIDR
\nType: Custom TCP Rule, Protocol: TCP, Range: 636, Source: Managed Microsoft AD VPC IPv4 CIDR
\nType: Custom TCP Rule, Protocol: TCP, Range: 1024-65535, Source:\n Managed Microsoft AD VPC IPv4 CIDR
\nType: Custom TCP Rule, Protocol: TCP, Range: 3268-33269, Source:\n Managed Microsoft AD VPC IPv4 CIDR
\nType: DNS (UDP), Protocol: UDP, Range: 53, Source: Managed Microsoft AD VPC IPv4 CIDR
\nType: DNS (TCP), Protocol: TCP, Range: 53, Source: Managed Microsoft AD VPC IPv4 CIDR
\nType: LDAP, Protocol: TCP, Range: 389, Source: Managed Microsoft AD VPC IPv4 CIDR
\nType: All ICMP, Protocol: All, Range: N/A, Source: Managed Microsoft AD VPC IPv4 CIDR
\nOutbound:
\nType: All traffic, Protocol: All, Range: All, Destination: 0.0.0.0/0
\nThese security rules impact an internal network interface that is not exposed\n publicly.
" } } }, @@ -1630,7 +1630,7 @@ "TrustPassword": { "target": "com.amazonaws.directoryservice#TrustPassword", "traits": { - "smithy.api#documentation": "The trust password. The must be the same password that was used when creating the trust\n relationship on the external domain.
", + "smithy.api#documentation": "The trust password. The trust password must be the same password that was used when creating the trust\n relationship on the external domain.
", "smithy.api#required": {} } }, @@ -1699,6 +1699,41 @@ "smithy.api#pattern": "^(?!.*\\\\|.*\"|.*\\/|.*\\[|.*\\]|.*:|.*;|.*\\||.*=|.*,|.*\\+|.*\\*|.*\\?|.*<|.*>|.*@).*$" } }, + "com.amazonaws.directoryservice#DataAccessStatus": { + "type": "enum", + "members": { + "DISABLED": { + "target": "smithy.api#Unit", + "traits": { + "smithy.api#enumValue": "Disabled" + } + }, + "DISABLING": { + "target": "smithy.api#Unit", + "traits": { + "smithy.api#enumValue": "Disabling" + } + }, + "ENABLED": { + "target": "smithy.api#Unit", + "traits": { + "smithy.api#enumValue": "Enabled" + } + }, + "ENABLING": { + "target": "smithy.api#Unit", + "traits": { + "smithy.api#enumValue": "Enabling" + } + }, + "FAILED": { + "target": "smithy.api#Unit", + "traits": { + "smithy.api#enumValue": "Failed" + } + } + } + }, "com.amazonaws.directoryservice#DeleteAssociatedConditionalForwarder": { "type": "boolean", "traits": { @@ -2442,6 +2477,64 @@ "smithy.api#output": {} } }, + "com.amazonaws.directoryservice#DescribeDirectoryDataAccess": { + "type": "operation", + "input": { + "target": "com.amazonaws.directoryservice#DescribeDirectoryDataAccessRequest" + }, + "output": { + "target": "com.amazonaws.directoryservice#DescribeDirectoryDataAccessResult" + }, + "errors": [ + { + "target": "com.amazonaws.directoryservice#AccessDeniedException" + }, + { + "target": "com.amazonaws.directoryservice#ClientException" + }, + { + "target": "com.amazonaws.directoryservice#DirectoryDoesNotExistException" + }, + { + "target": "com.amazonaws.directoryservice#ServiceException" + }, + { + "target": "com.amazonaws.directoryservice#UnsupportedOperationException" + } + ], + "traits": { + "smithy.api#documentation": "Obtains status of directory data access enablement through the Directory Service Data API for the specified directory.
" + } + }, + "com.amazonaws.directoryservice#DescribeDirectoryDataAccessRequest": { + "type": "structure", + "members": { + "DirectoryId": { + "target": "com.amazonaws.directoryservice#DirectoryId", + "traits": { + "smithy.api#documentation": "The directory identifier.
", + "smithy.api#required": {} + } + } + }, + "traits": { + "smithy.api#input": {} + } + }, + "com.amazonaws.directoryservice#DescribeDirectoryDataAccessResult": { + "type": "structure", + "members": { + "DataAccessStatus": { + "target": "com.amazonaws.directoryservice#DataAccessStatus", + "traits": { + "smithy.api#documentation": "The current status of data access through the Directory Service Data API.
" + } + } + }, + "traits": { + "smithy.api#output": {} + } + }, "com.amazonaws.directoryservice#DescribeDomainControllers": { "type": "operation", "input": { @@ -2849,7 +2942,7 @@ "NextToken": { "target": "com.amazonaws.directoryservice#NextToken", "traits": { - "smithy.api#documentation": "If not null, token that indicates that more results are available. Pass this value for the NextToken
parameter in a subsequent call to DescribeSettings
to retrieve the next set of items.
If not null, token that indicates that more results are available. \n Pass this value for the NextToken
parameter in a subsequent \n call to DescribeSettings
to retrieve the next set of items.
The directory size.
" + "smithy.api#documentation": "The directory type.
" } }, "VpcSettings": { @@ -3846,6 +3939,9 @@ { "target": "com.amazonaws.directoryservice#DescribeDirectories" }, + { + "target": "com.amazonaws.directoryservice#DescribeDirectoryDataAccess" + }, { "target": "com.amazonaws.directoryservice#DescribeDomainControllers" }, @@ -3876,6 +3972,9 @@ { "target": "com.amazonaws.directoryservice#DisableClientAuthentication" }, + { + "target": "com.amazonaws.directoryservice#DisableDirectoryDataAccess" + }, { "target": "com.amazonaws.directoryservice#DisableLDAPS" }, @@ -3888,6 +3987,9 @@ { "target": "com.amazonaws.directoryservice#EnableClientAuthentication" }, + { + "target": "com.amazonaws.directoryservice#EnableDirectoryDataAccess" + }, { "target": "com.amazonaws.directoryservice#EnableLDAPS" }, @@ -5063,6 +5165,12 @@ "traits": { "smithy.api#enumValue": "Failed" } + }, + "UPDATING": { + "target": "smithy.api#Unit", + "traits": { + "smithy.api#enumValue": "Updating" + } } } }, @@ -5106,7 +5214,7 @@ } }, "traits": { - "smithy.api#documentation": "The specified directory is unavailable or could not be found.
", + "smithy.api#documentation": "The specified directory is unavailable.
", "smithy.api#error": "client" } }, @@ -5209,7 +5317,7 @@ "Type": { "target": "com.amazonaws.directoryservice#ClientAuthenticationType", "traits": { - "smithy.api#documentation": "The type of client authentication to disable. Currently, only the parameter, SmartCard
is supported.
The type of client authentication to disable. Currently the only parameter \"SmartCard\"
is supported.
Deactivates access to directory data via the Directory Service Data API for the specified directory.
" + } + }, + "com.amazonaws.directoryservice#DisableDirectoryDataAccessRequest": { + "type": "structure", + "members": { + "DirectoryId": { + "target": "com.amazonaws.directoryservice#DirectoryId", + "traits": { + "smithy.api#documentation": "The directory identifier.
", + "smithy.api#required": {} + } + } + }, + "traits": { + "smithy.api#input": {} + } + }, + "com.amazonaws.directoryservice#DisableDirectoryDataAccessResult": { + "type": "structure", + "members": {}, + "traits": { + "smithy.api#output": {} + } + }, "com.amazonaws.directoryservice#DisableLDAPS": { "type": "operation", "input": { @@ -5546,6 +5711,12 @@ "traits": { "smithy.api#enumValue": "Failed" } + }, + "UPDATING": { + "target": "smithy.api#Unit", + "traits": { + "smithy.api#enumValue": "Updating" + } } } }, @@ -5622,6 +5793,63 @@ "smithy.api#output": {} } }, + "com.amazonaws.directoryservice#EnableDirectoryDataAccess": { + "type": "operation", + "input": { + "target": "com.amazonaws.directoryservice#EnableDirectoryDataAccessRequest" + }, + "output": { + "target": "com.amazonaws.directoryservice#EnableDirectoryDataAccessResult" + }, + "errors": [ + { + "target": "com.amazonaws.directoryservice#AccessDeniedException" + }, + { + "target": "com.amazonaws.directoryservice#ClientException" + }, + { + "target": "com.amazonaws.directoryservice#DirectoryDoesNotExistException" + }, + { + "target": "com.amazonaws.directoryservice#DirectoryInDesiredStateException" + }, + { + "target": "com.amazonaws.directoryservice#DirectoryUnavailableException" + }, + { + "target": "com.amazonaws.directoryservice#ServiceException" + }, + { + "target": "com.amazonaws.directoryservice#UnsupportedOperationException" + } + ], + "traits": { + "smithy.api#documentation": "Enables access to directory data via the Directory Service Data API for the specified directory.
" + } + }, + "com.amazonaws.directoryservice#EnableDirectoryDataAccessRequest": { + "type": "structure", + "members": { + "DirectoryId": { + "target": "com.amazonaws.directoryservice#DirectoryId", + "traits": { + "smithy.api#documentation": "The directory identifier.
", + "smithy.api#required": {} + } + } + }, + "traits": { + "smithy.api#input": {} + } + }, + "com.amazonaws.directoryservice#EnableDirectoryDataAccessResult": { + "type": "structure", + "members": {}, + "traits": { + "smithy.api#output": {} + } + }, "com.amazonaws.directoryservice#EnableLDAPS": { "type": "operation", "input": { @@ -7069,7 +7297,7 @@ "target": "com.amazonaws.directoryservice#RadiusRetries", "traits": { "smithy.api#default": 0, - "smithy.api#documentation": "The maximum number of times that communication with the RADIUS server is\n attempted.
" + "smithy.api#documentation": "The maximum number of times that communication with the RADIUS server is retried after the initial attempt.
" } }, "SharedSecret": { @@ -7141,7 +7369,7 @@ "traits": { "smithy.api#range": { "min": 1, - "max": 20 + "max": 50 } } }, @@ -7708,7 +7936,7 @@ } ], "traits": { - "smithy.api#documentation": "Resets the password for any user in your Managed Microsoft AD or Simple AD\n directory.
\nYou can reset the password for any user in your directory with the following\n exceptions:
\nFor Simple AD, you cannot reset the password for any user that is a member of either\n the Domain Admins or Enterprise\n Admins group except for the administrator user.
\nFor Managed Microsoft AD, you can only reset the password for a user that is in an\n OU based off of the NetBIOS name that you typed when you created your directory. For\n example, you cannot reset the password for a user in the Amazon Web Services\n Reserved OU. For more information about the OU structure for an Managed Microsoft AD directory, see What Gets Created in the Directory Service Administration\n Guide.
\nResets the password for any user in your Managed Microsoft AD or Simple AD\n directory. Disabled users will become enabled and can be authenticated following the API call.
\nYou can reset the password for any user in your directory with the following\n exceptions:
\nFor Simple AD, you cannot reset the password for any user that is a member of either\n the Domain Admins or Enterprise\n Admins group except for the administrator user.
\nFor Managed Microsoft AD, you can only reset the password for a user that is in an\n OU based off of the NetBIOS name that you typed when you created your directory. For\n example, you cannot reset the password for a user in the Amazon Web Services\n Reserved OU. For more information about the OU structure for an Managed Microsoft AD directory, see What Gets Created in the Directory Service Administration\n Guide.
\nRequired name of the tag. The string value can be Unicode characters and cannot be\n prefixed with \"aws:\". The string can contain only the set of Unicode letters, digits,\n white-space, '_', '.', '/', '=', '+', '-' (Java regex:\n \"^([\\\\p{L}\\\\p{Z}\\\\p{N}_.:/=+\\\\-]*)$\").
", + "smithy.api#documentation": "Required name of the tag. The string value can be Unicode characters and cannot be\n prefixed with \"aws:\". The string can contain only the set of Unicode letters, digits,\n white-space, '_', '.', '/', '=', '+', '-', ':', '@'(Java regex:\n \"^([\\\\p{L}\\\\p{Z}\\\\p{N}_.:/=+\\\\-]*)$\").
", "smithy.api#required": {} } }, "Value": { "target": "com.amazonaws.directoryservice#TagValue", "traits": { - "smithy.api#documentation": "The optional value of the tag. The string value can be Unicode characters. The string\n can contain only the set of Unicode letters, digits, white-space, '_', '.', '/', '=', '+', '-'\n (Java regex: \"^([\\\\p{L}\\\\p{Z}\\\\p{N}_.:/=+\\\\-]*)$\").
", + "smithy.api#documentation": "The optional value of the tag. The string value can be Unicode characters. The string\n can contain only the set of Unicode letters, digits, white-space, '_', '.', '/', '=', '+', '-', ':', '@'\n (Java regex: \"^([\\\\p{L}\\\\p{Z}\\\\p{N}_.:/=+\\\\-]*)$\").
", "smithy.api#required": {} } } diff --git a/aws-models/guardduty.json b/aws-models/guardduty.json index 8343b777698a..3e052ca8466c 100644 --- a/aws-models/guardduty.json +++ b/aws-models/guardduty.json @@ -4674,6 +4674,13 @@ "smithy.api#documentation": "The name of the task group that's associated with the task.
", "smithy.api#jsonName": "group" } + }, + "LaunchType": { + "target": "com.amazonaws.guardduty#String", + "traits": { + "smithy.api#documentation": "A capacity on which the task is running. For example, Fargate
and EC2
.
The IP of the Kubernetes API caller and the IPs of any proxies or load balancers between\n the caller and the API endpoint.
", - "smithy.api#jsonName": "sourceIps" + "smithy.api#jsonName": "sourceIPs" } }, "UserAgent": { diff --git a/aws-models/mailmanager.json b/aws-models/mailmanager.json index 39fb5641d2c9..49bd9a0ebebe 100644 --- a/aws-models/mailmanager.json +++ b/aws-models/mailmanager.json @@ -4415,7 +4415,7 @@ }, "aws.protocols#awsJson1_0": {}, "smithy.api#cors": {}, - "smithy.api#documentation": "\n AWS SES Mail Manager API contains operations and data types\n that comprise the Mail Manager feature of Amazon Simple Email Service.
\nMail Manager is a set of Amazon SES email gateway features designed to help you strengthen\n your organization's email infrastructure, simplify email workflow management, and\n streamline email compliance control. To learn more, see the Mail Manager chapter in the Amazon SES Developer\n Guide.
", + "smithy.api#documentation": "The Amazon SES Mail Manager API contains operations and data types\n that comprise the Mail Manager feature of Amazon Simple Email Service (SES).
\nMail Manager is a set of Amazon SES email gateway features designed to help you strengthen\n your organization's email infrastructure, simplify email workflow management, and\n streamline email compliance control. To learn more, see the Mail Manager chapter in the Amazon SES Developer\n Guide.
", "smithy.api#externalDocumentation": { "API Reference": "https://w.amazon.com/bin/view/AWS/Border" }, @@ -5120,6 +5120,12 @@ "smithy.api#documentation": "The textual body content of an email message.
" } }, + "com.amazonaws.mailmanager#MimeHeaderAttribute": { + "type": "string", + "traits": { + "smithy.api#pattern": "^X-[a-zA-Z0-9-]{1,256}$" + } + }, "com.amazonaws.mailmanager#NameOrArn": { "type": "string", "traits": { @@ -6400,6 +6406,12 @@ "traits": { "smithy.api#documentation": "The email attribute to evaluate in a string condition expression.
" } + }, + "MimeHeaderAttribute": { + "target": "com.amazonaws.mailmanager#MimeHeaderAttribute", + "traits": { + "smithy.api#documentation": "The email MIME X-Header attribute to evaluate in a string condition expression.
" + } } }, "traits": { @@ -7083,10 +7095,19 @@ "target": "smithy.api#String" } }, + "com.amazonaws.mailmanager#StringValue": { + "type": "string", + "traits": { + "smithy.api#length": { + "min": 1, + "max": 2048 + } + } + }, "com.amazonaws.mailmanager#StringValueList": { "type": "list", "member": { - "target": "smithy.api#String" + "target": "com.amazonaws.mailmanager#StringValue" }, "traits": { "smithy.api#length": { @@ -7658,7 +7679,7 @@ } ], "traits": { - "smithy.api#documentation": ">Update attributes of an already provisioned rule set.
", + "smithy.api#documentation": "Update attributes of an already provisioned rule set.
", "smithy.api#idempotent": {} } }, diff --git a/aws-models/rds.json b/aws-models/rds.json index 347818e57919..ebc1c71fb854 100644 --- a/aws-models/rds.json +++ b/aws-models/rds.json @@ -27805,7 +27805,7 @@ } ], "traits": { - "smithy.api#documentation": "Creates a new DB instance from a DB snapshot. The target database is created from the source database restore point with most\n of the source's original configuration, including the default security group and DB parameter group. By default, the new DB\n instance is created as a Single-AZ deployment, except when the instance is a SQL Server instance that has an option group\n associated with mirroring. In this case, the instance becomes a Multi-AZ deployment, not a Single-AZ deployment.
\nIf you want to replace your original DB instance with the new, restored DB instance, then rename your original DB instance\n before you call the RestoreDBInstanceFromDBSnapshot
operation. RDS doesn't allow two DB instances with the same name. After you\n have renamed your original DB instance with a different identifier, then you can pass the original name of the DB instance as\n the DBInstanceIdentifier
in the call to the RestoreDBInstanceFromDBSnapshot
operation. The result is that you replace the original\n DB instance with the DB instance created from the snapshot.
If you are restoring from a shared manual DB snapshot, the DBSnapshotIdentifier
\n must be the ARN of the shared DB snapshot.
This command doesn't apply to Aurora MySQL and Aurora PostgreSQL. For Aurora, use RestoreDBClusterFromSnapshot
.
Creates a new DB instance from a DB snapshot. The target database is created from the source database restore point with most\n of the source's original configuration, including the default security group and DB parameter group. By default, the new DB\n instance is created as a Single-AZ deployment, except when the instance is a SQL Server instance that has an option group\n associated with mirroring. In this case, the instance becomes a Multi-AZ deployment, not a Single-AZ deployment.
\nIf you want to replace your original DB instance with the new, restored DB instance, then rename your original DB instance\n before you call the RestoreDBInstanceFromDBSnapshot
operation. RDS doesn't allow two DB instances with the same name. After you\n have renamed your original DB instance with a different identifier, then you can pass the original name of the DB instance as\n the DBInstanceIdentifier
in the call to the RestoreDBInstanceFromDBSnapshot
operation. The result is that you replace the original\n DB instance with the DB instance created from the snapshot.
If you are restoring from a shared manual DB snapshot, the DBSnapshotIdentifier
\n must be the ARN of the shared DB snapshot.
To restore from a DB snapshot with an unsupported engine version, you must first upgrade the \n engine version of the snapshot. For more information about upgrading a RDS for MySQL DB snapshot engine version, see Upgrading a MySQL DB snapshot engine version. \n For more information about upgrading a RDS for PostgreSQL DB snapshot engine version, Upgrading a PostgreSQL DB snapshot engine version.
\nThis command doesn't apply to Aurora MySQL and Aurora PostgreSQL. For Aurora, use RestoreDBClusterFromSnapshot
.
Completes a multipart upload by assembling previously uploaded parts.
\nYou first initiate the multipart upload and then upload all parts using the UploadPart\n operation or the UploadPartCopy\n operation. After successfully uploading all relevant parts of an upload, you call this\n CompleteMultipartUpload
operation to complete the upload. Upon receiving this request, Amazon S3 concatenates all the parts\n in ascending order by part number to create a new object. In the CompleteMultipartUpload \n request, you must provide the parts list and ensure that the parts list is complete.\n The CompleteMultipartUpload API operation concatenates the parts that you provide in the list. For each part in the list,\n you must provide the PartNumber
value and the ETag
value that are returned after that part\n was uploaded.
The processing of a CompleteMultipartUpload request could take several minutes to\n finalize. After Amazon S3 begins processing the request, it sends an HTTP response header that\n specifies a 200 OK
response. While processing is in progress, Amazon S3 periodically sends white\n space characters to keep the connection from timing out. A request could fail after the\n initial 200 OK
response has been sent. This means that a 200 OK
response can\n contain either a success or an error. The error response might be embedded in the 200 OK
response. \n If you call this API operation directly, make sure to design\n your application to parse the contents of the response and handle it appropriately. If you\n use Amazon Web Services SDKs, SDKs handle this condition. The SDKs detect the embedded error and apply\n error handling per your configuration settings (including automatically retrying the\n request as appropriate). If the condition persists, the SDKs throw an exception (or, for\n the SDKs that don't use exceptions, they return an error).
Note that if CompleteMultipartUpload
fails, applications should be prepared\n to retry any failed requests (including 500 error responses). For more information, see Amazon S3 Error Best\n Practices.
You can't use Content-Type: application/x-www-form-urlencoded
for the \n CompleteMultipartUpload requests. Also, if you don't provide a\n Content-Type
header, CompleteMultipartUpload
can still return a 200\n OK
response.
For more information about multipart uploads, see Uploading Objects Using Multipart\n Upload in the Amazon S3\n User Guide.
\n\n Directory buckets - For directory buckets, you must make requests for this API operation to the Zonal endpoint. These endpoints support virtual-hosted-style requests in the format https://bucket_name.s3express-az_id.region.amazonaws.com/key-name\n
. Path-style requests are not supported. For more information, see Regional and Zonal endpoints in the\n Amazon S3 User Guide.
\n General purpose bucket permissions - For\n information about permissions required to use the multipart upload API, see\n Multipart Upload and\n Permissions in the Amazon S3 User Guide.
\n\n Directory bucket permissions - To grant access to this API operation on a directory bucket, we recommend that you use the \n CreateSession
\n API operation for session-based authorization. Specifically, you grant the s3express:CreateSession
permission to the directory bucket in a bucket policy or an IAM identity-based policy. Then, you make the CreateSession
API call on the bucket to obtain a session token. With the session token in your request header, you can make API requests to this operation. After the session token expires, you make another CreateSession
API call to generate a new session token for use. \nAmazon Web Services CLI or SDKs create session and refresh the session token automatically to avoid service interruptions when a session expires. For more information about authorization, see \n CreateSession
\n .
If you provide an additional checksum\n value in your MultipartUpload
requests and the\n object is encrypted with Key Management Service, you must have permission to use the\n kms:Decrypt
action for the\n CompleteMultipartUpload
request to succeed.
Error Code: EntityTooSmall
\n
Description: Your proposed upload is smaller than the minimum allowed object\n size. Each part must be at least 5 MB in size, except the last part.
\nHTTP Status Code: 400 Bad Request
\nError Code: InvalidPart
\n
Description: One or more of the specified parts could not be found. The part\n might not have been uploaded, or the specified ETag might not have\n matched the uploaded part's ETag.
\nHTTP Status Code: 400 Bad Request
\nError Code: InvalidPartOrder
\n
Description: The list of parts was not in ascending order. The parts list\n must be specified in order by part number.
\nHTTP Status Code: 400 Bad Request
\nError Code: NoSuchUpload
\n
Description: The specified multipart upload does not exist. The upload ID\n might be invalid, or the multipart upload might have been aborted or\n completed.
\nHTTP Status Code: 404 Not Found
\n\n Directory buckets - The HTTP Host header syntax is \n Bucket_name.s3express-az_id.region.amazonaws.com
.
The following operations are related to CompleteMultipartUpload
:
\n UploadPart\n
\n\n AbortMultipartUpload\n
\n\n ListParts\n
\n\n ListMultipartUploads\n
\nCompletes a multipart upload by assembling previously uploaded parts.
\nYou first initiate the multipart upload and then upload all parts using the UploadPart\n operation or the UploadPartCopy\n operation. After successfully uploading all relevant parts of an upload, you call this\n CompleteMultipartUpload
operation to complete the upload. Upon receiving this request, Amazon S3 concatenates all the parts\n in ascending order by part number to create a new object. In the CompleteMultipartUpload \n request, you must provide the parts list and ensure that the parts list is complete.\n The CompleteMultipartUpload API operation concatenates the parts that you provide in the list. For each part in the list,\n you must provide the PartNumber
value and the ETag
value that are returned after that part\n was uploaded.
The processing of a CompleteMultipartUpload request could take several minutes to\n finalize. After Amazon S3 begins processing the request, it sends an HTTP response header that\n specifies a 200 OK
response. While processing is in progress, Amazon S3 periodically sends white\n space characters to keep the connection from timing out. A request could fail after the\n initial 200 OK
response has been sent. This means that a 200 OK
response can\n contain either a success or an error. The error response might be embedded in the 200 OK
response. \n If you call this API operation directly, make sure to design\n your application to parse the contents of the response and handle it appropriately. If you\n use Amazon Web Services SDKs, SDKs handle this condition. The SDKs detect the embedded error and apply\n error handling per your configuration settings (including automatically retrying the\n request as appropriate). If the condition persists, the SDKs throw an exception (or, for\n the SDKs that don't use exceptions, they return an error).
Note that if CompleteMultipartUpload
fails, applications should be prepared\n to retry any failed requests (including 500 error responses). For more information, see Amazon S3 Error Best\n Practices.
You can't use Content-Type: application/x-www-form-urlencoded
for the \n CompleteMultipartUpload requests. Also, if you don't provide a\n Content-Type
header, CompleteMultipartUpload
can still return a 200\n OK
response.
For more information about multipart uploads, see Uploading Objects Using Multipart\n Upload in the Amazon S3\n User Guide.
\n\n Directory buckets - For directory buckets, you must make requests for this API operation to the Zonal endpoint. These endpoints support virtual-hosted-style requests in the format https://bucket_name.s3express-az_id.region.amazonaws.com/key-name\n
. Path-style requests are not supported. For more information, see Regional and Zonal endpoints in the\n Amazon S3 User Guide.
\n General purpose bucket permissions - For\n information about permissions required to use the multipart upload API, see\n Multipart Upload and\n Permissions in the Amazon S3 User Guide.
\nIf you provide an additional checksum\n value in your MultipartUpload
requests and the\n object is encrypted with Key Management Service, you must have permission to use the\n kms:Decrypt
action for the\n CompleteMultipartUpload
request to succeed.
\n Directory bucket permissions - To grant access to this API operation on a directory bucket, we recommend that you use the \n CreateSession
\n API operation for session-based authorization. Specifically, you grant the s3express:CreateSession
permission to the directory bucket in a bucket policy or an IAM identity-based policy. Then, you make the CreateSession
API call on the bucket to obtain a session token. With the session token in your request header, you can make API requests to this operation. After the session token expires, you make another CreateSession
API call to generate a new session token for use. \nAmazon Web Services CLI or SDKs create session and refresh the session token automatically to avoid service interruptions when a session expires. For more information about authorization, see \n CreateSession
\n .
If the object is encrypted with\n SSE-KMS, you must also have the\n kms:GenerateDataKey
and kms:Decrypt
permissions in IAM identity-based policies and KMS key policies for the KMS key.
Error Code: EntityTooSmall
\n
Description: Your proposed upload is smaller than the minimum allowed object\n size. Each part must be at least 5 MB in size, except the last part.
\nHTTP Status Code: 400 Bad Request
\nError Code: InvalidPart
\n
Description: One or more of the specified parts could not be found. The part\n might not have been uploaded, or the specified ETag might not have\n matched the uploaded part's ETag.
\nHTTP Status Code: 400 Bad Request
\nError Code: InvalidPartOrder
\n
Description: The list of parts was not in ascending order. The parts list\n must be specified in order by part number.
\nHTTP Status Code: 400 Bad Request
\nError Code: NoSuchUpload
\n
Description: The specified multipart upload does not exist. The upload ID\n might be invalid, or the multipart upload might have been aborted or\n completed.
\nHTTP Status Code: 404 Not Found
\n\n Directory buckets - The HTTP Host header syntax is \n Bucket_name.s3express-az_id.region.amazonaws.com
.
The following operations are related to CompleteMultipartUpload
:
\n UploadPart\n
\n\n AbortMultipartUpload\n
\n\n ListParts\n
\n\n ListMultipartUploads\n
\nThe server-side encryption algorithm used when storing this object in Amazon S3 (for example,\n AES256
, aws:kms
).
For directory buckets, only server-side encryption with Amazon S3 managed keys (SSE-S3) (AES256
) is supported.
The server-side encryption algorithm used when storing this object in Amazon S3 (for example,\n AES256
, aws:kms
).
If present, indicates the ID of the Key Management Service (KMS) symmetric encryption customer managed key\n that was used for the object.
\nThis functionality is not supported for directory buckets.
\nIf present, indicates the ID of the KMS key that was used for object encryption.
", "smithy.api#httpHeader": "x-amz-server-side-encryption-aws-kms-key-id" } }, "BucketKeyEnabled": { "target": "com.amazonaws.s3#BucketKeyEnabled", "traits": { - "smithy.api#documentation": "Indicates whether the multipart upload uses an S3 Bucket Key for server-side encryption\n with Key Management Service (KMS) keys (SSE-KMS).
\nThis functionality is not supported for directory buckets.
\nIndicates whether the multipart upload uses an S3 Bucket Key for server-side encryption\n with Key Management Service (KMS) keys (SSE-KMS).
", "smithy.api#httpHeader": "x-amz-server-side-encryption-bucket-key-enabled" } }, @@ -18701,7 +18701,7 @@ } ], "traits": { - "smithy.api#documentation": "Creates a copy of an object that is already stored in Amazon S3.
\nYou can store individual objects of up to 5 TB in Amazon S3. You create a copy of your\n object up to 5 GB in size in a single atomic action using this API. However, to copy an\n object greater than 5 GB, you must use the multipart upload Upload Part - Copy\n (UploadPartCopy) API. For more information, see Copy Object Using the\n REST Multipart Upload API.
\nYou can copy individual objects between general purpose buckets, between directory buckets, and \n between general purpose buckets and directory buckets.
\nAmazon S3 supports copy operations using Multi-Region Access Points only as a destination when using the Multi-Region Access Point ARN.
\n\n Directory buckets - For directory buckets, you must make requests for this API operation to the Zonal endpoint. These endpoints support virtual-hosted-style requests in the format https://bucket_name.s3express-az_id.region.amazonaws.com/key-name\n
. Path-style requests are not supported. For more information, see Regional and Zonal endpoints in the\n Amazon S3 User Guide.
VPC endpoints don't support cross-Region requests (including copies). If you're using VPC endpoints, your source and destination buckets should be in the same Amazon Web Services Region as your VPC endpoint.
\nBoth the\n Region that you want to copy the object from and the Region that you want to copy the\n object to must be enabled for your account. For more information about how to enable a Region for your account, see Enable \n or disable a Region for standalone accounts in the\n Amazon Web Services Account Management Guide.
\nAmazon S3 transfer acceleration does not support cross-Region copies. If you request a\n cross-Region copy using a transfer acceleration endpoint, you get a 400 Bad\n Request
error. For more information, see Transfer\n Acceleration.
All CopyObject
requests must be authenticated and signed by using IAM credentials (access key ID and secret access key for the IAM identities). All headers with the x-amz-
prefix, including\n x-amz-copy-source
, must be signed. For more information, see REST Authentication.
\n Directory buckets - You must use the IAM credentials to authenticate and authorize your access to the CopyObject
API operation, instead of using the \n temporary security credentials through the CreateSession
API operation.
Amazon Web Services CLI or SDKs handles authentication and authorization on your behalf.
\nYou must have\n read access to the source object and write\n access to the destination bucket.
\n\n General purpose bucket permissions -\n You must have permissions in an IAM policy based on the source and destination\n bucket types in a CopyObject
operation.
If the source object is in a general purpose bucket, you must have\n \n s3:GetObject
\n \n permission to read the source object that is being copied.
If the destination bucket is a general purpose bucket, you must have\n \n s3:PutObject
\n \n permission to write the object copy to the destination bucket.
\n Directory bucket permissions -\n You must have permissions in a bucket policy or an IAM identity-based policy based on the source and destination\n bucket types in a CopyObject
operation.
If the source object that you want to copy is in a\n directory bucket, you must have the \n s3express:CreateSession
\n permission in\n the Action
element of a policy to read the object. By default, the session is in the ReadWrite
mode. If you want to restrict the access, you can explicitly set the s3express:SessionMode
condition key to ReadOnly
on the copy source bucket.
If the copy destination is a directory bucket, you must have the \n s3express:CreateSession
\n permission in the\n Action
element of a policy to write the object\n to the destination. The s3express:SessionMode
condition\n key can't be set to ReadOnly
on the copy destination bucket.
For example policies, see Example bucket policies for S3 Express One Zone and Amazon Web Services Identity and Access Management (IAM) identity-based policies for S3 Express One Zone in the\n Amazon S3 User Guide.
\nWhen the request is an HTTP 1.1 request, the response is chunk encoded. When\n the request is not an HTTP 1.1 request, the response would not contain the\n Content-Length
. You always need to read the entire response body\n to check if the copy succeeds.
If the copy is successful, you receive a response with information about the copied\n object.
\nA copy request might return an error when Amazon S3 receives the copy request or while Amazon S3\n is copying the files. A 200 OK
response can contain either a success or an error.
If the error occurs before the copy action starts, you receive a\n standard Amazon S3 error.
\nIf the error occurs during the copy operation, the error response is\n embedded in the 200 OK
response. For example, in a cross-region copy, you \n may encounter throttling and receive a 200 OK
response. \n For more information, see Resolve \n the Error 200 response when copying objects to Amazon S3. \n The 200 OK
status code means the copy was accepted, but \n it doesn't mean the copy is complete. Another example is \n when you disconnect from Amazon S3 before the copy is complete, Amazon S3 might cancel the copy and you may receive a 200 OK
response. \n You must stay connected to Amazon S3 until the entire response is successfully received and processed.
If you call this API operation directly, make\n sure to design your application to parse the content of the response and handle it\n appropriately. If you use Amazon Web Services SDKs, SDKs handle this condition. The SDKs detect the\n embedded error and apply error handling per your configuration settings (including\n automatically retrying the request as appropriate). If the condition persists, the SDKs\n throw an exception (or, for the SDKs that don't use exceptions, they return an \n error).
\nThe copy request charge is based on the storage class and Region that you specify for\n the destination object. The request can also result in a data retrieval charge for the\n source if the source storage class bills for data retrieval. If the copy source is in a different region, the data transfer is billed to the copy source account. For pricing information, see\n Amazon S3 pricing.
\n\n Directory buckets - The HTTP Host header syntax is \n Bucket_name.s3express-az_id.region.amazonaws.com
.
The following operations are related to CopyObject
:
Creates a copy of an object that is already stored in Amazon S3.
\nYou can store individual objects of up to 5 TB in Amazon S3. You create a copy of your\n object up to 5 GB in size in a single atomic action using this API. However, to copy an\n object greater than 5 GB, you must use the multipart upload Upload Part - Copy\n (UploadPartCopy) API. For more information, see Copy Object Using the\n REST Multipart Upload API.
\nYou can copy individual objects between general purpose buckets, between directory buckets, and \n between general purpose buckets and directory buckets.
\nAmazon S3 supports copy operations using Multi-Region Access Points only as a destination when using the Multi-Region Access Point ARN.
\n\n Directory buckets - For directory buckets, you must make requests for this API operation to the Zonal endpoint. These endpoints support virtual-hosted-style requests in the format https://bucket_name.s3express-az_id.region.amazonaws.com/key-name\n
. Path-style requests are not supported. For more information, see Regional and Zonal endpoints in the\n Amazon S3 User Guide.
VPC endpoints don't support cross-Region requests (including copies). If you're using VPC endpoints, your source and destination buckets should be in the same Amazon Web Services Region as your VPC endpoint.
\nBoth the\n Region that you want to copy the object from and the Region that you want to copy the\n object to must be enabled for your account. For more information about how to enable a Region for your account, see Enable \n or disable a Region for standalone accounts in the\n Amazon Web Services Account Management Guide.
\nAmazon S3 transfer acceleration does not support cross-Region copies. If you request a\n cross-Region copy using a transfer acceleration endpoint, you get a 400 Bad\n Request
error. For more information, see Transfer\n Acceleration.
All CopyObject
requests must be authenticated and signed by using IAM credentials (access key ID and secret access key for the IAM identities). All headers with the x-amz-
prefix, including\n x-amz-copy-source
, must be signed. For more information, see REST Authentication.
\n Directory buckets - You must use the IAM credentials to authenticate and authorize your access to the CopyObject
API operation, instead of using the \n temporary security credentials through the CreateSession
API operation.
Amazon Web Services CLI or SDKs handles authentication and authorization on your behalf.
\nYou must have\n read access to the source object and write\n access to the destination bucket.
\n\n General purpose bucket permissions -\n You must have permissions in an IAM policy based on the source and destination\n bucket types in a CopyObject
operation.
If the source object is in a general purpose bucket, you must have\n \n s3:GetObject
\n \n permission to read the source object that is being copied.
If the destination bucket is a general purpose bucket, you must have\n \n s3:PutObject
\n \n permission to write the object copy to the destination bucket.
\n Directory bucket permissions -\n You must have permissions in a bucket policy or an IAM identity-based policy based on the source and destination\n bucket types in a CopyObject
operation.
If the source object that you want to copy is in a\n directory bucket, you must have the \n s3express:CreateSession
\n permission in\n the Action
element of a policy to read the object. By default, the session is in the ReadWrite
mode. If you want to restrict the access, you can explicitly set the s3express:SessionMode
condition key to ReadOnly
on the copy source bucket.
If the copy destination is a directory bucket, you must have the \n s3express:CreateSession
\n permission in the\n Action
element of a policy to write the object\n to the destination. The s3express:SessionMode
condition\n key can't be set to ReadOnly
on the copy destination bucket.
If the object is encrypted with\n SSE-KMS, you must also have the\n kms:GenerateDataKey
and kms:Decrypt
permissions in IAM identity-based policies and KMS key policies for the KMS key.
For example policies, see Example bucket policies for S3 Express One Zone and Amazon Web Services Identity and Access Management (IAM) identity-based policies for S3 Express One Zone in the\n Amazon S3 User Guide.
\nWhen the request is an HTTP 1.1 request, the response is chunk encoded. When\n the request is not an HTTP 1.1 request, the response would not contain the\n Content-Length
. You always need to read the entire response body\n to check if the copy succeeds.
If the copy is successful, you receive a response with information about the copied\n object.
\nA copy request might return an error when Amazon S3 receives the copy request or while Amazon S3\n is copying the files. A 200 OK
response can contain either a success or an error.
If the error occurs before the copy action starts, you receive a\n standard Amazon S3 error.
\nIf the error occurs during the copy operation, the error response is\n embedded in the 200 OK
response. For example, in a cross-region copy, you \n may encounter throttling and receive a 200 OK
response. \n For more information, see Resolve \n the Error 200 response when copying objects to Amazon S3. \n The 200 OK
status code means the copy was accepted, but \n it doesn't mean the copy is complete. Another example is \n when you disconnect from Amazon S3 before the copy is complete, Amazon S3 might cancel the copy and you may receive a 200 OK
response. \n You must stay connected to Amazon S3 until the entire response is successfully received and processed.
If you call this API operation directly, make\n sure to design your application to parse the content of the response and handle it\n appropriately. If you use Amazon Web Services SDKs, SDKs handle this condition. The SDKs detect the\n embedded error and apply error handling per your configuration settings (including\n automatically retrying the request as appropriate). If the condition persists, the SDKs\n throw an exception (or, for the SDKs that don't use exceptions, they return an \n error).
\nThe copy request charge is based on the storage class and Region that you specify for\n the destination object. The request can also result in a data retrieval charge for the\n source if the source storage class bills for data retrieval. If the copy source is in a different region, the data transfer is billed to the copy source account. For pricing information, see\n Amazon S3 pricing.
\n\n Directory buckets - The HTTP Host header syntax is \n Bucket_name.s3express-az_id.region.amazonaws.com
.
The following operations are related to CopyObject
:
The server-side encryption algorithm used when you store this object in Amazon S3 (for example,\n AES256
, aws:kms
, aws:kms:dsse
).
For directory buckets, only server-side encryption with Amazon S3 managed keys (SSE-S3) (AES256
) is supported.
The server-side encryption algorithm used when you store this object in Amazon S3 (for example,\n AES256
, aws:kms
, aws:kms:dsse
).
If present, indicates the ID of the Key Management Service (KMS) symmetric encryption customer managed key\n that was used for the object.
\nThis functionality is not supported for directory buckets.
\nIf present, indicates the ID of the KMS key that was used for object encryption.
", "smithy.api#httpHeader": "x-amz-server-side-encryption-aws-kms-key-id" } }, "SSEKMSEncryptionContext": { "target": "com.amazonaws.s3#SSEKMSEncryptionContext", "traits": { - "smithy.api#documentation": "If present, indicates the Amazon Web Services KMS Encryption Context to use for object encryption. The\n value of this header is a base64-encoded UTF-8 string holding JSON with the encryption\n context key-value pairs.
\nThis functionality is not supported for directory buckets.
\nIf present, indicates the Amazon Web Services KMS Encryption Context to use for object encryption. The\n value of this header is a base64-encoded UTF-8 string holding JSON with the encryption\n context key-value pairs.
", "smithy.api#httpHeader": "x-amz-server-side-encryption-context" } }, "BucketKeyEnabled": { "target": "com.amazonaws.s3#BucketKeyEnabled", "traits": { - "smithy.api#documentation": "Indicates whether the copied object uses an S3 Bucket Key for server-side encryption\n with Key Management Service (KMS) keys (SSE-KMS).
\nThis functionality is not supported for directory buckets.
\nIndicates whether the copied object uses an S3 Bucket Key for server-side encryption\n with Key Management Service (KMS) keys (SSE-KMS).
", "smithy.api#httpHeader": "x-amz-server-side-encryption-bucket-key-enabled" } }, @@ -18987,7 +18987,7 @@ "ServerSideEncryption": { "target": "com.amazonaws.s3#ServerSideEncryption", "traits": { - "smithy.api#documentation": "The server-side encryption algorithm used when storing this object in Amazon S3 (for example,\n AES256
, aws:kms
, aws:kms:dsse
). Unrecognized or unsupported values won’t write a destination object and will receive a 400 Bad Request
response.
Amazon S3 automatically encrypts all new objects that are copied to an S3 bucket.\n When copying an object, if you don't specify encryption information in your copy\n request, the encryption setting of the target object is set to the default\n encryption configuration of the destination bucket. By default, all buckets have a\n base level of encryption configuration that uses server-side encryption with Amazon S3\n managed keys (SSE-S3). If the destination bucket has a default encryption\n configuration that uses server-side encryption with Key Management Service (KMS) keys\n (SSE-KMS), dual-layer server-side encryption with Amazon Web Services KMS keys (DSSE-KMS), or\n server-side encryption with customer-provided encryption keys (SSE-C), Amazon S3 uses\n the corresponding KMS key, or a customer-provided key to encrypt the target\n object copy.
\nWhen you perform a CopyObject
operation, if you want to use a\n different type of encryption setting for the target object, you can specify \n appropriate encryption-related headers to encrypt the target object with an Amazon S3 managed key, a\n KMS key, or a customer-provided key. If the encryption setting in\n your request is different from the default encryption configuration of the\n destination bucket, the encryption setting in your request takes precedence.
With server-side\n encryption, Amazon S3 encrypts your data as it writes your data to disks in its data\n centers and decrypts the data when you access it. For more information about server-side encryption, see Using\n Server-Side Encryption in the\n Amazon S3 User Guide.
\nFor directory buckets, only server-side encryption with Amazon S3 managed keys (SSE-S3) (AES256
) is supported.
The server-side encryption algorithm used when storing this object in Amazon S3. Unrecognized or unsupported values won’t write a destination object and will receive a 400 Bad Request
response.
Amazon S3 automatically encrypts all new objects that are copied to an S3 bucket.\n When copying an object, if you don't specify encryption information in your copy\n request, the encryption setting of the target object is set to the default\n encryption configuration of the destination bucket. By default, all buckets have a\n base level of encryption configuration that uses server-side encryption with Amazon S3\n managed keys (SSE-S3). If the destination bucket has a different default encryption\n configuration, Amazon S3 uses\n the corresponding encryption key to encrypt the target\n object copy.
\nWith server-side\n encryption, Amazon S3 encrypts your data as it writes your data to disks in its data\n centers and decrypts the data when you access it. For more information about server-side encryption, see Using\n Server-Side Encryption in the\n Amazon S3 User Guide.
\n\n General purpose buckets \n
\nFor general purpose buckets, there are the following supported options for server-side encryption: server-side encryption with Key Management Service (KMS) keys\n (SSE-KMS), dual-layer server-side encryption with Amazon Web Services KMS keys (DSSE-KMS), and \n server-side encryption with customer-provided encryption keys (SSE-C). Amazon S3 uses\n the corresponding KMS key, or a customer-provided key to encrypt the target\n object copy.
\nWhen you perform a CopyObject
operation, if you want to use a\n different type of encryption setting for the target object, you can specify \n appropriate encryption-related headers to encrypt the target object with an Amazon S3 managed key, a\n KMS key, or a customer-provided key. If the encryption setting in\n your request is different from the default encryption configuration of the\n destination bucket, the encryption setting in your request takes precedence.
\n Directory buckets \n
\nFor directory buckets, there are only two supported options for server-side encryption: server-side encryption with Amazon S3 managed keys (SSE-S3) (AES256
) and server-side encryption with KMS keys (SSE-KMS) (aws:kms
). We recommend that the bucket's default encryption uses the desired encryption configuration and you don't override the bucket default encryption in your \n CreateSession
requests or PUT
object requests. Then, new objects \n are automatically encrypted with the desired encryption settings. For more\n information, see Protecting data with server-side encryption in the Amazon S3 User Guide. For more information about the encryption overriding behaviors in directory buckets, see Specifying server-side encryption with KMS for new object uploads.
To encrypt new object copies to a directory bucket with SSE-KMS, we recommend you specify SSE-KMS as the directory bucket's default encryption configuration with a KMS key (specifically, a customer managed key). \n Amazon Web Services managed key (aws/s3
) isn't supported. Your SSE-KMS configuration can only support 1 customer managed key per directory bucket for the lifetime of the bucket. After you specify a customer managed key for SSE-KMS, you can't override the customer managed key for the bucket's SSE-KMS configuration. \n Then, when you perform a CopyObject
operation and want to specify server-side encryption settings for new object copies with SSE-KMS in the encryption-related request headers, you must ensure the encryption key is the same customer managed key that you specified for the directory bucket's default encryption configuration. \n
Specifies the KMS ID (Key ID, Key ARN, or Key Alias) to use for object encryption. All GET and PUT requests for an\n object protected by KMS will fail if they're not made via SSL or using SigV4. For\n information about configuring any of the officially supported Amazon Web Services SDKs and Amazon Web Services CLI, see\n Specifying the\n Signature Version in Request Authentication in the\n Amazon S3 User Guide.
\nThis functionality is not supported when the destination bucket is a directory bucket.
\nSpecifies the KMS key ID (Key ID, Key ARN, or Key Alias) to use for object encryption. All GET and PUT requests for an\n object protected by KMS will fail if they're not made via SSL or using SigV4. For\n information about configuring any of the officially supported Amazon Web Services SDKs and Amazon Web Services CLI, see\n Specifying the\n Signature Version in Request Authentication in the\n Amazon S3 User Guide.
\n\n Directory buckets - If you specify x-amz-server-side-encryption
with aws:kms
, you must specify the \n x-amz-server-side-encryption-aws-kms-key-id
header with the ID (Key ID or Key ARN) of the KMS \n symmetric encryption customer managed key to use. Otherwise, you get an HTTP 400 Bad Request
error. Only use the key ID or key ARN. The key alias format of the KMS key isn't supported. Your SSE-KMS configuration can only support 1 customer managed key per directory bucket for the lifetime of the bucket. \nAmazon Web Services managed key (aws/s3
) isn't supported. \n
Specifies the Amazon Web Services KMS Encryption Context to use for object encryption. The value of\n this header is a base64-encoded UTF-8 string holding JSON with the encryption context\n key-value pairs. This value must be explicitly added to specify encryption context for \n CopyObject
requests.
This functionality is not supported when the destination bucket is a directory bucket.
\nSpecifies the Amazon Web Services KMS Encryption Context as an additional encryption context to use for the destination object encryption. The value of\n this header is a base64-encoded UTF-8 string holding JSON with the encryption context\n key-value pairs.
\n\n General purpose buckets - This value must be explicitly added to specify encryption context for \n CopyObject
requests if you want an additional encryption context for your destination object. The additional encryption context of the source object won't be copied to the destination object. For more information, see Encryption context in the Amazon S3 User Guide.
\n Directory buckets - You can optionally provide an explicit encryption context value. The value must match the default encryption context - the bucket Amazon Resource Name (ARN). An additional encryption context value is not supported.
", "smithy.api#httpHeader": "x-amz-server-side-encryption-context" } }, "BucketKeyEnabled": { "target": "com.amazonaws.s3#BucketKeyEnabled", "traits": { - "smithy.api#documentation": "Specifies whether Amazon S3 should use an S3 Bucket Key for object encryption with\n server-side encryption using Key Management Service (KMS) keys (SSE-KMS). If a target object uses SSE-KMS, you can enable an S3 Bucket Key for the\n object.
\nSetting this header to\n true
causes Amazon S3 to use an S3 Bucket Key for object encryption with\n SSE-KMS. Specifying this header with a COPY action doesn’t affect bucket-level settings for S3\n Bucket Key.
For more information, see Amazon S3 Bucket Keys in the\n Amazon S3 User Guide.
\nThis functionality is not supported when the destination bucket is a directory bucket.
\nSpecifies whether Amazon S3 should use an S3 Bucket Key for object encryption with\n server-side encryption using Key Management Service (KMS) keys (SSE-KMS). If a target object uses SSE-KMS, you can enable an S3 Bucket Key for the\n object.
\nSetting this header to\n true
causes Amazon S3 to use an S3 Bucket Key for object encryption with\n SSE-KMS. Specifying this header with a COPY action doesn’t affect bucket-level settings for S3\n Bucket Key.
For more information, see Amazon S3 Bucket Keys in the\n Amazon S3 User Guide.
\n\n Directory buckets - S3 Bucket Keys aren't supported, when you copy SSE-KMS encrypted objects from general purpose buckets \nto directory buckets, from directory buckets to general purpose buckets, or between directory buckets, through CopyObject. In this case, Amazon S3 makes a call to KMS every time a copy request is made for a KMS-encrypted object.
\nThis action initiates a multipart upload and returns an upload ID. This upload ID is\n used to associate all of the parts in the specific multipart upload. You specify this\n upload ID in each of your subsequent upload part requests (see UploadPart). You also include this\n upload ID in the final request to either complete or abort the multipart upload\n request. For more information about multipart uploads, see Multipart Upload Overview in the Amazon S3 User Guide.
\nAfter you initiate a multipart upload and upload one or more parts, to stop being\n charged for storing the uploaded parts, you must either complete or abort the multipart\n upload. Amazon S3 frees up the space used to store the parts and stops charging you for\n storing them only after you either complete or abort a multipart upload.
\nIf you have configured a lifecycle rule to abort incomplete multipart uploads, the created multipart \n upload must be completed within the number of days specified in the bucket lifecycle\n configuration. Otherwise, the incomplete multipart upload becomes eligible for an abort\n action and Amazon S3 aborts the multipart upload. For more information, see Aborting Incomplete Multipart Uploads Using a Bucket Lifecycle\n Configuration.
\n\n Directory buckets - S3 Lifecycle is not supported by directory buckets.
\n\n Directory buckets - For directory buckets, you must make requests for this API operation to the Zonal endpoint. These endpoints support virtual-hosted-style requests in the format https://bucket_name.s3express-az_id.region.amazonaws.com/key-name\n
. Path-style requests are not supported. For more information, see Regional and Zonal endpoints in the\n Amazon S3 User Guide.
For request signing, multipart upload is just a series of regular requests. You initiate\n a multipart upload, send one or more requests to upload parts, and then complete the\n multipart upload process. You sign each request individually. There is nothing special\n about signing multipart upload requests. For more information about signing, see Authenticating Requests (Amazon Web Services Signature Version 4) in the Amazon S3 User Guide.
\n\n General purpose bucket permissions - To\n perform a multipart upload with encryption using an Key Management Service (KMS)\n KMS key, the requester must have permission to the\n kms:Decrypt
and kms:GenerateDataKey
actions on\n the key. The requester must also have permissions for the\n kms:GenerateDataKey
action for the\n CreateMultipartUpload
API. Then, the requester needs\n permissions for the kms:Decrypt
action on the\n UploadPart
and UploadPartCopy
APIs. These\n permissions are required because Amazon S3 must decrypt and read data from the\n encrypted file parts before it completes the multipart upload. For more\n information, see Multipart upload API and permissions and Protecting data\n using server-side encryption with Amazon Web Services KMS in the\n Amazon S3 User Guide.
\n Directory bucket permissions - To grant access to this API operation on a directory bucket, we recommend that you use the \n CreateSession
\n API operation for session-based authorization. Specifically, you grant the s3express:CreateSession
permission to the directory bucket in a bucket policy or an IAM identity-based policy. Then, you make the CreateSession
API call on the bucket to obtain a session token. With the session token in your request header, you can make API requests to this operation. After the session token expires, you make another CreateSession
API call to generate a new session token for use. \nAmazon Web Services CLI or SDKs create session and refresh the session token automatically to avoid service interruptions when a session expires. For more information about authorization, see \n CreateSession
\n .
\n General purpose buckets - Server-side encryption is for data encryption at rest. Amazon S3 encrypts your data as it\n writes it to disks in its data centers and decrypts it when you access it. Amazon S3\n automatically encrypts all new objects that are uploaded to an S3 bucket. When doing a\n multipart upload, if you don't specify encryption information in your request, the\n encryption setting of the uploaded parts is set to the default encryption configuration of\n the destination bucket. By default, all buckets have a base level of encryption\n configuration that uses server-side encryption with Amazon S3 managed keys (SSE-S3). If the\n destination bucket has a default encryption configuration that uses server-side encryption\n with an Key Management Service (KMS) key (SSE-KMS), or a customer-provided encryption key (SSE-C),\n Amazon S3 uses the corresponding KMS key, or a customer-provided key to encrypt the uploaded\n parts. When you perform a CreateMultipartUpload operation, if you want to use a different\n type of encryption setting for the uploaded parts, you can request that Amazon S3 encrypts the\n object with a different encryption key (such as an Amazon S3 managed key, a KMS key, or a customer-provided key). When the encryption\n setting in your request is different from the default encryption configuration of the\n destination bucket, the encryption setting in your request takes precedence. If you choose\n to provide your own encryption key, the request headers you provide in UploadPart\n and UploadPartCopy requests must match the headers you used in the CreateMultipartUpload
request.
Use KMS keys (SSE-KMS) that include the Amazon Web Services managed key\n (aws/s3
) and KMS customer managed keys stored in Key Management Service (KMS) –\n If you want Amazon Web Services to manage the keys used to encrypt data, specify the\n following headers in the request.
\n x-amz-server-side-encryption
\n
\n x-amz-server-side-encryption-aws-kms-key-id
\n
\n x-amz-server-side-encryption-context
\n
If you specify x-amz-server-side-encryption:aws:kms
, but\n don't provide x-amz-server-side-encryption-aws-kms-key-id
,\n Amazon S3 uses the Amazon Web Services managed key (aws/s3
key) in KMS to\n protect the data.
To perform a multipart upload with encryption by using an Amazon Web Services KMS key, the requester\n must have permission to the kms:Decrypt
and kms:GenerateDataKey*
\n actions on the key. These permissions are required because Amazon S3 must decrypt and read data\n from the encrypted file parts before it completes the multipart upload. For more\n information, see Multipart upload API\n and permissions and Protecting data using\n server-side encryption with Amazon Web Services KMS in the\n Amazon S3 User Guide.
If your Identity and Access Management (IAM) user or role is in the same Amazon Web Services account as the KMS key,\n then you must have these permissions on the key policy. If your IAM user or role is in a different account from the key, then you must have the permissions on both the key\n policy and your IAM user or role.
\nAll GET
and PUT
requests for an object\n protected by KMS fail if you don't make them by using Secure Sockets\n Layer (SSL), Transport Layer Security (TLS), or Signature Version\n 4. For information about configuring any of the officially supported Amazon Web Services\n SDKs and Amazon Web Services CLI, see Specifying the Signature Version in Request Authentication\n in the Amazon S3 User Guide.
For more information about server-side encryption with KMS keys\n (SSE-KMS), see Protecting Data\n Using Server-Side Encryption with KMS keys in the Amazon S3 User Guide.
\nUse customer-provided encryption keys (SSE-C) – If you want to manage\n your own encryption keys, provide all the following headers in the\n request.
\n\n x-amz-server-side-encryption-customer-algorithm
\n
\n x-amz-server-side-encryption-customer-key
\n
\n x-amz-server-side-encryption-customer-key-MD5
\n
For more information about server-side encryption with customer-provided\n encryption keys (SSE-C), see \n Protecting data using server-side encryption with customer-provided\n encryption keys (SSE-C) in the Amazon S3 User Guide.
\n\n Directory buckets -For directory buckets, only server-side encryption with Amazon S3 managed keys (SSE-S3) (AES256
) is supported.
\n Directory buckets - The HTTP Host header syntax is \n Bucket_name.s3express-az_id.region.amazonaws.com
.
The following operations are related to CreateMultipartUpload
:
\n UploadPart\n
\n\n AbortMultipartUpload\n
\n\n ListParts\n
\n\n ListMultipartUploads\n
\nThis action initiates a multipart upload and returns an upload ID. This upload ID is\n used to associate all of the parts in the specific multipart upload. You specify this\n upload ID in each of your subsequent upload part requests (see UploadPart). You also include this\n upload ID in the final request to either complete or abort the multipart upload\n request. For more information about multipart uploads, see Multipart Upload Overview in the Amazon S3 User Guide.
\nAfter you initiate a multipart upload and upload one or more parts, to stop being\n charged for storing the uploaded parts, you must either complete or abort the multipart\n upload. Amazon S3 frees up the space used to store the parts and stops charging you for\n storing them only after you either complete or abort a multipart upload.
\nIf you have configured a lifecycle rule to abort incomplete multipart uploads, the created multipart \n upload must be completed within the number of days specified in the bucket lifecycle\n configuration. Otherwise, the incomplete multipart upload becomes eligible for an abort\n action and Amazon S3 aborts the multipart upload. For more information, see Aborting Incomplete Multipart Uploads Using a Bucket Lifecycle\n Configuration.
\n\n Directory buckets - S3 Lifecycle is not supported by directory buckets.
\n\n Directory buckets - For directory buckets, you must make requests for this API operation to the Zonal endpoint. These endpoints support virtual-hosted-style requests in the format https://bucket_name.s3express-az_id.region.amazonaws.com/key-name\n
. Path-style requests are not supported. For more information, see Regional and Zonal endpoints in the\n Amazon S3 User Guide.
For request signing, multipart upload is just a series of regular requests. You initiate\n a multipart upload, send one or more requests to upload parts, and then complete the\n multipart upload process. You sign each request individually. There is nothing special\n about signing multipart upload requests. For more information about signing, see Authenticating Requests (Amazon Web Services Signature Version 4) in the Amazon S3 User Guide.
\n\n General purpose bucket permissions - To\n perform a multipart upload with encryption using an Key Management Service (KMS)\n KMS key, the requester must have permission to the\n kms:Decrypt
and kms:GenerateDataKey
actions on\n the key. The requester must also have permissions for the\n kms:GenerateDataKey
action for the\n CreateMultipartUpload
API. Then, the requester needs\n permissions for the kms:Decrypt
action on the\n UploadPart
and UploadPartCopy
APIs. These\n permissions are required because Amazon S3 must decrypt and read data from the\n encrypted file parts before it completes the multipart upload. For more\n information, see Multipart upload API and permissions and Protecting data\n using server-side encryption with Amazon Web Services KMS in the\n Amazon S3 User Guide.
\n Directory bucket permissions - To grant access to this API operation on a directory bucket, we recommend that you use the \n CreateSession
\n API operation for session-based authorization. Specifically, you grant the s3express:CreateSession
permission to the directory bucket in a bucket policy or an IAM identity-based policy. Then, you make the CreateSession
API call on the bucket to obtain a session token. With the session token in your request header, you can make API requests to this operation. After the session token expires, you make another CreateSession
API call to generate a new session token for use. \nAmazon Web Services CLI or SDKs create session and refresh the session token automatically to avoid service interruptions when a session expires. For more information about authorization, see \n CreateSession
\n .
\n General purpose buckets - Server-side encryption is for data encryption at rest. Amazon S3 encrypts your data as it\n writes it to disks in its data centers and decrypts it when you access it. Amazon S3\n automatically encrypts all new objects that are uploaded to an S3 bucket. When doing a\n multipart upload, if you don't specify encryption information in your request, the\n encryption setting of the uploaded parts is set to the default encryption configuration of\n the destination bucket. By default, all buckets have a base level of encryption\n configuration that uses server-side encryption with Amazon S3 managed keys (SSE-S3). If the\n destination bucket has a default encryption configuration that uses server-side encryption\n with an Key Management Service (KMS) key (SSE-KMS), or a customer-provided encryption key (SSE-C),\n Amazon S3 uses the corresponding KMS key, or a customer-provided key to encrypt the uploaded\n parts. When you perform a CreateMultipartUpload operation, if you want to use a different\n type of encryption setting for the uploaded parts, you can request that Amazon S3 encrypts the\n object with a different encryption key (such as an Amazon S3 managed key, a KMS key, or a customer-provided key). When the encryption\n setting in your request is different from the default encryption configuration of the\n destination bucket, the encryption setting in your request takes precedence. If you choose\n to provide your own encryption key, the request headers you provide in UploadPart\n and UploadPartCopy requests must match the headers you used in the CreateMultipartUpload
request.
Use KMS keys (SSE-KMS) that include the Amazon Web Services managed key\n (aws/s3
) and KMS customer managed keys stored in Key Management Service (KMS) –\n If you want Amazon Web Services to manage the keys used to encrypt data, specify the\n following headers in the request.
\n x-amz-server-side-encryption
\n
\n x-amz-server-side-encryption-aws-kms-key-id
\n
\n x-amz-server-side-encryption-context
\n
If you specify x-amz-server-side-encryption:aws:kms
, but\n don't provide x-amz-server-side-encryption-aws-kms-key-id
,\n Amazon S3 uses the Amazon Web Services managed key (aws/s3
key) in KMS to\n protect the data.
To perform a multipart upload with encryption by using an Amazon Web Services KMS key, the requester\n must have permission to the kms:Decrypt
and kms:GenerateDataKey*
\n actions on the key. These permissions are required because Amazon S3 must decrypt and read data\n from the encrypted file parts before it completes the multipart upload. For more\n information, see Multipart upload API\n and permissions and Protecting data using\n server-side encryption with Amazon Web Services KMS in the\n Amazon S3 User Guide.
If your Identity and Access Management (IAM) user or role is in the same Amazon Web Services account as the KMS key,\n then you must have these permissions on the key policy. If your IAM user or role is in a different account from the key, then you must have the permissions on both the key\n policy and your IAM user or role.
\nAll GET
and PUT
requests for an object\n protected by KMS fail if you don't make them by using Secure Sockets\n Layer (SSL), Transport Layer Security (TLS), or Signature Version\n 4. For information about configuring any of the officially supported Amazon Web Services\n SDKs and Amazon Web Services CLI, see Specifying the Signature Version in Request Authentication\n in the Amazon S3 User Guide.
For more information about server-side encryption with KMS keys\n (SSE-KMS), see Protecting Data\n Using Server-Side Encryption with KMS keys in the Amazon S3 User Guide.
\nUse customer-provided encryption keys (SSE-C) – If you want to manage\n your own encryption keys, provide all the following headers in the\n request.
\n\n x-amz-server-side-encryption-customer-algorithm
\n
\n x-amz-server-side-encryption-customer-key
\n
\n x-amz-server-side-encryption-customer-key-MD5
\n
For more information about server-side encryption with customer-provided\n encryption keys (SSE-C), see \n Protecting data using server-side encryption with customer-provided\n encryption keys (SSE-C) in the Amazon S3 User Guide.
\n\n Directory buckets - For directory buckets, there are only two supported options for server-side encryption: server-side encryption with Amazon S3 managed keys (SSE-S3) (AES256
) and server-side encryption with KMS keys (SSE-KMS) (aws:kms
). We recommend that the bucket's default encryption uses the desired encryption configuration and you don't override the bucket default encryption in your \n CreateSession
requests or PUT
object requests. Then, new objects \n are automatically encrypted with the desired encryption settings. For more\n information, see Protecting data with server-side encryption in the Amazon S3 User Guide. For more information about the encryption overriding behaviors in directory buckets, see Specifying server-side encryption with KMS for new object uploads.
In the Zonal endpoint API calls (except CopyObject and UploadPartCopy) using the REST API, the encryption request headers must match the encryption settings that are specified in the CreateSession
request. \n You can't override the values of the encryption settings (x-amz-server-side-encryption
, x-amz-server-side-encryption-aws-kms-key-id
, x-amz-server-side-encryption-context
, and x-amz-server-side-encryption-bucket-key-enabled
) that are specified in the CreateSession
request. \n You don't need to explicitly specify these encryption settings values in Zonal endpoint API calls, and \n Amazon S3 will use the encryption settings values from the CreateSession
request to protect new objects in the directory bucket. \n
When you use the CLI or the Amazon Web Services SDKs, for CreateSession
, the session token refreshes automatically to avoid service interruptions when a session expires. The CLI or the Amazon Web Services SDKs use the bucket's default encryption configuration for the \n CreateSession
request. It's not supported to override the encryption settings values in the CreateSession
request. \n So in the Zonal endpoint API calls (except CopyObject and UploadPartCopy), \n the encryption request headers must match the default encryption configuration of the directory bucket.\n\n
For directory buckets, when you perform a CreateMultipartUpload
operation and an UploadPartCopy
operation, \n the request headers you provide in the CreateMultipartUpload
request must match the default encryption configuration of the destination bucket.
\n Directory buckets - The HTTP Host header syntax is \n Bucket_name.s3express-az_id.region.amazonaws.com
.
The following operations are related to CreateMultipartUpload
:
\n UploadPart\n
\n\n AbortMultipartUpload\n
\n\n ListParts\n
\n\n ListMultipartUploads\n
\nThe server-side encryption algorithm used when you store this object in Amazon S3 (for example,\n AES256
, aws:kms
).
For directory buckets, only server-side encryption with Amazon S3 managed keys (SSE-S3) (AES256
) is supported.
The server-side encryption algorithm used when you store this object in Amazon S3 (for example,\n AES256
, aws:kms
).
If present, indicates the ID of the Key Management Service (KMS) symmetric encryption customer managed key\n that was used for the object.
\nThis functionality is not supported for directory buckets.
\nIf present, indicates the ID of the KMS key that was used for object encryption.
", "smithy.api#httpHeader": "x-amz-server-side-encryption-aws-kms-key-id" } }, "SSEKMSEncryptionContext": { "target": "com.amazonaws.s3#SSEKMSEncryptionContext", "traits": { - "smithy.api#documentation": "If present, indicates the Amazon Web Services KMS Encryption Context to use for object encryption. The\n value of this header is a base64-encoded UTF-8 string holding JSON with the encryption\n context key-value pairs.
\nThis functionality is not supported for directory buckets.
\nIf present, indicates the Amazon Web Services KMS Encryption Context to use for object encryption. The value of\n this header is a Base64-encoded string of a UTF-8 encoded JSON, which contains the encryption context as key-value pairs.
", "smithy.api#httpHeader": "x-amz-server-side-encryption-context" } }, "BucketKeyEnabled": { "target": "com.amazonaws.s3#BucketKeyEnabled", "traits": { - "smithy.api#documentation": "Indicates whether the multipart upload uses an S3 Bucket Key for server-side encryption\n with Key Management Service (KMS) keys (SSE-KMS).
\nThis functionality is not supported for directory buckets.
\nIndicates whether the multipart upload uses an S3 Bucket Key for server-side encryption\n with Key Management Service (KMS) keys (SSE-KMS).
", "smithy.api#httpHeader": "x-amz-server-side-encryption-bucket-key-enabled" } }, @@ -19667,7 +19667,7 @@ "ServerSideEncryption": { "target": "com.amazonaws.s3#ServerSideEncryption", "traits": { - "smithy.api#documentation": "The server-side encryption algorithm used when you store this object in Amazon S3 (for example,\n AES256
, aws:kms
).
For directory buckets, only server-side encryption with Amazon S3 managed keys (SSE-S3) (AES256
) is supported.
The server-side encryption algorithm used when you store this object in Amazon S3 (for example,\n AES256
, aws:kms
).
\n Directory buckets - For directory buckets, there are only two supported options for server-side encryption: server-side encryption with Amazon S3 managed keys (SSE-S3) (AES256
) and server-side encryption with KMS keys (SSE-KMS) (aws:kms
). We recommend that the bucket's default encryption uses the desired encryption configuration and you don't override the bucket default encryption in your \n CreateSession
requests or PUT
object requests. Then, new objects \n are automatically encrypted with the desired encryption settings. For more\n information, see Protecting data with server-side encryption in the Amazon S3 User Guide. For more information about the encryption overriding behaviors in directory buckets, see Specifying server-side encryption with KMS for new object uploads. \n
In the Zonal endpoint API calls (except CopyObject and UploadPartCopy) using the REST API, the encryption request headers must match the encryption settings that are specified in the CreateSession
request. \n You can't override the values of the encryption settings (x-amz-server-side-encryption
, x-amz-server-side-encryption-aws-kms-key-id
, x-amz-server-side-encryption-context
, and x-amz-server-side-encryption-bucket-key-enabled
) that are specified in the CreateSession
request. \n You don't need to explicitly specify these encryption settings values in Zonal endpoint API calls, and \n Amazon S3 will use the encryption settings values from the CreateSession
request to protect new objects in the directory bucket. \n
When you use the CLI or the Amazon Web Services SDKs, for CreateSession
, the session token refreshes automatically to avoid service interruptions when a session expires. The CLI or the Amazon Web Services SDKs use the bucket's default encryption configuration for the \n CreateSession
request. It's not supported to override the encryption settings values in the CreateSession
request. \n So in the Zonal endpoint API calls (except CopyObject and UploadPartCopy), \n the encryption request headers must match the default encryption configuration of the directory bucket.\n\n
Specifies the ID (Key ID, Key ARN, or Key Alias) of the symmetric encryption customer managed key to use for object encryption.
\nThis functionality is not supported for directory buckets.
\nSpecifies the KMS key ID (Key ID, Key ARN, or Key Alias) to use for object encryption. If the KMS key doesn't exist in the same\n account that's issuing the command, you must use the full Key ARN not the Key ID.
\n\n General purpose buckets - If you specify x-amz-server-side-encryption
with aws:kms
or aws:kms:dsse
, this header specifies the ID (Key ID, Key ARN, or Key Alias) of the KMS \n key to use. If you specify\n x-amz-server-side-encryption:aws:kms
or\n x-amz-server-side-encryption:aws:kms:dsse
, but do not provide x-amz-server-side-encryption-aws-kms-key-id
, Amazon S3 uses the Amazon Web Services managed key\n (aws/s3
) to protect the data.
\n Directory buckets - If you specify x-amz-server-side-encryption
with aws:kms
, you must specify the \n x-amz-server-side-encryption-aws-kms-key-id
header with the ID (Key ID or Key ARN) of the KMS \n symmetric encryption customer managed key to use. Otherwise, you get an HTTP 400 Bad Request
error. Only use the key ID or key ARN. The key alias format of the KMS key isn't supported. Your SSE-KMS configuration can only support 1 customer managed key per directory bucket for the lifetime of the bucket. \nAmazon Web Services managed key (aws/s3
) isn't supported. \n
Specifies the Amazon Web Services KMS Encryption Context to use for object encryption. The value of\n this header is a base64-encoded UTF-8 string holding JSON with the encryption context\n key-value pairs.
\nThis functionality is not supported for directory buckets.
\nSpecifies the Amazon Web Services KMS Encryption Context to use for object encryption. The value of\n this header is a Base64-encoded string of a UTF-8 encoded JSON, which contains the encryption context as key-value pairs.
\n\n Directory buckets - You can optionally provide an explicit encryption context value. The value must match the default encryption context - the bucket Amazon Resource Name (ARN). An additional encryption context value is not supported.
", "smithy.api#httpHeader": "x-amz-server-side-encryption-context" } }, "BucketKeyEnabled": { "target": "com.amazonaws.s3#BucketKeyEnabled", "traits": { - "smithy.api#documentation": "Specifies whether Amazon S3 should use an S3 Bucket Key for object encryption with\n server-side encryption using Key Management Service (KMS) keys (SSE-KMS). Setting this header to\n true
causes Amazon S3 to use an S3 Bucket Key for object encryption with\n SSE-KMS.
Specifying this header with an object action doesn’t affect bucket-level settings for S3\n Bucket Key.
\nThis functionality is not supported for directory buckets.
\nSpecifies whether Amazon S3 should use an S3 Bucket Key for object encryption with\n server-side encryption using Key Management Service (KMS) keys (SSE-KMS).
\n\n General purpose buckets - Setting this header to\n true
causes Amazon S3 to use an S3 Bucket Key for object encryption with\n SSE-KMS. Also, specifying this header with a PUT action doesn't affect bucket-level settings for S3\n Bucket Key.
\n Directory buckets - S3 Bucket Keys are always enabled for GET
and PUT
operations in a directory bucket and can’t be disabled. S3 Bucket Keys aren't supported, when you copy SSE-KMS encrypted objects from general purpose buckets \nto directory buckets, from directory buckets to general purpose buckets, or between directory buckets, through CopyObject, UploadPartCopy, the Copy operation in Batch Operations, or \n the import jobs. In this case, Amazon S3 makes a call to KMS every time a copy request is made for a KMS-encrypted object.
Creates a session that establishes temporary security credentials to support fast authentication and authorization for the Zonal endpoint APIs on directory buckets. \n For more information about Zonal endpoint APIs that include the Availability Zone in the request endpoint, see \n S3 Express One Zone APIs in the Amazon S3 User Guide. \n
\nTo make Zonal endpoint API requests on a directory bucket, use the CreateSession
\n API operation. Specifically, you grant s3express:CreateSession
permission to a\n bucket in a bucket policy or an IAM identity-based policy. Then, you use IAM credentials to make the\n CreateSession
API request on the bucket, which returns temporary security\n credentials that include the access key ID, secret access key, session token, and\n expiration. These credentials have associated permissions to access the Zonal endpoint APIs. After\n the session is created, you don’t need to use other policies to grant permissions to each\n Zonal endpoint API individually. Instead, in your Zonal endpoint API requests, you sign your requests by\n applying the temporary security credentials of the session to the request headers and\n following the SigV4 protocol for authentication. You also apply the session token to the\n x-amz-s3session-token
request header for authorization. Temporary security\n credentials are scoped to the bucket and expire after 5 minutes. After the expiration time,\n any calls that you make with those credentials will fail. You must use IAM credentials\n again to make a CreateSession
API request that generates a new set of\n temporary credentials for use. Temporary credentials cannot be extended or refreshed beyond\n the original specified interval.
If you use Amazon Web Services SDKs, SDKs handle the session token refreshes automatically to avoid\n service interruptions when a session expires. We recommend that you use the Amazon Web Services SDKs to\n initiate and manage requests to the CreateSession API. For more information, see Performance guidelines and design patterns in the\n Amazon S3 User Guide.
\nYou must make requests for this API operation to the Zonal endpoint. These endpoints support virtual-hosted-style requests in the format https://bucket_name.s3express-az_id.region.amazonaws.com
. Path-style requests are not supported. For more information, see Regional and Zonal endpoints in the\n Amazon S3 User Guide.
\n \n CopyObject
API operation - Unlike other Zonal endpoint APIs, the CopyObject
API operation doesn't use the temporary security credentials returned from the CreateSession
API operation for authentication and authorization. For information about authentication and authorization of the CopyObject
API operation on directory buckets, see CopyObject.
\n \n HeadBucket
API operation - Unlike other Zonal endpoint APIs, the HeadBucket
API operation doesn't use the temporary security credentials returned from the CreateSession
API operation for authentication and authorization. For information about authentication and authorization of the HeadBucket
API operation on directory buckets, see HeadBucket.
To obtain temporary security credentials, you must create a bucket policy or an IAM identity-based policy that\n grants s3express:CreateSession
permission to the bucket. In a\n policy, you can have the s3express:SessionMode
condition key to\n control who can create a ReadWrite
or ReadOnly
session.\n For more information about ReadWrite
or ReadOnly
\n sessions, see \n x-amz-create-session-mode
\n . For example policies, see\n Example bucket policies for S3 Express One Zone and Amazon Web Services Identity and Access Management (IAM) identity-based policies for S3 Express One Zone in the\n Amazon S3 User Guide.
To grant cross-account access to Zonal endpoint APIs, the bucket policy should also grant both accounts the s3express:CreateSession
permission.
\n Directory buckets - The HTTP Host header syntax is \n Bucket_name.s3express-az_id.region.amazonaws.com
.
Creates a session that establishes temporary security credentials to support fast authentication and authorization for the Zonal endpoint API operations on directory buckets. \n For more information about Zonal endpoint API operations that include the Availability Zone in the request endpoint, see \n S3 Express One Zone APIs in the Amazon S3 User Guide. \n
\nTo make Zonal endpoint API requests on a directory bucket, use the CreateSession
\n API operation. Specifically, you grant s3express:CreateSession
permission to a\n bucket in a bucket policy or an IAM identity-based policy. Then, you use IAM credentials to make the\n CreateSession
API request on the bucket, which returns temporary security\n credentials that include the access key ID, secret access key, session token, and\n expiration. These credentials have associated permissions to access the Zonal endpoint API operations. After\n the session is created, you don’t need to use other policies to grant permissions to each\n Zonal endpoint API individually. Instead, in your Zonal endpoint API requests, you sign your requests by\n applying the temporary security credentials of the session to the request headers and\n following the SigV4 protocol for authentication. You also apply the session token to the\n x-amz-s3session-token
request header for authorization. Temporary security\n credentials are scoped to the bucket and expire after 5 minutes. After the expiration time,\n any calls that you make with those credentials will fail. You must use IAM credentials\n again to make a CreateSession
API request that generates a new set of\n temporary credentials for use. Temporary credentials cannot be extended or refreshed beyond\n the original specified interval.
If you use Amazon Web Services SDKs, SDKs handle the session token refreshes automatically to avoid\n service interruptions when a session expires. We recommend that you use the Amazon Web Services SDKs to\n initiate and manage requests to the CreateSession API. For more information, see Performance guidelines and design patterns in the\n Amazon S3 User Guide.
\nYou must make requests for this API operation to the Zonal endpoint. These endpoints support virtual-hosted-style requests in the format https://bucket_name.s3express-az_id.region.amazonaws.com
. Path-style requests are not supported. For more information, see Regional and Zonal endpoints in the\n Amazon S3 User Guide.
\n \n CopyObject
API operation - Unlike other Zonal endpoint API operations, the CopyObject
API operation doesn't use the temporary security credentials returned from the CreateSession
API operation for authentication and authorization. For information about authentication and authorization of the CopyObject
API operation on directory buckets, see CopyObject.
\n \n HeadBucket
API operation - Unlike other Zonal endpoint API operations, the HeadBucket
API operation doesn't use the temporary security credentials returned from the CreateSession
API operation for authentication and authorization. For information about authentication and authorization of the HeadBucket
API operation on directory buckets, see HeadBucket.
To obtain temporary security credentials, you must create a bucket policy or an IAM identity-based policy that\n grants s3express:CreateSession
permission to the bucket. In a\n policy, you can have the s3express:SessionMode
condition key to\n control who can create a ReadWrite
or ReadOnly
session.\n For more information about ReadWrite
or ReadOnly
\n sessions, see \n x-amz-create-session-mode
\n . For example policies, see\n Example bucket policies for S3 Express One Zone and Amazon Web Services Identity and Access Management (IAM) identity-based policies for S3 Express One Zone in the\n Amazon S3 User Guide.
To grant cross-account access to Zonal endpoint API operations, the bucket policy should also grant both accounts the s3express:CreateSession
permission.
If you want to encrypt objects with SSE-KMS, you must also have the kms:GenerateDataKey
and the kms:Decrypt
permissions in IAM identity-based policies and KMS key policies for the target KMS key.
For directory buckets, there are only two supported options for server-side encryption: server-side encryption with Amazon S3 managed keys (SSE-S3) (AES256
) and server-side encryption with KMS keys (SSE-KMS) (aws:kms
). We recommend that the bucket's default encryption uses the desired encryption configuration and you don't override the bucket default encryption in your \n CreateSession
requests or PUT
object requests. Then, new objects \n are automatically encrypted with the desired encryption settings. For more\n information, see Protecting data with server-side encryption in the Amazon S3 User Guide. For more information about the encryption overriding behaviors in directory buckets, see Specifying server-side encryption with KMS for new object uploads.
For Zonal endpoint (object-level) API operations except CopyObject and UploadPartCopy, \nyou authenticate and authorize requests through CreateSession for low latency. \n To encrypt new objects in a directory bucket with SSE-KMS, you must specify SSE-KMS as the directory bucket's default encryption configuration with a KMS key (specifically, a customer managed key). Then, when a session is created for Zonal endpoint API operations, new objects are automatically encrypted and decrypted with SSE-KMS and S3 Bucket Keys during the session.
\n\n Only 1 customer managed key is supported per directory bucket for the lifetime of the bucket. Amazon Web Services managed key (aws/s3
) isn't supported. \n After you specify SSE-KMS as your bucket's default encryption configuration with a customer managed key, you can't change the customer managed key for the bucket's SSE-KMS configuration.\n
In the Zonal endpoint API calls (except CopyObject and UploadPartCopy) using the REST API, \n you can't override the values of the encryption settings (x-amz-server-side-encryption
, x-amz-server-side-encryption-aws-kms-key-id
, x-amz-server-side-encryption-context
, and x-amz-server-side-encryption-bucket-key-enabled
) from the CreateSession
request. \n You don't need to explicitly specify these encryption settings values in Zonal endpoint API calls, and \n Amazon S3 will use the encryption settings values from the CreateSession
request to protect new objects in the directory bucket. \n
When you use the CLI or the Amazon Web Services SDKs, for CreateSession
, the session token refreshes automatically to avoid service interruptions when a session expires. The CLI or the Amazon Web Services SDKs use the bucket's default encryption configuration for the \n CreateSession
request. It's not supported to override the encryption settings values in the CreateSession
request. \n Also, in the Zonal endpoint API calls (except CopyObject and UploadPartCopy), \n it's not supported to override the values of the encryption settings from the CreateSession
request. \n\n
\n Directory buckets - The HTTP Host header syntax is \n Bucket_name.s3express-az_id.region.amazonaws.com
.
The server-side encryption algorithm used when you store objects in the directory bucket.
", + "smithy.api#httpHeader": "x-amz-server-side-encryption" + } + }, + "SSEKMSKeyId": { + "target": "com.amazonaws.s3#SSEKMSKeyId", + "traits": { + "smithy.api#documentation": "If you specify x-amz-server-side-encryption
with aws:kms
, this header indicates the ID of the KMS \n symmetric encryption customer managed key that was used for object encryption.
If present, indicates the Amazon Web Services KMS Encryption Context to use for object encryption. The value of\n this header is a Base64-encoded string of a UTF-8 encoded JSON, which contains the encryption context as key-value pairs. \n This value is stored as object metadata and automatically gets\n passed on to Amazon Web Services KMS for future GetObject
\n operations on this object.
Indicates whether to use an S3 Bucket Key for server-side encryption\n with KMS keys (SSE-KMS).
", + "smithy.api#httpHeader": "x-amz-server-side-encryption-bucket-key-enabled" + } + }, "Credentials": { "target": "com.amazonaws.s3#SessionCredentials", "traits": { @@ -19829,7 +19857,7 @@ "SessionMode": { "target": "com.amazonaws.s3#SessionMode", "traits": { - "smithy.api#documentation": "Specifies the mode of the session that will be created, either ReadWrite
or\n ReadOnly
. By default, a ReadWrite
session is created. A\n ReadWrite
session is capable of executing all the Zonal endpoint APIs on a\n directory bucket. A ReadOnly
session is constrained to execute the following\n Zonal endpoint APIs: GetObject
, HeadObject
, ListObjectsV2
,\n GetObjectAttributes
, ListParts
, and\n ListMultipartUploads
.
Specifies the mode of the session that will be created, either ReadWrite
or\n ReadOnly
. By default, a ReadWrite
session is created. A\n ReadWrite
session is capable of executing all the Zonal endpoint API operations on a\n directory bucket. A ReadOnly
session is constrained to execute the following\n Zonal endpoint API operations: GetObject
, HeadObject
, ListObjectsV2
,\n GetObjectAttributes
, ListParts
, and\n ListMultipartUploads
.
The server-side encryption algorithm to use when you store objects in the directory bucket.
\nFor directory buckets, there are only two supported options for server-side encryption: server-side encryption with Amazon S3 managed keys (SSE-S3) (AES256
) and server-side encryption with KMS keys (SSE-KMS) (aws:kms
). By default, Amazon S3 encrypts data with SSE-S3. \n For more\n information, see Protecting data with server-side encryption in the Amazon S3 User Guide.
If you specify x-amz-server-side-encryption
with aws:kms
, you must specify the \n x-amz-server-side-encryption-aws-kms-key-id
header with the ID (Key ID or Key ARN) of the KMS \n symmetric encryption customer managed key to use. Otherwise, you get an HTTP 400 Bad Request
error. Only use the key ID or key ARN. The key alias format of the KMS key isn't supported. Also, if the KMS key doesn't exist in the same\n account that't issuing the command, you must use the full Key ARN not the Key ID.
Your SSE-KMS configuration can only support 1 customer managed key per directory bucket for the lifetime of the bucket. \nAmazon Web Services managed key (aws/s3
) isn't supported. \n
Specifies the Amazon Web Services KMS Encryption Context as an additional encryption context to use for object encryption. The value of\n this header is a Base64-encoded string of a UTF-8 encoded JSON, which contains the encryption context as key-value pairs. \n This value is stored as object metadata and automatically gets passed on\n to Amazon Web Services KMS for future GetObject
operations on\n this object.
\n General purpose buckets - This value must be explicitly added during CopyObject
operations if you want an additional encryption context for your object. For more information, see Encryption context in the Amazon S3 User Guide.
\n Directory buckets - You can optionally provide an explicit encryption context value. The value must match the default encryption context - the bucket Amazon Resource Name (ARN). An additional encryption context value is not supported.
", + "smithy.api#httpHeader": "x-amz-server-side-encryption-context" + } + }, + "BucketKeyEnabled": { + "target": "com.amazonaws.s3#BucketKeyEnabled", + "traits": { + "smithy.api#documentation": "Specifies whether Amazon S3 should use an S3 Bucket Key for object encryption with\n server-side encryption using KMS keys (SSE-KMS).
\nS3 Bucket Keys are always enabled for GET
and PUT
operations in a directory bucket and can’t be disabled. S3 Bucket Keys aren't supported, when you copy SSE-KMS encrypted objects from general purpose buckets \nto directory buckets, from directory buckets to general purpose buckets, or between directory buckets, through CopyObject, UploadPartCopy, the Copy operation in Batch Operations, or \n the import jobs. In this case, Amazon S3 makes a call to KMS every time a copy request is made for a KMS-encrypted object.
This operation is not supported by directory buckets.
\nThis implementation of the DELETE action resets the default encryption for the bucket as\n server-side encryption with Amazon S3 managed keys (SSE-S3). For information about the bucket\n default encryption feature, see Amazon S3 Bucket Default Encryption\n in the Amazon S3 User Guide.
\nTo use this operation, you must have permissions to perform the\n s3:PutEncryptionConfiguration
action. The bucket owner has this permission\n by default. The bucket owner can grant this permission to others. For more information\n about permissions, see Permissions Related to Bucket Subresource Operations and Managing\n Access Permissions to your Amazon S3 Resources in the\n Amazon S3 User Guide.
The following operations are related to DeleteBucketEncryption
:
\n PutBucketEncryption\n
\n\n GetBucketEncryption\n
\nThis implementation of the DELETE action resets the default encryption for the bucket as\n server-side encryption with Amazon S3 managed keys (SSE-S3).
\n\n General purpose buckets - For information about the bucket default encryption feature, see Amazon S3 Bucket\n Default Encryption in the Amazon S3 User Guide.
\n\n Directory buckets - For directory buckets, there are only two supported options for server-side encryption: SSE-S3 and SSE-KMS. For information about the default encryption configuration in directory buckets, see Setting default server-side encryption behavior \n for directory buckets.
\n\n General purpose bucket permissions - The s3:PutEncryptionConfiguration
permission is required in a policy. \n The bucket owner has this permission\n by default. The bucket owner can grant this permission to others. For more information\n about permissions, see Permissions Related to Bucket Operations and Managing\n Access Permissions to Your Amazon S3 Resources.
\n Directory bucket permissions - To grant access to this API operation, you must have the s3express:PutEncryptionConfiguration
permission in an IAM identity-based policy instead of a bucket policy. Cross-account access to this API operation isn't supported. This operation can only be performed by the Amazon Web Services account that owns the resource. For more information about directory bucket policies and permissions, see Amazon Web Services Identity and Access Management (IAM) for S3 Express One Zone in the Amazon S3 User Guide.
\n Directory buckets - The HTTP Host header syntax is s3express-control.region.amazonaws.com
.
The following operations are related to DeleteBucketEncryption
:
\n PutBucketEncryption\n
\n\n GetBucketEncryption\n
\nThe name of the bucket containing the server-side encryption configuration to\n delete.
", + "smithy.api#documentation": "The name of the bucket containing the server-side encryption configuration to\n delete.
\n\n Directory buckets - When you use this operation with a directory bucket, you must use path-style requests in the format https://s3express-control.region_code.amazonaws.com/bucket-name\n
. Virtual-hosted-style requests aren't supported. Directory bucket names must be unique in the chosen Availability Zone. Bucket names must also follow the format \n bucket_base_name--az_id--x-s3
(for example, \n DOC-EXAMPLE-BUCKET--usw2-az1--x-s3
). For information about bucket naming restrictions, see Directory bucket naming rules in the Amazon S3 User Guide\n
The account ID of the expected bucket owner. If the account ID that you provide does not match the actual owner of the bucket, the request fails with the HTTP status code 403 Forbidden
(access denied).
The account ID of the expected bucket owner. If the account ID that you provide does not match the actual owner of the bucket, the request fails with the HTTP status code 403 Forbidden
(access denied).
For directory buckets, this header is not supported in this API operation. If you specify this header, the request fails with the HTTP status code \n501 Not Implemented
.
Optional configuration to replicate existing source bucket objects. For more\n information, see Replicating Existing Objects in the Amazon S3 User Guide.\n
" + "smithy.api#documentation": "Optional configuration to replicate existing source bucket objects. \n
\nThis parameter is no longer supported. To replicate existing objects, see Replicating existing objects with S3 Batch Replication in the Amazon S3 User Guide.
\nThis operation is not supported by directory buckets.
\nReturns the default encryption configuration for an Amazon S3 bucket. By default, all buckets\n have a default encryption configuration that uses server-side encryption with Amazon S3 managed\n keys (SSE-S3). For information about the bucket default encryption feature, see Amazon S3 Bucket\n Default Encryption in the Amazon S3 User Guide.
\nTo use this operation, you must have permission to perform the\n s3:GetEncryptionConfiguration
action. The bucket owner has this permission\n by default. The bucket owner can grant this permission to others. For more information\n about permissions, see Permissions Related to Bucket Subresource Operations and Managing\n Access Permissions to Your Amazon S3 Resources.
The following operations are related to GetBucketEncryption
:
\n PutBucketEncryption\n
\nReturns the default encryption configuration for an Amazon S3 bucket. By default, all buckets\n have a default encryption configuration that uses server-side encryption with Amazon S3 managed\n keys (SSE-S3).
\n\n General purpose buckets - For information about the bucket default encryption feature, see Amazon S3 Bucket\n Default Encryption in the Amazon S3 User Guide.
\n\n Directory buckets - For directory buckets, there are only two supported options for server-side encryption: SSE-S3 and SSE-KMS. For information about the default encryption configuration in directory buckets, see Setting default server-side encryption behavior \n for directory buckets.
\n\n General purpose bucket permissions - The s3:GetEncryptionConfiguration
permission is required in a policy. \n The bucket owner has this permission\n by default. The bucket owner can grant this permission to others. For more information\n about permissions, see Permissions Related to Bucket Operations and Managing\n Access Permissions to Your Amazon S3 Resources.
\n Directory bucket permissions - To grant access to this API operation, you must have the s3express:GetEncryptionConfiguration
permission in an IAM identity-based policy instead of a bucket policy. Cross-account access to this API operation isn't supported. This operation can only be performed by the Amazon Web Services account that owns the resource. For more information about directory bucket policies and permissions, see Amazon Web Services Identity and Access Management (IAM) for S3 Express One Zone in the Amazon S3 User Guide.
\n Directory buckets - The HTTP Host header syntax is s3express-control.region.amazonaws.com
.
The following operations are related to GetBucketEncryption
:
\n PutBucketEncryption\n
\nThe name of the bucket from which the server-side encryption configuration is\n retrieved.
", + "smithy.api#documentation": "The name of the bucket from which the server-side encryption configuration is\n retrieved.
\n\n Directory buckets - When you use this operation with a directory bucket, you must use path-style requests in the format https://s3express-control.region_code.amazonaws.com/bucket-name\n
. Virtual-hosted-style requests aren't supported. Directory bucket names must be unique in the chosen Availability Zone. Bucket names must also follow the format \n bucket_base_name--az_id--x-s3
(for example, \n DOC-EXAMPLE-BUCKET--usw2-az1--x-s3
). For information about bucket naming restrictions, see Directory bucket naming rules in the Amazon S3 User Guide\n
The account ID of the expected bucket owner. If the account ID that you provide does not match the actual owner of the bucket, the request fails with the HTTP status code 403 Forbidden
(access denied).
The account ID of the expected bucket owner. If the account ID that you provide does not match the actual owner of the bucket, the request fails with the HTTP status code 403 Forbidden
(access denied).
For directory buckets, this header is not supported in this API operation. If you specify this header, the request fails with the HTTP status code \n501 Not Implemented
.
Retrieves an object from Amazon S3.
\nIn the GetObject
request, specify the full key name for the object.
\n General purpose buckets - Both the virtual-hosted-style requests and the path-style requests are supported. For a virtual hosted-style request example, if you have\n the object photos/2006/February/sample.jpg
, specify the object key name as\n /photos/2006/February/sample.jpg
. For a path-style request example, if you\n have the object photos/2006/February/sample.jpg
in the bucket named\n examplebucket
, specify the object key name as\n /examplebucket/photos/2006/February/sample.jpg
. For more information about\n request types, see HTTP Host\n Header Bucket Specification in the Amazon S3 User Guide.
\n Directory buckets - Only virtual-hosted-style requests are supported. For a virtual hosted-style request example, if you have the object photos/2006/February/sample.jpg
in the bucket named examplebucket--use1-az5--x-s3
, specify the object key name as /photos/2006/February/sample.jpg
. Also, when you make requests to this API operation, your requests are sent to the Zonal endpoint. These endpoints support virtual-hosted-style requests in the format https://bucket_name.s3express-az_id.region.amazonaws.com/key-name\n
. Path-style requests are not supported. For more information, see Regional and Zonal endpoints in the\n Amazon S3 User Guide.
\n General purpose bucket permissions - You must have the required permissions in a policy. To use GetObject
, you must have the READ
\n access to the object (or version). If you grant READ
access to the anonymous user, the GetObject
operation \n returns the object without using an authorization header. For more information, see Specifying permissions in\n a policy in the Amazon S3 User Guide.
If you include a versionId
in your request header, you must have the\n s3:GetObjectVersion
permission to access a specific\n version of an object. The s3:GetObject
permission is not required in this scenario.
If you request the\n current version of an object without a specific versionId
in the request header, only\n the s3:GetObject
permission is required. The s3:GetObjectVersion
permission is not required in this scenario.\n
If the object that you request doesn’t exist, the error that\n Amazon S3 returns depends on whether you also have the s3:ListBucket
\n permission.
If you have the s3:ListBucket
permission on the bucket, Amazon S3\n returns an HTTP status code 404 Not Found
error.
If you don’t have the s3:ListBucket
permission, Amazon S3 returns an\n HTTP status code 403 Access Denied
error.
\n Directory bucket permissions - To grant access to this API operation on a directory bucket, we recommend that you use the \n CreateSession
\n API operation for session-based authorization. Specifically, you grant the s3express:CreateSession
permission to the directory bucket in a bucket policy or an IAM identity-based policy. Then, you make the CreateSession
API call on the bucket to obtain a session token. With the session token in your request header, you can make API requests to this operation. After the session token expires, you make another CreateSession
API call to generate a new session token for use. \nAmazon Web Services CLI or SDKs create session and refresh the session token automatically to avoid service interruptions when a session expires. For more information about authorization, see \n CreateSession
\n .
If the object you are retrieving is stored in the S3 Glacier Flexible Retrieval storage class, the \n S3 Glacier Deep Archive storage class, the S3 Intelligent-Tiering Archive Access tier, or the \n S3 Intelligent-Tiering Deep Archive Access tier, before you can retrieve the object you must first restore a\n copy using RestoreObject. Otherwise, this operation returns an\n InvalidObjectState
error. For information about restoring archived objects,\n see Restoring\n Archived Objects in the Amazon S3 User Guide.
\n Directory buckets - For directory buckets, only the S3 Express One Zone storage class is supported to store newly created objects. \nUnsupported storage class values won't write a destination object and will respond with the HTTP status code 400 Bad Request
.
Encryption request headers, like x-amz-server-side-encryption
, should not\n be sent for the GetObject
requests, if your object uses server-side encryption with Amazon S3 managed encryption keys (SSE-S3), server-side encryption with Key Management Service (KMS)\n keys (SSE-KMS), or dual-layer server-side encryption with Amazon Web Services KMS keys (DSSE-KMS). If you include the header in your GetObject
requests for the object that uses \n these types of keys, you’ll get an HTTP 400 Bad Request
error.
There are times when you want to override certain response header values of a\n GetObject
response. For example, you might override the\n Content-Disposition
response header value through your GetObject
\n request.
You can override values for a set of response headers. These modified response header values are included only in a successful response, that is, when the HTTP status code 200 OK
is returned. \n The headers you can override using the following query parameters in the request are a subset of the headers that Amazon S3 accepts when you create an object. \n
The response headers that you can override for the\n GetObject
response are Cache-Control
, Content-Disposition
, \n Content-Encoding
, Content-Language
, Content-Type
, and Expires
.
To override values for a set of response headers in the\n GetObject
response, you can use the following query\n parameters in the request.
\n response-cache-control
\n
\n response-content-disposition
\n
\n response-content-encoding
\n
\n response-content-language
\n
\n response-content-type
\n
\n response-expires
\n
When you use these parameters, you must sign the request by using either an Authorization header or a\n presigned URL. These parameters cannot be used with an\n unsigned (anonymous) request.
\n\n Directory buckets - The HTTP Host header syntax is \n Bucket_name.s3express-az_id.region.amazonaws.com
.
The following operations are related to GetObject
:
\n ListBuckets\n
\n\n GetObjectAcl\n
\nRetrieves an object from Amazon S3.
\nIn the GetObject
request, specify the full key name for the object.
\n General purpose buckets - Both the virtual-hosted-style requests and the path-style requests are supported. For a virtual hosted-style request example, if you have\n the object photos/2006/February/sample.jpg
, specify the object key name as\n /photos/2006/February/sample.jpg
. For a path-style request example, if you\n have the object photos/2006/February/sample.jpg
in the bucket named\n examplebucket
, specify the object key name as\n /examplebucket/photos/2006/February/sample.jpg
. For more information about\n request types, see HTTP Host\n Header Bucket Specification in the Amazon S3 User Guide.
\n Directory buckets - Only virtual-hosted-style requests are supported. For a virtual hosted-style request example, if you have the object photos/2006/February/sample.jpg
in the bucket named examplebucket--use1-az5--x-s3
, specify the object key name as /photos/2006/February/sample.jpg
. Also, when you make requests to this API operation, your requests are sent to the Zonal endpoint. These endpoints support virtual-hosted-style requests in the format https://bucket_name.s3express-az_id.region.amazonaws.com/key-name\n
. Path-style requests are not supported. For more information, see Regional and Zonal endpoints in the\n Amazon S3 User Guide.
\n General purpose bucket permissions - You must have the required permissions in a policy. To use GetObject
, you must have the READ
\n access to the object (or version). If you grant READ
access to the anonymous user, the GetObject
operation \n returns the object without using an authorization header. For more information, see Specifying permissions in\n a policy in the Amazon S3 User Guide.
If you include a versionId
in your request header, you must have the\n s3:GetObjectVersion
permission to access a specific\n version of an object. The s3:GetObject
permission is not required in this scenario.
If you request the\n current version of an object without a specific versionId
in the request header, only\n the s3:GetObject
permission is required. The s3:GetObjectVersion
permission is not required in this scenario.\n
If the object that you request doesn’t exist, the error that\n Amazon S3 returns depends on whether you also have the s3:ListBucket
\n permission.
If you have the s3:ListBucket
permission on the bucket, Amazon S3\n returns an HTTP status code 404 Not Found
error.
If you don’t have the s3:ListBucket
permission, Amazon S3 returns an\n HTTP status code 403 Access Denied
error.
\n Directory bucket permissions - To grant access to this API operation on a directory bucket, we recommend that you use the \n CreateSession
\n API operation for session-based authorization. Specifically, you grant the s3express:CreateSession
permission to the directory bucket in a bucket policy or an IAM identity-based policy. Then, you make the CreateSession
API call on the bucket to obtain a session token. With the session token in your request header, you can make API requests to this operation. After the session token expires, you make another CreateSession
API call to generate a new session token for use. \nAmazon Web Services CLI or SDKs create session and refresh the session token automatically to avoid service interruptions when a session expires. For more information about authorization, see \n CreateSession
\n .
If the object is encrypted using \n SSE-KMS, you must also have the\n kms:GenerateDataKey
and kms:Decrypt
permissions in IAM identity-based policies and KMS key policies for the KMS key.
If the object you are retrieving is stored in the S3 Glacier Flexible Retrieval storage class, the \n S3 Glacier Deep Archive storage class, the S3 Intelligent-Tiering Archive Access tier, or the \n S3 Intelligent-Tiering Deep Archive Access tier, before you can retrieve the object you must first restore a\n copy using RestoreObject. Otherwise, this operation returns an\n InvalidObjectState
error. For information about restoring archived objects,\n see Restoring\n Archived Objects in the Amazon S3 User Guide.
\n Directory buckets - For directory buckets, only the S3 Express One Zone storage class is supported to store newly created objects. \nUnsupported storage class values won't write a destination object and will respond with the HTTP status code 400 Bad Request
.
Encryption request headers, like x-amz-server-side-encryption
, should not\n be sent for the GetObject
requests, if your object uses server-side encryption with Amazon S3 managed encryption keys (SSE-S3), server-side encryption with Key Management Service (KMS)\n keys (SSE-KMS), or dual-layer server-side encryption with Amazon Web Services KMS keys (DSSE-KMS). If you include the header in your GetObject
requests for the object that uses \n these types of keys, you’ll get an HTTP 400 Bad Request
error.
\n Directory buckets - For directory buckets, there are only two supported options for server-side encryption: SSE-S3 and SSE-KMS. SSE-C isn't supported. For more\n information, see Protecting data with server-side encryption in the Amazon S3 User Guide.
\nThere are times when you want to override certain response header values of a\n GetObject
response. For example, you might override the\n Content-Disposition
response header value through your GetObject
\n request.
You can override values for a set of response headers. These modified response header values are included only in a successful response, that is, when the HTTP status code 200 OK
is returned. \n The headers you can override using the following query parameters in the request are a subset of the headers that Amazon S3 accepts when you create an object. \n
The response headers that you can override for the\n GetObject
response are Cache-Control
, Content-Disposition
, \n Content-Encoding
, Content-Language
, Content-Type
, and Expires
.
To override values for a set of response headers in the\n GetObject
response, you can use the following query\n parameters in the request.
\n response-cache-control
\n
\n response-content-disposition
\n
\n response-content-encoding
\n
\n response-content-language
\n
\n response-content-type
\n
\n response-expires
\n
When you use these parameters, you must sign the request by using either an Authorization header or a\n presigned URL. These parameters cannot be used with an\n unsigned (anonymous) request.
\n\n Directory buckets - The HTTP Host header syntax is \n Bucket_name.s3express-az_id.region.amazonaws.com
.
The following operations are related to GetObject
:
\n ListBuckets\n
\n\n GetObjectAcl\n
\nRetrieves all the metadata from an object without returning the object itself. This\n operation is useful if you're interested only in an object's metadata.
\n\n GetObjectAttributes
combines the functionality of HeadObject
\n and ListParts
. All of the data returned with each of those individual calls\n can be returned with a single call to GetObjectAttributes
.
\n Directory buckets - For directory buckets, you must make requests for this API operation to the Zonal endpoint. These endpoints support virtual-hosted-style requests in the format https://bucket_name.s3express-az_id.region.amazonaws.com/key-name\n
. Path-style requests are not supported. For more information, see Regional and Zonal endpoints in the\n Amazon S3 User Guide.
\n General purpose bucket permissions - To use\n GetObjectAttributes
, you must have READ access to the object. The permissions that you need to use this operation with depend on whether the\n bucket is versioned. If the bucket is versioned, you need both the\n s3:GetObjectVersion
and s3:GetObjectVersionAttributes
\n permissions for this operation. If the bucket is not versioned, you need the\n s3:GetObject
and s3:GetObjectAttributes
permissions.\n For more information, see Specifying Permissions in\n a Policy in the Amazon S3 User Guide. If the object\n that you request does not exist, the error Amazon S3 returns depends on whether you\n also have the s3:ListBucket
permission.
If you have the s3:ListBucket
permission on the bucket, Amazon S3\n returns an HTTP status code 404 Not Found
(\"no such key\")\n error.
If you don't have the s3:ListBucket
permission, Amazon S3 returns\n an HTTP status code 403 Forbidden
(\"access denied\")\n error.
\n Directory bucket permissions - To grant access to this API operation on a directory bucket, we recommend that you use the \n CreateSession
\n API operation for session-based authorization. Specifically, you grant the s3express:CreateSession
permission to the directory bucket in a bucket policy or an IAM identity-based policy. Then, you make the CreateSession
API call on the bucket to obtain a session token. With the session token in your request header, you can make API requests to this operation. After the session token expires, you make another CreateSession
API call to generate a new session token for use. \nAmazon Web Services CLI or SDKs create session and refresh the session token automatically to avoid service interruptions when a session expires. For more information about authorization, see \n CreateSession
\n .
Encryption request headers, like x-amz-server-side-encryption
,\n should not be sent for HEAD
requests if your object uses server-side\n encryption with Key Management Service (KMS) keys (SSE-KMS), dual-layer server-side\n encryption with Amazon Web Services KMS keys (DSSE-KMS), or server-side encryption with Amazon S3\n managed encryption keys (SSE-S3). The x-amz-server-side-encryption
header is used when you PUT
an object to S3 and want to specify the encryption method. \n If you include this header in a GET
request for an object that uses these types of keys, \n you’ll get an HTTP 400 Bad Request
error. It's because the encryption method can't be changed when you retrieve the object.
If you encrypt an object by using server-side encryption with customer-provided\n encryption keys (SSE-C) when you store the object in Amazon S3, then when you retrieve the\n metadata from the object, you must use the following headers to provide the encryption key for the server to be able to retrieve the object's metadata. The headers are:
\n\n x-amz-server-side-encryption-customer-algorithm
\n
\n x-amz-server-side-encryption-customer-key
\n
\n x-amz-server-side-encryption-customer-key-MD5
\n
For more information about SSE-C, see Server-Side Encryption\n (Using Customer-Provided Encryption Keys) in the Amazon S3\n User Guide.
\n\n Directory bucket permissions - For directory buckets, only server-side encryption with Amazon S3 managed keys (SSE-S3) (AES256
) is supported.
\n Directory buckets - S3 Versioning isn't enabled and supported for directory buckets. For this API operation, only the null
value of the version ID is supported by directory buckets. You can only specify null
\n to the versionId
query parameter in the request.
Consider the following when using request headers:
\nIf both of the If-Match
and If-Unmodified-Since
headers\n are present in the request as follows, then Amazon S3 returns the HTTP status code\n 200 OK
and the data requested:
\n If-Match
condition evaluates to true
.
\n If-Unmodified-Since
condition evaluates to\n false
.
For more information about conditional requests, see RFC 7232.
\nIf both of the If-None-Match
and If-Modified-Since
\n headers are present in the request as follows, then Amazon S3 returns the HTTP status code\n 304 Not Modified
:
\n If-None-Match
condition evaluates to false
.
\n If-Modified-Since
condition evaluates to\n true
.
For more information about conditional requests, see RFC 7232.
\n\n Directory buckets - The HTTP Host header syntax is \n Bucket_name.s3express-az_id.region.amazonaws.com
.
The following actions are related to GetObjectAttributes
:
\n GetObject\n
\n\n GetObjectAcl\n
\n\n GetObjectLegalHold\n
\n\n GetObjectRetention\n
\n\n GetObjectTagging\n
\n\n HeadObject\n
\n\n ListParts\n
\nRetrieves all the metadata from an object without returning the object itself. This\n operation is useful if you're interested only in an object's metadata.
\n\n GetObjectAttributes
combines the functionality of HeadObject
\n and ListParts
. All of the data returned with each of those individual calls\n can be returned with a single call to GetObjectAttributes
.
\n Directory buckets - For directory buckets, you must make requests for this API operation to the Zonal endpoint. These endpoints support virtual-hosted-style requests in the format https://bucket_name.s3express-az_id.region.amazonaws.com/key-name\n
. Path-style requests are not supported. For more information, see Regional and Zonal endpoints in the\n Amazon S3 User Guide.
\n General purpose bucket permissions - To use\n GetObjectAttributes
, you must have READ access to the object. The permissions that you need to use this operation depend on whether the\n bucket is versioned. If the bucket is versioned, you need both the\n s3:GetObjectVersion
and s3:GetObjectVersionAttributes
\n permissions for this operation. If the bucket is not versioned, you need the\n s3:GetObject
and s3:GetObjectAttributes
permissions.\n For more information, see Specifying Permissions in\n a Policy in the Amazon S3 User Guide. If the object\n that you request does not exist, the error Amazon S3 returns depends on whether you\n also have the s3:ListBucket
permission.
If you have the s3:ListBucket
permission on the bucket, Amazon S3\n returns an HTTP status code 404 Not Found
(\"no such key\")\n error.
If you don't have the s3:ListBucket
permission, Amazon S3 returns\n an HTTP status code 403 Forbidden
(\"access denied\")\n error.
\n Directory bucket permissions - To grant access to this API operation on a directory bucket, we recommend that you use the \n CreateSession
\n API operation for session-based authorization. Specifically, you grant the s3express:CreateSession
permission to the directory bucket in a bucket policy or an IAM identity-based policy. Then, you make the CreateSession
API call on the bucket to obtain a session token. With the session token in your request header, you can make API requests to this operation. After the session token expires, you make another CreateSession
API call to generate a new session token for use. \nAmazon Web Services CLI or SDKs create session and refresh the session token automatically to avoid service interruptions when a session expires. For more information about authorization, see \n CreateSession
\n .
If the object is encrypted with\n SSE-KMS, you must also have the\n kms:GenerateDataKey
and kms:Decrypt
permissions in IAM identity-based policies and KMS key policies for the KMS key.
Encryption request headers, like x-amz-server-side-encryption
,\n should not be sent for HEAD
requests if your object uses server-side\n encryption with Key Management Service (KMS) keys (SSE-KMS), dual-layer server-side\n encryption with Amazon Web Services KMS keys (DSSE-KMS), or server-side encryption with Amazon S3\n managed encryption keys (SSE-S3). The x-amz-server-side-encryption
header is used when you PUT
an object to S3 and want to specify the encryption method. \n If you include this header in a GET
request for an object that uses these types of keys, \n you’ll get an HTTP 400 Bad Request
error. It's because the encryption method can't be changed when you retrieve the object.
If you encrypt an object by using server-side encryption with customer-provided\n encryption keys (SSE-C) when you store the object in Amazon S3, then when you retrieve the\n metadata from the object, you must use the following headers to provide the encryption key for the server to be able to retrieve the object's metadata. The headers are:
\n\n x-amz-server-side-encryption-customer-algorithm
\n
\n x-amz-server-side-encryption-customer-key
\n
\n x-amz-server-side-encryption-customer-key-MD5
\n
For more information about SSE-C, see Server-Side Encryption\n (Using Customer-Provided Encryption Keys) in the Amazon S3\n User Guide.
\n\n Directory bucket permissions - For directory buckets, there are only two supported options for server-side encryption: server-side encryption with Amazon S3 managed keys (SSE-S3) (AES256
) and server-side encryption with KMS keys (SSE-KMS) (aws:kms
). We recommend that the bucket's default encryption uses the desired encryption configuration and you don't override the bucket default encryption in your \n CreateSession
requests or PUT
object requests. Then, new objects \n are automatically encrypted with the desired encryption settings. For more\n information, see Protecting data with server-side encryption in the Amazon S3 User Guide. For more information about the encryption overriding behaviors in directory buckets, see Specifying server-side encryption with KMS for new object uploads.
\n Directory buckets - S3 Versioning isn't enabled and supported for directory buckets. For this API operation, only the null
value of the version ID is supported by directory buckets. You can only specify null
\n to the versionId
query parameter in the request.
Consider the following when using request headers:
\nIf both of the If-Match
and If-Unmodified-Since
headers\n are present in the request as follows, then Amazon S3 returns the HTTP status code\n 200 OK
and the data requested:
\n If-Match
condition evaluates to true
.
\n If-Unmodified-Since
condition evaluates to\n false
.
For more information about conditional requests, see RFC 7232.
\nIf both of the If-None-Match
and If-Modified-Since
\n headers are present in the request as follows, then Amazon S3 returns the HTTP status code\n 304 Not Modified
:
\n If-None-Match
condition evaluates to false
.
\n If-Modified-Since
condition evaluates to\n true
.
For more information about conditional requests, see RFC 7232.
\n\n Directory buckets - The HTTP Host header syntax is \n Bucket_name.s3express-az_id.region.amazonaws.com
.
The following actions are related to GetObjectAttributes
:
\n GetObject\n
\n\n GetObjectAcl\n
\n\n GetObjectLegalHold\n
\n\n GetObjectRetention\n
\n\n GetObjectTagging\n
\n\n HeadObject\n
\n\n ListParts\n
\nThe server-side encryption algorithm used when you store this object in Amazon S3 (for example,\n AES256
, aws:kms
, aws:kms:dsse
).
For directory buckets, only server-side encryption with Amazon S3 managed keys (SSE-S3) (AES256
) is supported.
The server-side encryption algorithm used when you store this object in Amazon S3.
", "smithy.api#httpHeader": "x-amz-server-side-encryption" } }, @@ -23930,14 +23986,14 @@ "SSEKMSKeyId": { "target": "com.amazonaws.s3#SSEKMSKeyId", "traits": { - "smithy.api#documentation": "If present, indicates the ID of the Key Management Service (KMS) symmetric encryption customer managed key\n that was used for the object.
\nThis functionality is not supported for directory buckets.
\nIf present, indicates the ID of the KMS key that was used for object encryption.
", "smithy.api#httpHeader": "x-amz-server-side-encryption-aws-kms-key-id" } }, "BucketKeyEnabled": { "target": "com.amazonaws.s3#BucketKeyEnabled", "traits": { - "smithy.api#documentation": "Indicates whether the object uses an S3 Bucket Key for server-side encryption\n with Key Management Service (KMS) keys (SSE-KMS).
\nThis functionality is not supported for directory buckets.
\nIndicates whether the object uses an S3 Bucket Key for server-side encryption\n with Key Management Service (KMS) keys (SSE-KMS).
", "smithy.api#httpHeader": "x-amz-server-side-encryption-bucket-key-enabled" } }, @@ -24154,7 +24210,7 @@ "ChecksumMode": { "target": "com.amazonaws.s3#ChecksumMode", "traits": { - "smithy.api#documentation": "To retrieve the checksum, this mode must be enabled.
\nIn addition, if you enable checksum mode and the object is uploaded with a \n checksum \n and encrypted with an Key Management Service (KMS) key, you must have permission to use the \n kms:Decrypt
action to retrieve the checksum.
To retrieve the checksum, this mode must be enabled.
\n\n General purpose buckets - In addition, if you enable checksum mode and the object is uploaded with a \n checksum \n and encrypted with an Key Management Service (KMS) key, you must have permission to use the \n kms:Decrypt
action to retrieve the checksum.
The HEAD
operation retrieves metadata from an object without returning the\n object itself. This operation is useful if you're interested only in an object's metadata.
A HEAD
request has the same options as a GET
operation on an\n object. The response is identical to the GET
response except that there is no\n response body. Because of this, if the HEAD
request generates an error, it\n returns a generic code, such as 400 Bad Request
, 403 Forbidden
, 404 Not\n Found
, 405 Method Not Allowed
, 412 Precondition Failed
, or 304 Not Modified
. \n It's not possible to retrieve the exact exception of these error codes.
Request headers are limited to 8 KB in size. For more information, see Common\n Request Headers.
\n\n General purpose bucket permissions - To\n use HEAD
, you must have the s3:GetObject
permission. You need the relevant read object (or version) permission for this operation.\n For more information, see Actions, resources, and condition\n keys for Amazon S3 in the Amazon S3\n User Guide.
If the object you request doesn't exist, the error that\n Amazon S3 returns depends on whether you also have the s3:ListBucket
permission.
If you have the s3:ListBucket
permission on the bucket, Amazon S3\n returns an HTTP status code 404 Not Found
error.
If you don’t have the s3:ListBucket
permission, Amazon S3 returns\n an HTTP status code 403 Forbidden
error.
\n Directory bucket permissions - To grant access to this API operation on a directory bucket, we recommend that you use the \n CreateSession
\n API operation for session-based authorization. Specifically, you grant the s3express:CreateSession
permission to the directory bucket in a bucket policy or an IAM identity-based policy. Then, you make the CreateSession
API call on the bucket to obtain a session token. With the session token in your request header, you can make API requests to this operation. After the session token expires, you make another CreateSession
API call to generate a new session token for use. \nAmazon Web Services CLI or SDKs create session and refresh the session token automatically to avoid service interruptions when a session expires. For more information about authorization, see \n CreateSession
\n .
Encryption request headers, like x-amz-server-side-encryption
,\n should not be sent for HEAD
requests if your object uses server-side\n encryption with Key Management Service (KMS) keys (SSE-KMS), dual-layer server-side\n encryption with Amazon Web Services KMS keys (DSSE-KMS), or server-side encryption with Amazon S3\n managed encryption keys (SSE-S3). The x-amz-server-side-encryption
header is used when you PUT
an object to S3 and want to specify the encryption method. \n If you include this header in a HEAD
request for an object that uses these types of keys, \n you’ll get an HTTP 400 Bad Request
error. It's because the encryption method can't be changed when you retrieve the object.
If you encrypt an object by using server-side encryption with customer-provided\n encryption keys (SSE-C) when you store the object in Amazon S3, then when you retrieve the\n metadata from the object, you must use the following headers to provide the encryption key for the server to be able to retrieve the object's metadata. The headers are:
\n\n x-amz-server-side-encryption-customer-algorithm
\n
\n x-amz-server-side-encryption-customer-key
\n
\n x-amz-server-side-encryption-customer-key-MD5
\n
For more information about SSE-C, see Server-Side Encryption\n (Using Customer-Provided Encryption Keys) in the Amazon S3\n User Guide.
\n\n Directory bucket permissions - For directory buckets, only server-side encryption with Amazon S3 managed keys (SSE-S3) (AES256
) is supported.
If the current version of the object is a delete marker, Amazon S3 behaves as if the object was deleted and includes x-amz-delete-marker: true
in the response.
If the specified version is a delete marker, the response returns a 405 Method Not Allowed
error and the Last-Modified: timestamp
response header.
\n Directory buckets - Delete marker is not supported by directory buckets.
\n\n Directory buckets - S3 Versioning isn't enabled and supported for directory buckets. For this API operation, only the null
value of the version ID is supported by directory buckets. You can only specify null
\n to the versionId
query parameter in the request.
\n Directory buckets - The HTTP Host header syntax is \n Bucket_name.s3express-az_id.region.amazonaws.com
.
For directory buckets, you must make requests for this API operation to the Zonal endpoint. These endpoints support virtual-hosted-style requests in the format https://bucket_name.s3express-az_id.region.amazonaws.com/key-name\n
. Path-style requests are not supported. For more information, see Regional and Zonal endpoints in the\n Amazon S3 User Guide.
The following actions are related to HeadObject
:
\n GetObject\n
\n\n GetObjectAttributes\n
\nThe HEAD
operation retrieves metadata from an object without returning the\n object itself. This operation is useful if you're interested only in an object's metadata.
A HEAD
request has the same options as a GET
operation on an\n object. The response is identical to the GET
response except that there is no\n response body. Because of this, if the HEAD
request generates an error, it\n returns a generic code, such as 400 Bad Request
, 403 Forbidden
, 404 Not\n Found
, 405 Method Not Allowed
, 412 Precondition Failed
, or 304 Not Modified
. \n It's not possible to retrieve the exact exception of these error codes.
Request headers are limited to 8 KB in size. For more information, see Common\n Request Headers.
\n\n General purpose bucket permissions - To\n use HEAD
, you must have the s3:GetObject
permission. You need the relevant read object (or version) permission for this operation.\n For more information, see Actions, resources, and condition\n keys for Amazon S3 in the Amazon S3\n User Guide.
If the object you request doesn't exist, the error that\n Amazon S3 returns depends on whether you also have the s3:ListBucket
permission.
If you have the s3:ListBucket
permission on the bucket, Amazon S3\n returns an HTTP status code 404 Not Found
error.
If you don’t have the s3:ListBucket
permission, Amazon S3 returns\n an HTTP status code 403 Forbidden
error.
\n Directory bucket permissions - To grant access to this API operation on a directory bucket, we recommend that you use the \n CreateSession
\n API operation for session-based authorization. Specifically, you grant the s3express:CreateSession
permission to the directory bucket in a bucket policy or an IAM identity-based policy. Then, you make the CreateSession
API call on the bucket to obtain a session token. With the session token in your request header, you can make API requests to this operation. After the session token expires, you make another CreateSession
API call to generate a new session token for use. \nAmazon Web Services CLI or SDKs create session and refresh the session token automatically to avoid service interruptions when a session expires. For more information about authorization, see \n CreateSession
\n .
If you enable x-amz-checksum-mode
in the request and the object is encrypted with\n Amazon Web Services Key Management Service (Amazon Web Services KMS), you must also have the\n kms:GenerateDataKey
and kms:Decrypt
permissions in IAM identity-based policies and KMS key policies for the KMS key to retrieve the checksum of the object.
Encryption request headers, like x-amz-server-side-encryption
,\n should not be sent for HEAD
requests if your object uses server-side\n encryption with Key Management Service (KMS) keys (SSE-KMS), dual-layer server-side\n encryption with Amazon Web Services KMS keys (DSSE-KMS), or server-side encryption with Amazon S3\n managed encryption keys (SSE-S3). The x-amz-server-side-encryption
header is used when you PUT
an object to S3 and want to specify the encryption method. \n If you include this header in a HEAD
request for an object that uses these types of keys, \n you’ll get an HTTP 400 Bad Request
error. It's because the encryption method can't be changed when you retrieve the object.
If you encrypt an object by using server-side encryption with customer-provided\n encryption keys (SSE-C) when you store the object in Amazon S3, then when you retrieve the\n metadata from the object, you must use the following headers to provide the encryption key for the server to be able to retrieve the object's metadata. The headers are:
\n\n x-amz-server-side-encryption-customer-algorithm
\n
\n x-amz-server-side-encryption-customer-key
\n
\n x-amz-server-side-encryption-customer-key-MD5
\n
For more information about SSE-C, see Server-Side Encryption\n (Using Customer-Provided Encryption Keys) in the Amazon S3\n User Guide.
\n\n Directory bucket - For directory buckets, there are only two supported options for server-side encryption: SSE-S3 and SSE-KMS. SSE-C isn't supported. For more\n information, see Protecting data with server-side encryption in the Amazon S3 User Guide.
\nIf the current version of the object is a delete marker, Amazon S3 behaves as if the object was deleted and includes x-amz-delete-marker: true
in the response.
If the specified version is a delete marker, the response returns a 405 Method Not Allowed
error and the Last-Modified: timestamp
response header.
\n Directory buckets - Delete marker is not supported by directory buckets.
\n\n Directory buckets - S3 Versioning isn't enabled and supported for directory buckets. For this API operation, only the null
value of the version ID is supported by directory buckets. You can only specify null
\n to the versionId
query parameter in the request.
\n Directory buckets - The HTTP Host header syntax is \n Bucket_name.s3express-az_id.region.amazonaws.com
.
For directory buckets, you must make requests for this API operation to the Zonal endpoint. These endpoints support virtual-hosted-style requests in the format https://bucket_name.s3express-az_id.region.amazonaws.com/key-name\n
. Path-style requests are not supported. For more information, see Regional and Zonal endpoints in the\n Amazon S3 User Guide.
The following actions are related to HeadObject
:
\n GetObject\n
\n\n GetObjectAttributes\n
\nThe server-side encryption algorithm used when you store this object in Amazon S3 (for example,\n AES256
, aws:kms
, aws:kms:dsse
).
For directory buckets, only server-side encryption with Amazon S3 managed keys (SSE-S3) (AES256
) is supported.
The server-side encryption algorithm used when you store this object in Amazon S3 (for example,\n AES256
, aws:kms
, aws:kms:dsse
).
If present, indicates the ID of the Key Management Service (KMS) symmetric encryption customer managed key\n that was used for the object.
\nThis functionality is not supported for directory buckets.
\nIf present, indicates the ID of the KMS key that was used for object encryption.
", "smithy.api#httpHeader": "x-amz-server-side-encryption-aws-kms-key-id" } }, "BucketKeyEnabled": { "target": "com.amazonaws.s3#BucketKeyEnabled", "traits": { - "smithy.api#documentation": "Indicates whether the object uses an S3 Bucket Key for server-side encryption\n with Key Management Service (KMS) keys (SSE-KMS).
\nThis functionality is not supported for directory buckets.
\nIndicates whether the object uses an S3 Bucket Key for server-side encryption\n with Key Management Service (KMS) keys (SSE-KMS).
", "smithy.api#httpHeader": "x-amz-server-side-encryption-bucket-key-enabled" } }, @@ -25204,7 +25260,7 @@ "ChecksumMode": { "target": "com.amazonaws.s3#ChecksumMode", "traits": { - "smithy.api#documentation": "To retrieve the checksum, this parameter must be enabled.
\nIn addition, if you enable checksum mode and the object is uploaded with a \n checksum \n and encrypted with an Key Management Service (KMS) key, you must have permission to use the \n kms:Decrypt
action to retrieve the checksum.
To retrieve the checksum, this parameter must be enabled.
\n\n General purpose buckets - If you enable checksum mode and the object is uploaded with a \n checksum \n and encrypted with an Key Management Service (KMS) key, you must have permission to use the \n kms:Decrypt
action to retrieve the checksum.
\n Directory buckets - If you enable ChecksumMode
and the object is encrypted with\n Amazon Web Services Key Management Service (Amazon Web Services KMS), you must also have the\n kms:GenerateDataKey
and kms:Decrypt
permissions in IAM identity-based policies and KMS key policies for the KMS key to retrieve the checksum of the object.
This operation is not supported by directory buckets.
\nThis action uses the encryption
subresource to configure default encryption\n and Amazon S3 Bucket Keys for an existing bucket.
By default, all buckets have a default encryption configuration that uses server-side\n encryption with Amazon S3 managed keys (SSE-S3). You can optionally configure default encryption\n for a bucket by using server-side encryption with Key Management Service (KMS) keys (SSE-KMS) or\n dual-layer server-side encryption with Amazon Web Services KMS keys (DSSE-KMS). If you specify default encryption by using\n SSE-KMS, you can also configure Amazon S3 Bucket\n Keys. If you use PutBucketEncryption to set your default bucket encryption to SSE-KMS, you should verify that your KMS key ID is correct. Amazon S3 does not validate the KMS key ID provided in PutBucketEncryption requests.
\nIf you're specifying a customer managed KMS key, we recommend using a fully qualified\n KMS key ARN. If you use a KMS key alias instead, then KMS resolves the key within the\n requester’s account. This behavior can result in data that's encrypted with a KMS key\n that belongs to the requester, and not the bucket owner.
\nAlso, this action requires Amazon Web Services Signature Version 4. For more information, see \n Authenticating Requests (Amazon Web Services Signature Version 4).
\nTo use this operation, you must have permission to perform the\n s3:PutEncryptionConfiguration
action. The bucket owner has this permission\n by default. The bucket owner can grant this permission to others. For more information\n about permissions, see Permissions Related to Bucket Subresource Operations and Managing\n Access Permissions to Your Amazon S3 Resources in the\n Amazon S3 User Guide.
The following operations are related to PutBucketEncryption
:
\n GetBucketEncryption\n
\nThis operation configures default encryption \n and Amazon S3 Bucket Keys for an existing bucket.
\n\n Directory buckets - For directory buckets, you must make requests for this API operation to the Regional endpoint. These endpoints support path-style requests in the format https://s3express-control.region_code.amazonaws.com/bucket-name\n
. Virtual-hosted-style requests aren't supported. \nFor more information, see Regional and Zonal endpoints in the\n Amazon S3 User Guide.
By default, all buckets have a default encryption configuration that uses server-side\n encryption with Amazon S3 managed keys (SSE-S3).
\n\n General purpose buckets\n
\nYou can optionally configure default encryption\n for a bucket by using server-side encryption with Key Management Service (KMS) keys (SSE-KMS) or\n dual-layer server-side encryption with Amazon Web Services KMS keys (DSSE-KMS). \n If you specify default encryption by using\n SSE-KMS, you can also configure Amazon S3 Bucket\n Keys. For information about the bucket default\n encryption feature, see Amazon S3 Bucket Default Encryption\n in the Amazon S3 User Guide.\n
\nIf you use PutBucketEncryption to set your default bucket encryption to SSE-KMS, you should verify that your KMS key ID is correct. Amazon S3 doesn't validate the KMS key ID provided in PutBucketEncryption requests.
\n\n Directory buckets - You can optionally configure default encryption\n for a bucket by using server-side encryption with Key Management Service (KMS) keys (SSE-KMS).
\nWe recommend that the bucket's default encryption uses the desired encryption configuration and you don't override the bucket default encryption in your \n CreateSession
requests or PUT
object requests. Then, new objects \n are automatically encrypted with the desired encryption settings. For more information about the encryption overriding behaviors in directory buckets, see Specifying server-side encryption with KMS for new object uploads.
Your SSE-KMS configuration can only support 1 customer managed key per directory bucket for the lifetime of the bucket. \nAmazon Web Services managed key (aws/s3
) isn't supported. \n
S3 Bucket Keys are always enabled for GET
and PUT
operations in a directory bucket and can’t be disabled. S3 Bucket Keys aren't supported, when you copy SSE-KMS encrypted objects from general purpose buckets \nto directory buckets, from directory buckets to general purpose buckets, or between directory buckets, through CopyObject, UploadPartCopy, the Copy operation in Batch Operations, or \n the import jobs. In this case, Amazon S3 makes a call to KMS every time a copy request is made for a KMS-encrypted object.
When you specify an KMS customer managed key for encryption in your directory bucket, only use the key ID or key ARN. The key alias format of the KMS key isn't supported.
\nFor directory buckets, if you use PutBucketEncryption to set your default bucket encryption to SSE-KMS, Amazon S3 validates the KMS key ID provided in PutBucketEncryption requests.
\nIf you're specifying a customer managed KMS key, we recommend using a fully qualified\n KMS key ARN. If you use a KMS key alias instead, then KMS resolves the key within the\n requester’s account. This behavior can result in data that's encrypted with a KMS key\n that belongs to the requester, and not the bucket owner.
\nAlso, this action requires Amazon Web Services Signature Version 4. For more information, see \n Authenticating Requests (Amazon Web Services Signature Version 4).
\n\n General purpose bucket permissions - The s3:PutEncryptionConfiguration
permission is required in a policy. \n The bucket owner has this permission\n by default. The bucket owner can grant this permission to others. For more information\n about permissions, see Permissions Related to Bucket Operations and Managing\n Access Permissions to Your Amazon S3 Resources in the\n Amazon S3 User Guide.
\n Directory bucket permissions - To grant access to this API operation, you must have the s3express:PutEncryptionConfiguration
permission in an IAM identity-based policy instead of a bucket policy. Cross-account access to this API operation isn't supported. This operation can only be performed by the Amazon Web Services account that owns the resource. For more information about directory bucket policies and permissions, see Amazon Web Services Identity and Access Management (IAM) for S3 Express One Zone in the Amazon S3 User Guide.
To set a directory bucket default encryption with SSE-KMS, you must also have the kms:GenerateDataKey
and the kms:Decrypt
permissions in IAM identity-based policies and KMS key policies for the target KMS key.
\n Directory buckets - The HTTP Host header syntax is s3express-control.region.amazonaws.com
.
The following operations are related to PutBucketEncryption
:
\n GetBucketEncryption\n
\nSpecifies default encryption for a bucket using server-side encryption with different\n key options. By default, all buckets have a default encryption configuration that uses\n server-side encryption with Amazon S3 managed keys (SSE-S3). You can optionally configure\n default encryption for a bucket by using server-side encryption with an Amazon Web Services KMS key\n (SSE-KMS) or a customer-provided key (SSE-C). For information about the bucket default\n encryption feature, see Amazon S3 Bucket Default Encryption\n in the Amazon S3 User Guide.
", + "smithy.api#documentation": "Specifies default encryption for a bucket using server-side encryption with different\n key options.
\n\n Directory buckets - When you use this operation with a directory bucket, you must use path-style requests in the format https://s3express-control.region_code.amazonaws.com/bucket-name\n
. Virtual-hosted-style requests aren't supported. Directory bucket names must be unique in the chosen Availability Zone. Bucket names must also follow the format \n bucket_base_name--az_id--x-s3
(for example, \n DOC-EXAMPLE-BUCKET--usw2-az1--x-s3
). For information about bucket naming restrictions, see Directory bucket naming rules in the Amazon S3 User Guide\n
The base64-encoded 128-bit MD5 digest of the server-side encryption\n configuration.
\nFor requests made using the Amazon Web Services Command Line Interface (CLI) or Amazon Web Services SDKs, this field is calculated automatically.
", + "smithy.api#documentation": "The base64-encoded 128-bit MD5 digest of the server-side encryption\n configuration.
\nFor requests made using the Amazon Web Services Command Line Interface (CLI) or Amazon Web Services SDKs, this field is calculated automatically.
\nThis functionality is not supported for directory buckets.
\nIndicates the algorithm used to create the checksum for the object when you use the SDK. This header will not provide any\n additional functionality if you don't use the SDK. When you send this header, there must be a corresponding x-amz-checksum
or\n x-amz-trailer
header sent. Otherwise, Amazon S3 fails the request with the HTTP status code 400 Bad Request
. For more\n information, see Checking object integrity in\n the Amazon S3 User Guide.
If you provide an individual checksum, Amazon S3 ignores any provided\n ChecksumAlgorithm
parameter.
Indicates the algorithm used to create the checksum for the object when you use the SDK. This header will not provide any\n additional functionality if you don't use the SDK. When you send this header, there must be a corresponding x-amz-checksum
or\n x-amz-trailer
header sent. Otherwise, Amazon S3 fails the request with the HTTP status code 400 Bad Request
. For more\n information, see Checking object integrity in\n the Amazon S3 User Guide.
If you provide an individual checksum, Amazon S3 ignores any provided\n ChecksumAlgorithm
parameter.
For directory buckets, when you use Amazon Web Services SDKs, CRC32
is the default checksum algorithm that's used for performance.
The account ID of the expected bucket owner. If the account ID that you provide does not match the actual owner of the bucket, the request fails with the HTTP status code 403 Forbidden
(access denied).
The account ID of the expected bucket owner. If the account ID that you provide does not match the actual owner of the bucket, the request fails with the HTTP status code 403 Forbidden
(access denied).
For directory buckets, this header is not supported in this API operation. If you specify this header, the request fails with the HTTP status code \n501 Not Implemented
.
Adds an object to a bucket.
\nAmazon S3 never adds partial objects; if you receive a success response, Amazon S3 added the\n entire object to the bucket. You cannot use PutObject
to only update a\n single piece of metadata for an existing object. You must put the entire object with\n updated metadata if you want to update some values.
If your bucket uses the bucket owner enforced setting for Object Ownership, ACLs are disabled and no longer affect permissions. All\n objects written to the bucket by any account will be owned by the bucket owner.
\n\n Directory buckets - For directory buckets, you must make requests for this API operation to the Zonal endpoint. These endpoints support virtual-hosted-style requests in the format https://bucket_name.s3express-az_id.region.amazonaws.com/key-name\n
. Path-style requests are not supported. For more information, see Regional and Zonal endpoints in the\n Amazon S3 User Guide.
Amazon S3 is a distributed system. If it receives multiple write requests for the same object\n simultaneously, it overwrites all but the last object written. However, Amazon S3 provides features that can modify this behavior:
\n\n S3 Object Lock - To prevent objects from\n being deleted or overwritten, you can use Amazon S3 Object\n Lock in the Amazon S3 User Guide.
\nThis functionality is not supported for directory buckets.
\n\n S3 Versioning - When you enable\n versioning for a bucket, if Amazon S3 receives multiple write requests for the same object\n simultaneously, it stores all versions of the objects. For each write request that is made to the same object, Amazon S3 automatically generates a unique version ID\n of that object being stored in Amazon S3. \n You can retrieve, replace, or delete any version of the object. For more information about versioning, see\n Adding Objects to\n Versioning-Enabled Buckets in the Amazon S3\n User Guide. For information about returning the versioning state\n of a bucket, see GetBucketVersioning.
\nThis functionality is not supported for directory buckets.
\n\n General purpose bucket permissions - The following permissions are required in your policies when your \n PutObject
request includes specific headers.
\n \n s3:PutObject
\n - To successfully complete the PutObject
request, you must always have the s3:PutObject
permission on a bucket to add an object\n to it.
\n \n s3:PutObjectAcl
\n - To successfully change the objects ACL of your PutObject
request, you must have the s3:PutObjectAcl
.
\n \n s3:PutObjectTagging
\n - To successfully set the tag-set with your PutObject
request, you\n must have the s3:PutObjectTagging
.
\n Directory bucket permissions - To grant access to this API operation on a directory bucket, we recommend that you use the \n CreateSession
\n API operation for session-based authorization. Specifically, you grant the s3express:CreateSession
permission to the directory bucket in a bucket policy or an IAM identity-based policy. Then, you make the CreateSession
API call on the bucket to obtain a session token. With the session token in your request header, you can make API requests to this operation. After the session token expires, you make another CreateSession
API call to generate a new session token for use. \nAmazon Web Services CLI or SDKs create session and refresh the session token automatically to avoid service interruptions when a session expires. For more information about authorization, see \n CreateSession
\n .
\n General purpose bucket - To ensure that data is not corrupted traversing the network, use the\n Content-MD5
header. When you use this header, Amazon S3 checks the object\n against the provided MD5 value and, if they do not match, Amazon S3 returns an error. Alternatively, when the object's ETag is its MD5 digest, \n you can calculate the MD5 while putting the object to Amazon S3 and compare the returned ETag to\n the calculated MD5 value.
\n Directory bucket - This functionality is not supported for directory buckets.
\n\n Directory buckets - The HTTP Host header syntax is \n Bucket_name.s3express-az_id.region.amazonaws.com
.
For more information about related Amazon S3 APIs, see the following:
\n\n CopyObject\n
\n\n DeleteObject\n
\nAdds an object to a bucket.
\nAmazon S3 never adds partial objects; if you receive a success response, Amazon S3 added the\n entire object to the bucket. You cannot use PutObject
to only update a\n single piece of metadata for an existing object. You must put the entire object with\n updated metadata if you want to update some values.
If your bucket uses the bucket owner enforced setting for Object Ownership, ACLs are disabled and no longer affect permissions. All\n objects written to the bucket by any account will be owned by the bucket owner.
\n\n Directory buckets - For directory buckets, you must make requests for this API operation to the Zonal endpoint. These endpoints support virtual-hosted-style requests in the format https://bucket_name.s3express-az_id.region.amazonaws.com/key-name\n
. Path-style requests are not supported. For more information, see Regional and Zonal endpoints in the\n Amazon S3 User Guide.
Amazon S3 is a distributed system. If it receives multiple write requests for the same object\n simultaneously, it overwrites all but the last object written. However, Amazon S3 provides features that can modify this behavior:
\n\n S3 Object Lock - To prevent objects from\n being deleted or overwritten, you can use Amazon S3 Object\n Lock in the Amazon S3 User Guide.
\nThis functionality is not supported for directory buckets.
\n\n S3 Versioning - When you enable\n versioning for a bucket, if Amazon S3 receives multiple write requests for the same object\n simultaneously, it stores all versions of the objects. For each write request that is made to the same object, Amazon S3 automatically generates a unique version ID\n of that object being stored in Amazon S3. \n You can retrieve, replace, or delete any version of the object. For more information about versioning, see\n Adding Objects to\n Versioning-Enabled Buckets in the Amazon S3\n User Guide. For information about returning the versioning state\n of a bucket, see GetBucketVersioning.
\nThis functionality is not supported for directory buckets.
\n\n General purpose bucket permissions - The following permissions are required in your policies when your \n PutObject
request includes specific headers.
\n \n s3:PutObject
\n - To successfully complete the PutObject
request, you must always have the s3:PutObject
permission on a bucket to add an object\n to it.
\n \n s3:PutObjectAcl
\n - To successfully change the objects ACL of your PutObject
request, you must have the s3:PutObjectAcl
.
\n \n s3:PutObjectTagging
\n - To successfully set the tag-set with your PutObject
request, you\n must have the s3:PutObjectTagging
.
\n Directory bucket permissions - To grant access to this API operation on a directory bucket, we recommend that you use the \n CreateSession
\n API operation for session-based authorization. Specifically, you grant the s3express:CreateSession
permission to the directory bucket in a bucket policy or an IAM identity-based policy. Then, you make the CreateSession
API call on the bucket to obtain a session token. With the session token in your request header, you can make API requests to this operation. After the session token expires, you make another CreateSession
API call to generate a new session token for use. \nAmazon Web Services CLI or SDKs create session and refresh the session token automatically to avoid service interruptions when a session expires. For more information about authorization, see \n CreateSession
\n .
If the object is encrypted with\n SSE-KMS, you must also have the\n kms:GenerateDataKey
and kms:Decrypt
permissions in IAM identity-based policies and KMS key policies for the KMS key.
\n General purpose bucket - To ensure that data is not corrupted traversing the network, use the\n Content-MD5
header. When you use this header, Amazon S3 checks the object\n against the provided MD5 value and, if they do not match, Amazon S3 returns an error. Alternatively, when the object's ETag is its MD5 digest, \n you can calculate the MD5 while putting the object to Amazon S3 and compare the returned ETag to\n the calculated MD5 value.
\n Directory bucket - This functionality is not supported for directory buckets.
\n\n Directory buckets - The HTTP Host header syntax is \n Bucket_name.s3express-az_id.region.amazonaws.com
.
For more information about related Amazon S3 APIs, see the following:
\n\n CopyObject\n
\n\n DeleteObject\n
\nThe server-side encryption algorithm used when you store this object in Amazon S3 (for example,\n AES256
, aws:kms
, aws:kms:dsse
).
For directory buckets, only server-side encryption with Amazon S3 managed keys (SSE-S3) (AES256
) is supported.
The server-side encryption algorithm used when you store this object in Amazon S3.
", "smithy.api#httpHeader": "x-amz-server-side-encryption" } }, @@ -31296,21 +31352,21 @@ "SSEKMSKeyId": { "target": "com.amazonaws.s3#SSEKMSKeyId", "traits": { - "smithy.api#documentation": "If x-amz-server-side-encryption
has a valid value of aws:kms
\n or aws:kms:dsse
, this header indicates the ID of the Key Management Service (KMS)\n symmetric encryption customer managed key that was used for the object.
This functionality is not supported for directory buckets.
\nIf present, indicates the ID of the KMS key that was used for object encryption.
", "smithy.api#httpHeader": "x-amz-server-side-encryption-aws-kms-key-id" } }, "SSEKMSEncryptionContext": { "target": "com.amazonaws.s3#SSEKMSEncryptionContext", "traits": { - "smithy.api#documentation": "If present, indicates the Amazon Web Services KMS Encryption Context to use for object encryption. The\n value of this header is a base64-encoded UTF-8 string holding JSON with the encryption\n context key-value pairs. This value is stored as object metadata and automatically gets\n passed on to Amazon Web Services KMS for future GetObject
or CopyObject
\n operations on this object.
This functionality is not supported for directory buckets.
\nIf present, indicates the Amazon Web Services KMS Encryption Context to use for object encryption. The value of\n this header is a Base64-encoded string of a UTF-8 encoded JSON, which contains the encryption context as key-value pairs. \n This value is stored as object metadata and automatically gets\n passed on to Amazon Web Services KMS for future GetObject
\n operations on this object.
Indicates whether the uploaded object uses an S3 Bucket Key for server-side encryption\n with Key Management Service (KMS) keys (SSE-KMS).
\nThis functionality is not supported for directory buckets.
\nIndicates whether the uploaded object uses an S3 Bucket Key for server-side encryption\n with Key Management Service (KMS) keys (SSE-KMS).
", "smithy.api#httpHeader": "x-amz-server-side-encryption-bucket-key-enabled" } }, @@ -31501,7 +31557,7 @@ "ServerSideEncryption": { "target": "com.amazonaws.s3#ServerSideEncryption", "traits": { - "smithy.api#documentation": "The server-side encryption algorithm that was used when you store this object in Amazon S3 (for example,\n AES256
, aws:kms
, aws:kms:dsse
).
\n General purpose buckets - You have four mutually exclusive options to protect data using server-side encryption in\n Amazon S3, depending on how you choose to manage the encryption keys. Specifically, the\n encryption key options are Amazon S3 managed keys (SSE-S3), Amazon Web Services KMS keys (SSE-KMS or\n DSSE-KMS), and customer-provided keys (SSE-C). Amazon S3 encrypts data with server-side\n encryption by using Amazon S3 managed keys (SSE-S3) by default. You can optionally tell Amazon S3 to\n encrypt data at rest by using server-side encryption with other key options. For more\n information, see Using Server-Side\n Encryption in the Amazon S3 User Guide.
\n\n Directory buckets - For directory buckets, only the server-side encryption with Amazon S3 managed keys (SSE-S3) (AES256
) value is supported.
The server-side encryption algorithm that was used when you store this object in Amazon S3 (for example,\n AES256
, aws:kms
, aws:kms:dsse
).
\n General purpose buckets - You have four mutually exclusive options to protect data using server-side encryption in\n Amazon S3, depending on how you choose to manage the encryption keys. Specifically, the\n encryption key options are Amazon S3 managed keys (SSE-S3), Amazon Web Services KMS keys (SSE-KMS or\n DSSE-KMS), and customer-provided keys (SSE-C). Amazon S3 encrypts data with server-side\n encryption by using Amazon S3 managed keys (SSE-S3) by default. You can optionally tell Amazon S3 to\n encrypt data at rest by using server-side encryption with other key options. For more\n information, see Using Server-Side\n Encryption in the Amazon S3 User Guide.
\n\n Directory buckets - For directory buckets, there are only two supported options for server-side encryption: server-side encryption with Amazon S3 managed keys (SSE-S3) (AES256
) and server-side encryption with KMS keys (SSE-KMS) (aws:kms
). We recommend that the bucket's default encryption uses the desired encryption configuration and you don't override the bucket default encryption in your \n CreateSession
requests or PUT
object requests. Then, new objects \n are automatically encrypted with the desired encryption settings. For more\n information, see Protecting data with server-side encryption in the Amazon S3 User Guide. For more information about the encryption overriding behaviors in directory buckets, see Specifying server-side encryption with KMS for new object uploads. \n
In the Zonal endpoint API calls (except CopyObject and UploadPartCopy) using the REST API, the encryption request headers must match the encryption settings that are specified in the CreateSession
request. \n You can't override the values of the encryption settings (x-amz-server-side-encryption
, x-amz-server-side-encryption-aws-kms-key-id
, x-amz-server-side-encryption-context
, and x-amz-server-side-encryption-bucket-key-enabled
) that are specified in the CreateSession
request. \n You don't need to explicitly specify these encryption settings values in Zonal endpoint API calls, and \n Amazon S3 will use the encryption settings values from the CreateSession
request to protect new objects in the directory bucket. \n
When you use the CLI or the Amazon Web Services SDKs, for CreateSession
, the session token refreshes automatically to avoid service interruptions when a session expires. The CLI or the Amazon Web Services SDKs use the bucket's default encryption configuration for the \n CreateSession
request. It's not supported to override the encryption settings values in the CreateSession
request. \n So in the Zonal endpoint API calls (except CopyObject and UploadPartCopy), \n the encryption request headers must match the default encryption configuration of the directory bucket.\n\n
If x-amz-server-side-encryption
has a valid value of aws:kms
\n or aws:kms:dsse
, this header specifies the ID (Key ID, Key ARN, or Key Alias) of the Key Management Service (KMS)\n symmetric encryption customer managed key that was used for the object. If you specify\n x-amz-server-side-encryption:aws:kms
or\n x-amz-server-side-encryption:aws:kms:dsse
, but do not provide\n x-amz-server-side-encryption-aws-kms-key-id
, Amazon S3 uses the Amazon Web Services managed key\n (aws/s3
) to protect the data. If the KMS key does not exist in the same\n account that's issuing the command, you must use the full ARN and not just the ID.
This functionality is not supported for directory buckets.
\nSpecifies the KMS key ID (Key ID, Key ARN, or Key Alias) to use for object encryption. If the KMS key doesn't exist in the same\n account that's issuing the command, you must use the full Key ARN not the Key ID.
\n\n General purpose buckets - If you specify x-amz-server-side-encryption
with aws:kms
or aws:kms:dsse
, this header specifies the ID (Key ID, Key ARN, or Key Alias) of the KMS \n key to use. If you specify\n x-amz-server-side-encryption:aws:kms
or\n x-amz-server-side-encryption:aws:kms:dsse
, but do not provide x-amz-server-side-encryption-aws-kms-key-id
, Amazon S3 uses the Amazon Web Services managed key\n (aws/s3
) to protect the data.
\n Directory buckets - If you specify x-amz-server-side-encryption
with aws:kms
, you must specify the \n x-amz-server-side-encryption-aws-kms-key-id
header with the ID (Key ID or Key ARN) of the KMS \n symmetric encryption customer managed key to use. Otherwise, you get an HTTP 400 Bad Request
error. Only use the key ID or key ARN. The key alias format of the KMS key isn't supported. Your SSE-KMS configuration can only support 1 customer managed key per directory bucket for the lifetime of the bucket. \nAmazon Web Services managed key (aws/s3
) isn't supported. \n
Specifies the Amazon Web Services KMS Encryption Context to use for object encryption. The value of\n this header is a base64-encoded UTF-8 string holding JSON with the encryption context\n key-value pairs. This value is stored as object metadata and automatically gets passed on\n to Amazon Web Services KMS for future GetObject
or CopyObject
operations on\n this object. This value must be explicitly added during CopyObject
operations.
This functionality is not supported for directory buckets.
\nSpecifies the Amazon Web Services KMS Encryption Context as an additional encryption context to use for object encryption. The value of\n this header is a Base64-encoded string of a UTF-8 encoded JSON, which contains the encryption context as key-value pairs. \n This value is stored as object metadata and automatically gets passed on\n to Amazon Web Services KMS for future GetObject
operations on\n this object.
\n General purpose buckets - This value must be explicitly added during CopyObject
operations if you want an additional encryption context for your object. For more information, see Encryption context in the Amazon S3 User Guide.
\n Directory buckets - You can optionally provide an explicit encryption context value. The value must match the default encryption context - the bucket Amazon Resource Name (ARN). An additional encryption context value is not supported.
", "smithy.api#httpHeader": "x-amz-server-side-encryption-context" } }, "BucketKeyEnabled": { "target": "com.amazonaws.s3#BucketKeyEnabled", "traits": { - "smithy.api#documentation": "Specifies whether Amazon S3 should use an S3 Bucket Key for object encryption with\n server-side encryption using Key Management Service (KMS) keys (SSE-KMS). Setting this header to\n true
causes Amazon S3 to use an S3 Bucket Key for object encryption with\n SSE-KMS.
Specifying this header with a PUT action doesn’t affect bucket-level settings for S3\n Bucket Key.
\nThis functionality is not supported for directory buckets.
\nSpecifies whether Amazon S3 should use an S3 Bucket Key for object encryption with\n server-side encryption using Key Management Service (KMS) keys (SSE-KMS).
\n\n General purpose buckets - Setting this header to\n true
causes Amazon S3 to use an S3 Bucket Key for object encryption with\n SSE-KMS. Also, specifying this header with a PUT action doesn't affect bucket-level settings for S3\n Bucket Key.
\n Directory buckets - S3 Bucket Keys are always enabled for GET
and PUT
operations in a directory bucket and can’t be disabled. S3 Bucket Keys aren't supported, when you copy SSE-KMS encrypted objects from general purpose buckets \nto directory buckets, from directory buckets to general purpose buckets, or between directory buckets, through CopyObject, UploadPartCopy, the Copy operation in Batch Operations, or \n the import jobs. In this case, Amazon S3 makes a call to KMS every time a copy request is made for a KMS-encrypted object.
Optional configuration to replicate existing source bucket objects. For more\n information, see Replicating Existing Objects in the Amazon S3 User Guide.\n
" + "smithy.api#documentation": "Optional configuration to replicate existing source bucket objects.\n
\nThis parameter is no longer supported. To replicate existing objects, see Replicating existing objects with S3 Batch Replication in the Amazon S3 User Guide.
\nServer-side encryption algorithm to use for the default encryption.
", + "smithy.api#documentation": "Server-side encryption algorithm to use for the default encryption.
\nFor directory buckets, there are only two supported values for server-side encryption: AES256
and aws:kms
.
Amazon Web Services Key Management Service (KMS) customer Amazon Web Services KMS key ID to use for the default\n encryption. This parameter is allowed if and only if SSEAlgorithm
is set to\n aws:kms
or aws:kms:dsse
.
You can specify the key ID, key alias, or the Amazon Resource Name (ARN) of the KMS\n key.
\nKey ID: 1234abcd-12ab-34cd-56ef-1234567890ab
\n
Key ARN: arn:aws:kms:us-east-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab
\n
Key Alias: alias/alias-name
\n
If you use a key ID, you can run into a LogDestination undeliverable error when creating\n a VPC flow log.
\nIf you are using encryption with cross-account or Amazon Web Services service operations you must use\n a fully qualified KMS key ARN. For more information, see Using encryption for cross-account operations.
\nAmazon S3 only supports symmetric encryption KMS keys. For more information, see Asymmetric keys in Amazon Web Services KMS in the Amazon Web Services Key Management Service\n Developer Guide.
\nAmazon Web Services Key Management Service (KMS) customer managed key ID to use for the default\n encryption.
\n\n General purpose buckets - This parameter is allowed if and only if SSEAlgorithm
is set to\n aws:kms
or aws:kms:dsse
.
\n Directory buckets - This parameter is allowed if and only if SSEAlgorithm
is set to\n aws:kms
.
You can specify the key ID, key alias, or the Amazon Resource Name (ARN) of the KMS\n key.
\nKey ID: 1234abcd-12ab-34cd-56ef-1234567890ab
\n
Key ARN: arn:aws:kms:us-east-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab
\n
Key Alias: alias/alias-name
\n
If you are using encryption with cross-account or Amazon Web Services service operations, you must use\n a fully qualified KMS key ARN. For more information, see Using encryption for cross-account operations.
\n\n General purpose buckets - If you're specifying a customer managed KMS key, we recommend using a fully qualified\n KMS key ARN. If you use a KMS key alias instead, then KMS resolves the key within the\n requester’s account. This behavior can result in data that's encrypted with a KMS key\n that belongs to the requester, and not the bucket owner. Also, if you use a key ID, you can run into a LogDestination undeliverable error when creating\n a VPC flow log. \n
\n\n Directory buckets - When you specify an KMS customer managed key for encryption in your directory bucket, only use the key ID or key ARN. The key alias format of the KMS key isn't supported.
\nAmazon S3 only supports symmetric encryption KMS keys. For more information, see Asymmetric keys in Amazon Web Services KMS in the Amazon Web Services Key Management Service\n Developer Guide.
\nDescribes the default server-side encryption to apply to new objects in the bucket. If a\n PUT Object request doesn't specify any server-side encryption, this default encryption will\n be applied. If you don't specify a customer managed key at configuration, Amazon S3 automatically creates\n an Amazon Web Services KMS key in your Amazon Web Services account the first time that you add an object encrypted\n with SSE-KMS to a bucket. By default, Amazon S3 uses this KMS key for SSE-KMS. For more\n information, see PUT Bucket encryption in\n the Amazon S3 API Reference.
\nIf you're specifying a customer managed KMS key, we recommend using a fully qualified\n KMS key ARN. If you use a KMS key alias instead, then KMS resolves the key within the\n requester’s account. This behavior can result in data that's encrypted with a KMS key\n that belongs to the requester, and not the bucket owner.
\nDescribes the default server-side encryption to apply to new objects in the bucket. If a\n PUT Object request doesn't specify any server-side encryption, this default encryption will\n be applied. For more\n information, see PutBucketEncryption.
\n\n General purpose buckets - If you don't specify a customer managed key at configuration, Amazon S3 automatically creates\n an Amazon Web Services KMS key (aws/s3
) in your Amazon Web Services account the first time that you add an object encrypted\n with SSE-KMS to a bucket. By default, Amazon S3 uses this KMS key for SSE-KMS.
\n Directory buckets - Your SSE-KMS configuration can only support 1 customer managed key per directory bucket for the lifetime of the bucket. \nAmazon Web Services managed key (aws/s3
) isn't supported. \n
\n Directory buckets - For directory buckets, there are only two supported options for server-side encryption: SSE-S3 and SSE-KMS.
\nSpecifies whether Amazon S3 should use an S3 Bucket Key with server-side encryption using KMS\n (SSE-KMS) for new objects in the bucket. Existing objects are not affected. Setting the\n BucketKeyEnabled
element to true
causes Amazon S3 to use an S3\n Bucket Key. By default, S3 Bucket Key is not enabled.
For more information, see Amazon S3 Bucket Keys in the\n Amazon S3 User Guide.
" + "smithy.api#documentation": "Specifies whether Amazon S3 should use an S3 Bucket Key with server-side encryption using KMS\n (SSE-KMS) for new objects in the bucket. Existing objects are not affected. Setting the\n BucketKeyEnabled
element to true
causes Amazon S3 to use an S3\n Bucket Key.
\n General purpose buckets - By default, S3 Bucket Key is not enabled. For more information, see Amazon S3 Bucket Keys in the\n Amazon S3 User Guide.
\n\n Directory buckets - S3 Bucket Keys are always enabled for GET
and PUT
operations in a directory bucket and can’t be disabled. S3 Bucket Keys aren't supported, when you copy SSE-KMS encrypted objects from general purpose buckets \nto directory buckets, from directory buckets to general purpose buckets, or between directory buckets, through CopyObject, UploadPartCopy, the Copy operation in Batch Operations, or \n the import jobs. In this case, Amazon S3 makes a call to KMS every time a copy request is made for a KMS-encrypted object.
Specifies the default server-side encryption configuration.
\nIf you're specifying a customer managed KMS key, we recommend using a fully qualified\n KMS key ARN. If you use a KMS key alias instead, then KMS resolves the key within the\n requester’s account. This behavior can result in data that's encrypted with a KMS key\n that belongs to the requester, and not the bucket owner.
\nSpecifies the default server-side encryption configuration.
\n\n General purpose buckets - If you're specifying a customer managed KMS key, we recommend using a fully qualified\n KMS key ARN. If you use a KMS key alias instead, then KMS resolves the key within the\n requester’s account. This behavior can result in data that's encrypted with a KMS key\n that belongs to the requester, and not the bucket owner.
\n\n Directory buckets - When you specify an KMS customer managed key for encryption in your directory bucket, only use the key ID or key ARN. The key alias format of the KMS key isn't supported.
\nThe established temporary security credentials of the session.
\n\n Directory buckets - These session credentials are only supported for the authentication and authorization of Zonal endpoint APIs on directory buckets.
\nThe established temporary security credentials of the session.
\n\n Directory buckets - These session credentials are only supported for the authentication and authorization of Zonal endpoint API operations on directory buckets.
\nUploads a part in a multipart upload.
\nIn this operation, you provide new data as a part of an object in your request. However, you have an option\n to specify your existing Amazon S3 object as a data source for the part you are uploading. To\n upload a part from an existing object, you use the UploadPartCopy operation.\n
\nYou must initiate a multipart upload (see CreateMultipartUpload)\n before you can upload any part. In response to your initiate request, Amazon S3 returns an\n upload ID, a unique identifier that you must include in your upload part request.
\nPart numbers can be any number from 1 to 10,000, inclusive. A part number uniquely\n identifies a part and also defines its position within the object being created. If you\n upload a new part using the same part number that was used with a previous part, the\n previously uploaded part is overwritten.
\nFor information about maximum and minimum part sizes and other multipart upload\n specifications, see Multipart upload limits in the Amazon S3 User Guide.
\nAfter you initiate multipart upload and upload\n one or more parts, you must either complete or abort multipart upload in order to stop\n getting charged for storage of the uploaded parts. Only after you either complete or abort\n multipart upload, Amazon S3 frees up the parts storage and stops charging you for the parts\n storage.
\nFor more information on multipart uploads, go to Multipart Upload Overview in the\n Amazon S3 User Guide .
\n\n Directory buckets - For directory buckets, you must make requests for this API operation to the Zonal endpoint. These endpoints support virtual-hosted-style requests in the format https://bucket_name.s3express-az_id.region.amazonaws.com/key-name\n
. Path-style requests are not supported. For more information, see Regional and Zonal endpoints in the\n Amazon S3 User Guide.
\n General purpose bucket permissions - To\n perform a multipart upload with encryption using an Key Management Service key, the\n requester must have permission to the kms:Decrypt
and\n kms:GenerateDataKey
actions on the key. The requester must\n also have permissions for the kms:GenerateDataKey
action for\n the CreateMultipartUpload
API. Then, the requester needs\n permissions for the kms:Decrypt
action on the\n UploadPart
and UploadPartCopy
APIs.
These permissions are required because Amazon S3 must decrypt and read data\n from the encrypted file parts before it completes the multipart upload. For\n more information about KMS permissions, see Protecting data\n using server-side encryption with KMS in the\n Amazon S3 User Guide. For information about the\n permissions required to use the multipart upload API, see Multipart upload and permissions and Multipart upload API and permissions in the\n Amazon S3 User Guide.
\n\n Directory bucket permissions - To grant access to this API operation on a directory bucket, we recommend that you use the \n CreateSession
\n API operation for session-based authorization. Specifically, you grant the s3express:CreateSession
permission to the directory bucket in a bucket policy or an IAM identity-based policy. Then, you make the CreateSession
API call on the bucket to obtain a session token. With the session token in your request header, you can make API requests to this operation. After the session token expires, you make another CreateSession
API call to generate a new session token for use. \nAmazon Web Services CLI or SDKs create session and refresh the session token automatically to avoid service interruptions when a session expires. For more information about authorization, see \n CreateSession
\n .
\n General purpose bucket - To ensure that data is not corrupted traversing the network, specify the\n Content-MD5
header in the upload part request. Amazon S3 checks the part data against the provided MD5 value. If they do not match, Amazon S3 returns an error. If the upload request is signed with Signature Version 4, then Amazon Web Services S3 uses the\n x-amz-content-sha256
header as a checksum instead of\n Content-MD5
. For more information see Authenticating\n Requests: Using the Authorization Header (Amazon Web Services Signature Version 4).
\n Directory buckets - MD5 is not supported by directory buckets. You can use checksum algorithms to check object integrity.
\n\n General purpose bucket - Server-side encryption is for data encryption at rest. Amazon S3 encrypts your data as it\n writes it to disks in its data centers and decrypts it when you access it. You have \n mutually exclusive options to protect data using server-side encryption in Amazon S3, depending\n on how you choose to manage the encryption keys. Specifically, the encryption key options\n are Amazon S3 managed keys (SSE-S3), Amazon Web Services KMS keys (SSE-KMS), and Customer-Provided Keys\n (SSE-C). Amazon S3 encrypts data with server-side encryption using Amazon S3 managed keys (SSE-S3) by\n default. You can optionally tell Amazon S3 to encrypt data at rest using server-side encryption\n with other key options. The option you use depends on whether you want to use KMS keys\n (SSE-KMS) or provide your own encryption key (SSE-C).
\nServer-side encryption is supported by the S3 Multipart Upload operations. Unless you are\n using a customer-provided encryption key (SSE-C), you don't need to specify the encryption\n parameters in each UploadPart request. Instead, you only need to specify the server-side\n encryption parameters in the initial Initiate Multipart request. For more information, see\n CreateMultipartUpload.
\nIf you request server-side encryption using a customer-provided encryption key (SSE-C)\n in your initiate multipart upload request, you must provide identical encryption\n information in each part upload using the following request headers.
\nx-amz-server-side-encryption-customer-algorithm
\nx-amz-server-side-encryption-customer-key
\nx-amz-server-side-encryption-customer-key-MD5
\n\n Directory bucket - For directory buckets, only server-side encryption with Amazon S3 managed keys (SSE-S3) (AES256
) is supported.
\n For more information, see Using Server-Side\n Encryption in the Amazon S3 User Guide.
\nError Code: NoSuchUpload
\n
Description: The specified multipart upload does not exist. The upload ID might be invalid, or the multipart upload might have been aborted or completed.
\nHTTP Status Code: 404 Not Found
\nSOAP Fault Code Prefix: Client
\n\n Directory buckets - The HTTP Host header syntax is \n Bucket_name.s3express-az_id.region.amazonaws.com
.
The following operations are related to UploadPart
:
\n AbortMultipartUpload\n
\n\n ListParts\n
\n\n ListMultipartUploads\n
\nUploads a part in a multipart upload.
\nIn this operation, you provide new data as a part of an object in your request. However, you have an option\n to specify your existing Amazon S3 object as a data source for the part you are uploading. To\n upload a part from an existing object, you use the UploadPartCopy operation.\n
\nYou must initiate a multipart upload (see CreateMultipartUpload)\n before you can upload any part. In response to your initiate request, Amazon S3 returns an\n upload ID, a unique identifier that you must include in your upload part request.
\nPart numbers can be any number from 1 to 10,000, inclusive. A part number uniquely\n identifies a part and also defines its position within the object being created. If you\n upload a new part using the same part number that was used with a previous part, the\n previously uploaded part is overwritten.
\nFor information about maximum and minimum part sizes and other multipart upload\n specifications, see Multipart upload limits in the Amazon S3 User Guide.
\nAfter you initiate multipart upload and upload\n one or more parts, you must either complete or abort multipart upload in order to stop\n getting charged for storage of the uploaded parts. Only after you either complete or abort\n multipart upload, Amazon S3 frees up the parts storage and stops charging you for the parts\n storage.
\nFor more information on multipart uploads, go to Multipart Upload Overview in the\n Amazon S3 User Guide .
\n\n Directory buckets - For directory buckets, you must make requests for this API operation to the Zonal endpoint. These endpoints support virtual-hosted-style requests in the format https://bucket_name.s3express-az_id.region.amazonaws.com/key-name\n
. Path-style requests are not supported. For more information, see Regional and Zonal endpoints in the\n Amazon S3 User Guide.
\n General purpose bucket permissions - To\n perform a multipart upload with encryption using an Key Management Service key, the\n requester must have permission to the kms:Decrypt
and\n kms:GenerateDataKey
actions on the key. The requester must\n also have permissions for the kms:GenerateDataKey
action for\n the CreateMultipartUpload
API. Then, the requester needs\n permissions for the kms:Decrypt
action on the\n UploadPart
and UploadPartCopy
APIs.
These permissions are required because Amazon S3 must decrypt and read data\n from the encrypted file parts before it completes the multipart upload. For\n more information about KMS permissions, see Protecting data\n using server-side encryption with KMS in the\n Amazon S3 User Guide. For information about the\n permissions required to use the multipart upload API, see Multipart upload and permissions and Multipart upload API and permissions in the\n Amazon S3 User Guide.
\n\n Directory bucket permissions - To grant access to this API operation on a directory bucket, we recommend that you use the \n CreateSession
\n API operation for session-based authorization. Specifically, you grant the s3express:CreateSession
permission to the directory bucket in a bucket policy or an IAM identity-based policy. Then, you make the CreateSession
API call on the bucket to obtain a session token. With the session token in your request header, you can make API requests to this operation. After the session token expires, you make another CreateSession
API call to generate a new session token for use. \nAmazon Web Services CLI or SDKs create session and refresh the session token automatically to avoid service interruptions when a session expires. For more information about authorization, see \n CreateSession
\n .
If the object is encrypted with\n SSE-KMS, you must also have the\n kms:GenerateDataKey
and kms:Decrypt
permissions in IAM identity-based policies and KMS key policies for the KMS key.
\n General purpose bucket - To ensure that data is not corrupted traversing the network, specify the\n Content-MD5
header in the upload part request. Amazon S3 checks the part data against the provided MD5 value. If they do not match, Amazon S3 returns an error. If the upload request is signed with Signature Version 4, then Amazon Web Services S3 uses the\n x-amz-content-sha256
header as a checksum instead of\n Content-MD5
. For more information see Authenticating\n Requests: Using the Authorization Header (Amazon Web Services Signature Version 4).
\n Directory buckets - MD5 is not supported by directory buckets. You can use checksum algorithms to check object integrity.
\n\n General purpose bucket - Server-side encryption is for data encryption at rest. Amazon S3 encrypts your data as it\n writes it to disks in its data centers and decrypts it when you access it. You have \n mutually exclusive options to protect data using server-side encryption in Amazon S3, depending\n on how you choose to manage the encryption keys. Specifically, the encryption key options\n are Amazon S3 managed keys (SSE-S3), Amazon Web Services KMS keys (SSE-KMS), and Customer-Provided Keys\n (SSE-C). Amazon S3 encrypts data with server-side encryption using Amazon S3 managed keys (SSE-S3) by\n default. You can optionally tell Amazon S3 to encrypt data at rest using server-side encryption\n with other key options. The option you use depends on whether you want to use KMS keys\n (SSE-KMS) or provide your own encryption key (SSE-C).
\nServer-side encryption is supported by the S3 Multipart Upload operations. Unless you are\n using a customer-provided encryption key (SSE-C), you don't need to specify the encryption\n parameters in each UploadPart request. Instead, you only need to specify the server-side\n encryption parameters in the initial Initiate Multipart request. For more information, see\n CreateMultipartUpload.
\nIf you request server-side encryption using a customer-provided encryption key (SSE-C)\n in your initiate multipart upload request, you must provide identical encryption\n information in each part upload using the following request headers.
\nx-amz-server-side-encryption-customer-algorithm
\nx-amz-server-side-encryption-customer-key
\nx-amz-server-side-encryption-customer-key-MD5
\n\n For more information, see Using Server-Side\n Encryption in the Amazon S3 User Guide.
\n\n Directory buckets - For directory buckets, there are only two supported options for server-side encryption: server-side encryption with Amazon S3 managed keys (SSE-S3) (AES256
) and server-side encryption with KMS keys (SSE-KMS) (aws:kms
).
Error Code: NoSuchUpload
\n
Description: The specified multipart upload does not exist. The upload ID might be invalid, or the multipart upload might have been aborted or completed.
\nHTTP Status Code: 404 Not Found
\nSOAP Fault Code Prefix: Client
\n\n Directory buckets - The HTTP Host header syntax is \n Bucket_name.s3express-az_id.region.amazonaws.com
.
The following operations are related to UploadPart
:
\n AbortMultipartUpload\n
\n\n ListParts\n
\n\n ListMultipartUploads\n
\nUploads a part by copying data from an existing object as data source. To specify the\n data source, you add the request header x-amz-copy-source
in your request. To specify \n a byte range, you add the request header x-amz-copy-source-range
in your\n request.
For information about maximum and minimum part sizes and other multipart upload\n specifications, see Multipart upload limits in the Amazon S3 User Guide.
\nInstead of copying data from an existing object as part data, you might use the UploadPart\n action to upload new data as a part of an object in your request.
\nYou must initiate a multipart upload before you can upload any part. In response to your\n initiate request, Amazon S3 returns the upload ID, a unique identifier that you must include in\n your upload part request.
\nFor conceptual information about multipart uploads, see Uploading\n Objects Using Multipart Upload in the\n Amazon S3 User Guide. For information about copying objects using a single atomic action vs. a multipart\n upload, see Operations on Objects in\n the Amazon S3 User Guide.
\n\n Directory buckets - For directory buckets, you must make requests for this API operation to the Zonal endpoint. These endpoints support virtual-hosted-style requests in the format https://bucket_name.s3express-az_id.region.amazonaws.com/key-name\n
. Path-style requests are not supported. For more information, see Regional and Zonal endpoints in the\n Amazon S3 User Guide.
All UploadPartCopy
requests must be authenticated and signed by using IAM credentials (access key ID and secret access key for the IAM identities). All headers with the x-amz-
prefix, including\n x-amz-copy-source
, must be signed. For more information, see REST Authentication.
\n Directory buckets - You must use IAM credentials to authenticate and authorize your access to the UploadPartCopy
API operation, instead of using the \n temporary security credentials through the CreateSession
API operation.
Amazon Web Services CLI or SDKs handles authentication and authorization on your behalf.
\nYou must have READ
access to the source object and WRITE
\n access to the destination bucket.
\n General purpose bucket permissions - You\n must have the permissions in a policy based on the bucket types of your\n source bucket and destination bucket in an UploadPartCopy
\n operation.
If the source object is in a general purpose bucket, you must have the\n \n s3:GetObject
\n \n permission to read the source object that is being copied.
If the destination bucket is a general purpose bucket, you must have the\n \n s3:PutObject
\n \n permission to write the object copy to the destination bucket.
To perform a multipart upload with encryption using an Key Management Service\n key, the requester must have permission to the\n kms:Decrypt
and kms:GenerateDataKey
\n actions on the key. The requester must also have permissions for the\n kms:GenerateDataKey
action for the\n CreateMultipartUpload
API. Then, the requester needs\n permissions for the kms:Decrypt
action on the\n UploadPart
and UploadPartCopy
APIs. These\n permissions are required because Amazon S3 must decrypt and read data from\n the encrypted file parts before it completes the multipart upload. For\n more information about KMS permissions, see Protecting\n data using server-side encryption with KMS in the\n Amazon S3 User Guide. For information about the\n permissions required to use the multipart upload API, see Multipart upload\n and permissions and Multipart upload API and permissions in the\n Amazon S3 User Guide.
\n Directory bucket permissions -\n You must have permissions in a bucket policy or an IAM identity-based policy based on the source and destination\n bucket types in an UploadPartCopy
operation.
If the source object that you want to copy is in a\n directory bucket, you must have the \n s3express:CreateSession
\n permission in\n the Action
element of a policy to read the object. By\n default, the session is in the ReadWrite
mode. If you\n want to restrict the access, you can explicitly set the\n s3express:SessionMode
condition key to\n ReadOnly
on the copy source bucket.
If the copy destination is a directory bucket, you must have the \n \n s3express:CreateSession
\n permission in the\n Action
element of a policy to write the object\n to the destination. The s3express:SessionMode
condition\n key cannot be set to ReadOnly
on the copy destination.
For example policies, see Example bucket policies for S3 Express One Zone and Amazon Web Services Identity and Access Management (IAM) identity-based policies for S3 Express One Zone in the\n Amazon S3 User Guide.
\n\n General purpose buckets - \n \n For information about using server-side encryption with customer-provided\n encryption keys with the UploadPartCopy
operation, see CopyObject and UploadPart.\n
\n Directory buckets - For directory buckets, only server-side encryption with Amazon S3 managed keys (SSE-S3) (AES256
) is supported.
Error Code: NoSuchUpload
\n
Description: The specified multipart upload does not exist. The\n upload ID might be invalid, or the multipart upload might have been\n aborted or completed.
\nHTTP Status Code: 404 Not Found
\nError Code: InvalidRequest
\n
Description: The specified copy source is not supported as a\n byte-range copy source.
\nHTTP Status Code: 400 Bad Request
\n\n Directory buckets - The HTTP Host header syntax is \n Bucket_name.s3express-az_id.region.amazonaws.com
.
The following operations are related to UploadPartCopy
:
\n UploadPart\n
\n\n AbortMultipartUpload\n
\n\n ListParts\n
\n\n ListMultipartUploads\n
\nUploads a part by copying data from an existing object as data source. To specify the\n data source, you add the request header x-amz-copy-source
in your request. To specify \n a byte range, you add the request header x-amz-copy-source-range
in your\n request.
For information about maximum and minimum part sizes and other multipart upload\n specifications, see Multipart upload limits in the Amazon S3 User Guide.
\nInstead of copying data from an existing object as part data, you might use the UploadPart\n action to upload new data as a part of an object in your request.
\nYou must initiate a multipart upload before you can upload any part. In response to your\n initiate request, Amazon S3 returns the upload ID, a unique identifier that you must include in\n your upload part request.
\nFor conceptual information about multipart uploads, see Uploading\n Objects Using Multipart Upload in the\n Amazon S3 User Guide. For information about copying objects using a single atomic action vs. a multipart\n upload, see Operations on Objects in\n the Amazon S3 User Guide.
\n\n Directory buckets - For directory buckets, you must make requests for this API operation to the Zonal endpoint. These endpoints support virtual-hosted-style requests in the format https://bucket_name.s3express-az_id.region.amazonaws.com/key-name\n
. Path-style requests are not supported. For more information, see Regional and Zonal endpoints in the\n Amazon S3 User Guide.
All UploadPartCopy
requests must be authenticated and signed by using IAM credentials (access key ID and secret access key for the IAM identities). All headers with the x-amz-
prefix, including\n x-amz-copy-source
, must be signed. For more information, see REST Authentication.
\n Directory buckets - You must use IAM credentials to authenticate and authorize your access to the UploadPartCopy
API operation, instead of using the \n temporary security credentials through the CreateSession
API operation.
Amazon Web Services CLI or SDKs handles authentication and authorization on your behalf.
\nYou must have READ
access to the source object and WRITE
\n access to the destination bucket.
\n General purpose bucket permissions - You\n must have the permissions in a policy based on the bucket types of your\n source bucket and destination bucket in an UploadPartCopy
\n operation.
If the source object is in a general purpose bucket, you must have the\n \n s3:GetObject
\n \n permission to read the source object that is being copied.
If the destination bucket is a general purpose bucket, you must have the\n \n s3:PutObject
\n \n permission to write the object copy to the destination bucket.
To perform a multipart upload with encryption using an Key Management Service\n key, the requester must have permission to the\n kms:Decrypt
and kms:GenerateDataKey
\n actions on the key. The requester must also have permissions for the\n kms:GenerateDataKey
action for the\n CreateMultipartUpload
API. Then, the requester needs\n permissions for the kms:Decrypt
action on the\n UploadPart
and UploadPartCopy
APIs. These\n permissions are required because Amazon S3 must decrypt and read data from\n the encrypted file parts before it completes the multipart upload. For\n more information about KMS permissions, see Protecting\n data using server-side encryption with KMS in the\n Amazon S3 User Guide. For information about the\n permissions required to use the multipart upload API, see Multipart upload\n and permissions and Multipart upload API and permissions in the\n Amazon S3 User Guide.
\n Directory bucket permissions -\n You must have permissions in a bucket policy or an IAM identity-based policy based on the source and destination\n bucket types in an UploadPartCopy
operation.
If the source object that you want to copy is in a\n directory bucket, you must have the \n s3express:CreateSession
\n permission in\n the Action
element of a policy to read the object. By\n default, the session is in the ReadWrite
mode. If you\n want to restrict the access, you can explicitly set the\n s3express:SessionMode
condition key to\n ReadOnly
on the copy source bucket.
If the copy destination is a directory bucket, you must have the \n \n s3express:CreateSession
\n permission in the\n Action
element of a policy to write the object\n to the destination. The s3express:SessionMode
condition\n key cannot be set to ReadOnly
on the copy destination.
If the object is encrypted with\n SSE-KMS, you must also have the\n kms:GenerateDataKey
and kms:Decrypt
permissions in IAM identity-based policies and KMS key policies for the KMS key.
For example policies, see Example bucket policies for S3 Express One Zone and Amazon Web Services Identity and Access Management (IAM) identity-based policies for S3 Express One Zone in the\n Amazon S3 User Guide.
\n\n General purpose buckets - \n \n For information about using server-side encryption with customer-provided\n encryption keys with the UploadPartCopy
operation, see CopyObject and UploadPart.\n
\n Directory buckets - For directory buckets, there are only two supported options for server-side encryption: server-side encryption with Amazon S3 managed keys (SSE-S3) (AES256
) and server-side encryption with KMS keys (SSE-KMS) (aws:kms
). For more\n information, see Protecting data with server-side encryption in the Amazon S3 User Guide.
For directory buckets, when you perform a CreateMultipartUpload
operation and an UploadPartCopy
operation, \n the request headers you provide in the CreateMultipartUpload
request must match the default encryption configuration of the destination bucket.
S3 Bucket Keys aren't supported, when you copy SSE-KMS encrypted objects from general purpose buckets \nto directory buckets, from directory buckets to general purpose buckets, or between directory buckets, through UploadPartCopy. In this case, Amazon S3 makes a call to KMS every time a copy request is made for a KMS-encrypted object.
\nError Code: NoSuchUpload
\n
Description: The specified multipart upload does not exist. The\n upload ID might be invalid, or the multipart upload might have been\n aborted or completed.
\nHTTP Status Code: 404 Not Found
\nError Code: InvalidRequest
\n
Description: The specified copy source is not supported as a\n byte-range copy source.
\nHTTP Status Code: 400 Bad Request
\n\n Directory buckets - The HTTP Host header syntax is \n Bucket_name.s3express-az_id.region.amazonaws.com
.
The following operations are related to UploadPartCopy
:
\n UploadPart\n
\n\n AbortMultipartUpload\n
\n\n ListParts\n
\n\n ListMultipartUploads\n
\nThe server-side encryption algorithm used when you store this object in Amazon S3 (for example,\n AES256
, aws:kms
).
For directory buckets, only server-side encryption with Amazon S3 managed keys (SSE-S3) (AES256
) is supported.
The server-side encryption algorithm used when you store this object in Amazon S3 (for example,\n AES256
, aws:kms
).
If present, indicates the ID of the Key Management Service (KMS) symmetric encryption customer managed key\n that was used for the object.
\nThis functionality is not supported for directory buckets.
\nIf present, indicates the ID of the KMS key that was used for object encryption.
", "smithy.api#httpHeader": "x-amz-server-side-encryption-aws-kms-key-id" } }, "BucketKeyEnabled": { "target": "com.amazonaws.s3#BucketKeyEnabled", "traits": { - "smithy.api#documentation": "Indicates whether the multipart upload uses an S3 Bucket Key for server-side encryption\n with Key Management Service (KMS) keys (SSE-KMS).
\nThis functionality is not supported for directory buckets.
\nIndicates whether the multipart upload uses an S3 Bucket Key for server-side encryption\n with Key Management Service (KMS) keys (SSE-KMS).
", "smithy.api#httpHeader": "x-amz-server-side-encryption-bucket-key-enabled" } }, @@ -34015,7 +34071,7 @@ "ServerSideEncryption": { "target": "com.amazonaws.s3#ServerSideEncryption", "traits": { - "smithy.api#documentation": "The server-side encryption algorithm used when you store this object in Amazon S3 (for example,\n AES256
, aws:kms
).
For directory buckets, only server-side encryption with Amazon S3 managed keys (SSE-S3) (AES256
) is supported.
The server-side encryption algorithm used when you store this object in Amazon S3 (for example,\n AES256
, aws:kms
).
If present, indicates the ID of the Key Management Service (KMS) symmetric encryption customer managed key\n that was used for the object.
\nThis functionality is not supported for directory buckets.
\nIf present, indicates the ID of the KMS key that was used for object encryption.
", "smithy.api#httpHeader": "x-amz-server-side-encryption-aws-kms-key-id" } }, "BucketKeyEnabled": { "target": "com.amazonaws.s3#BucketKeyEnabled", "traits": { - "smithy.api#documentation": "Indicates whether the multipart upload uses an S3 Bucket Key for server-side encryption\n with Key Management Service (KMS) keys (SSE-KMS).
\nThis functionality is not supported for directory buckets.
\nIndicates whether the multipart upload uses an S3 Bucket Key for server-side encryption\n with Key Management Service (KMS) keys (SSE-KMS).
", "smithy.api#httpHeader": "x-amz-server-side-encryption-bucket-key-enabled" } }, diff --git a/examples/cross_service/detect_faces/Cargo.toml b/examples/cross_service/detect_faces/Cargo.toml index 64a592614493..696455954ec5 100644 --- a/examples/cross_service/detect_faces/Cargo.toml +++ b/examples/cross_service/detect_faces/Cargo.toml @@ -8,7 +8,7 @@ publish = false [dependencies] aws-config= { version = "1.5.6", path = "../../../sdk/aws-config" } aws-sdk-rekognition= { version = "1.46.0", path = "../../../sdk/rekognition" } -aws-sdk-s3= { version = "1.50.0", path = "../../../sdk/s3" } +aws-sdk-s3= { version = "1.51.0", path = "../../../sdk/s3" } aws-smithy-types= { version = "1.2.6", path = "../../../sdk/aws-smithy-types", features = ["rt-tokio"] } [dependencies.tokio] diff --git a/examples/cross_service/detect_labels/Cargo.toml b/examples/cross_service/detect_labels/Cargo.toml index f2708b94c423..9f9f57fbaf50 100644 --- a/examples/cross_service/detect_labels/Cargo.toml +++ b/examples/cross_service/detect_labels/Cargo.toml @@ -10,7 +10,7 @@ kamadak-exif = "0.5.4" aws-config= { version = "1.5.6", path = "../../../sdk/aws-config" } aws-sdk-dynamodb= { version = "1.46.0", path = "../../../sdk/dynamodb" } aws-sdk-rekognition= { version = "1.46.0", path = "../../../sdk/rekognition" } -aws-sdk-s3= { version = "1.50.0", path = "../../../sdk/s3" } +aws-sdk-s3= { version = "1.51.0", path = "../../../sdk/s3" } aws-smithy-types= { version = "1.2.6", path = "../../../sdk/aws-smithy-types", features = ["rt-tokio"] } aws-types= { version = "1.3.3", path = "../../../sdk/aws-types" } diff --git a/examples/cross_service/photo_asset_management/Cargo.toml b/examples/cross_service/photo_asset_management/Cargo.toml index c339e7e3c69c..f0d38e98f525 100644 --- a/examples/cross_service/photo_asset_management/Cargo.toml +++ b/examples/cross_service/photo_asset_management/Cargo.toml @@ -33,7 +33,7 @@ tracing = "0.1.37" aws-config= { version = "1.5.6", path = "../../../sdk/aws-config" } aws-sdk-dynamodb= { version = "1.46.0", path = "../../../sdk/dynamodb" } aws-sdk-rekognition= { version = "1.46.0", path = "../../../sdk/rekognition" } -aws-sdk-s3= { version = "1.50.0", path = "../../../sdk/s3" } +aws-sdk-s3= { version = "1.51.0", path = "../../../sdk/s3" } aws-sdk-sns= { version = "1.44.0", path = "../../../sdk/sns" } aws-smithy-runtime= { version = "1.7.1", path = "../../../sdk/aws-smithy-runtime" } aws-smithy-types-convert= { version = "0.60.8", path = "../../../sdk/aws-smithy-types-convert", features = ["convert-chrono"] } diff --git a/examples/cross_service/telephone/Cargo.toml b/examples/cross_service/telephone/Cargo.toml index f9352402c3fd..8889985d9236 100644 --- a/examples/cross_service/telephone/Cargo.toml +++ b/examples/cross_service/telephone/Cargo.toml @@ -13,7 +13,7 @@ reqwest = "0.11.4" serde_json = "1.0" aws-config= { version = "1.5.6", path = "../../../sdk/aws-config" } aws-sdk-polly= { version = "1.45.0", path = "../../../sdk/polly" } -aws-sdk-s3= { version = "1.50.0", path = "../../../sdk/s3" } +aws-sdk-s3= { version = "1.51.0", path = "../../../sdk/s3" } aws-sdk-transcribe= { version = "1.44.0", path = "../../../sdk/transcribe" } aws-smithy-types= { version = "1.2.6", path = "../../../sdk/aws-smithy-types", features = ["rt-tokio"] } diff --git a/examples/examples/aurora/Cargo.toml b/examples/examples/aurora/Cargo.toml index a5db141c4ba8..86e5abcfd1e5 100644 --- a/examples/examples/aurora/Cargo.toml +++ b/examples/examples/aurora/Cargo.toml @@ -16,7 +16,7 @@ tracing = "0.1.37" aws-config= { version = "1.5.6", path = "../../../sdk/aws-config", features = ["behavior-version-latest"] } aws-smithy-types= { version = "1.2.6", path = "../../../sdk/aws-smithy-types" } aws-smithy-runtime-api= { version = "1.7.2", path = "../../../sdk/aws-smithy-runtime-api" } -aws-sdk-rds= { version = "1.55.0", path = "../../../sdk/rds" } +aws-sdk-rds= { version = "1.55.1", path = "../../../sdk/rds" } [dependencies.phf] version = "0.11.2" diff --git a/examples/examples/concurrency/Cargo.toml b/examples/examples/concurrency/Cargo.toml index 5afb89b2b040..47c2d931339d 100644 --- a/examples/examples/concurrency/Cargo.toml +++ b/examples/examples/concurrency/Cargo.toml @@ -26,5 +26,5 @@ features = ["env-filter"] [dev-dependencies] fastrand = "1.8.0" aws-config= { version = "1.5.6", path = "../../../sdk/aws-config" } -aws-sdk-s3= { version = "1.50.0", path = "../../../sdk/s3" } +aws-sdk-s3= { version = "1.51.0", path = "../../../sdk/s3" } aws-sdk-sqs= { version = "1.43.0", path = "../../../sdk/sqs" } diff --git a/examples/examples/custom-root-certificates/Cargo.toml b/examples/examples/custom-root-certificates/Cargo.toml index 8041075f2ad3..bdbe07aec454 100644 --- a/examples/examples/custom-root-certificates/Cargo.toml +++ b/examples/examples/custom-root-certificates/Cargo.toml @@ -10,7 +10,7 @@ publish = false [dependencies] rustls = "0.21.9" aws-config= { version = "1.5.6", path = "../../../sdk/aws-config", features = ["behavior-version-latest"] } -aws-sdk-s3= { version = "1.50.0", path = "../../../sdk/s3", default-features = false } +aws-sdk-s3= { version = "1.51.0", path = "../../../sdk/s3", default-features = false } aws-smithy-runtime= { version = "1.7.1", path = "../../../sdk/aws-smithy-runtime", features = ["tls-rustls"] } [dependencies.tokio] diff --git a/examples/examples/glue/Cargo.toml b/examples/examples/glue/Cargo.toml index 7fa926529153..3f9b21e70d8a 100644 --- a/examples/examples/glue/Cargo.toml +++ b/examples/examples/glue/Cargo.toml @@ -20,7 +20,7 @@ futures = "0.3.25" tracing-bunyan-formatter = "0.3.4" aws-config= { version = "1.5.6", path = "../../../sdk/aws-config", features = ["behavior-version-latest"] } aws-sdk-glue= { version = "1.61.0", path = "../../../sdk/glue" } -aws-sdk-s3= { version = "1.50.0", path = "../../../sdk/s3" } +aws-sdk-s3= { version = "1.51.0", path = "../../../sdk/s3" } aws-http= { version = "0.60.6", path = "../../../sdk/aws-http" } aws-smithy-types= { version = "1.2.6", path = "../../../sdk/aws-smithy-types" } aws-types= { version = "1.3.3", path = "../../../sdk/aws-types" } diff --git a/examples/examples/iam/Cargo.toml b/examples/examples/iam/Cargo.toml index cb907bd94898..5bbd2a659dfc 100644 --- a/examples/examples/iam/Cargo.toml +++ b/examples/examples/iam/Cargo.toml @@ -24,7 +24,7 @@ lazy_static = "1.4.0" aws-config= { version = "1.5.6", path = "../../../sdk/aws-config", features = ["behavior-version-latest"] } aws-credential-types= { version = "1.2.1", path = "../../../sdk/aws-credential-types", features = ["hardcoded-credentials"] } aws-sdk-iam= { version = "1.45.0", path = "../../../sdk/iam" } -aws-sdk-s3= { version = "1.50.0", path = "../../../sdk/s3" } +aws-sdk-s3= { version = "1.51.0", path = "../../../sdk/s3" } aws-sdk-sts= { version = "1.43.0", path = "../../../sdk/sts" } [dependencies.sdk-examples-test-utils] diff --git a/examples/examples/lambda/Cargo.toml b/examples/examples/lambda/Cargo.toml index 387730c54994..c4cb79d098df 100644 --- a/examples/examples/lambda/Cargo.toml +++ b/examples/examples/lambda/Cargo.toml @@ -15,7 +15,7 @@ aws-config= { version = "1.5.6", path = "../../../sdk/aws-config", features = [" aws-sdk-ec2= { version = "1.73.0", path = "../../../sdk/ec2" } aws-sdk-iam= { version = "1.45.0", path = "../../../sdk/iam" } aws-sdk-lambda= { version = "1.46.0", path = "../../../sdk/lambda" } -aws-sdk-s3= { version = "1.50.0", path = "../../../sdk/s3" } +aws-sdk-s3= { version = "1.51.0", path = "../../../sdk/s3" } aws-smithy-types= { version = "1.2.6", path = "../../../sdk/aws-smithy-types" } aws-types= { version = "1.3.3", path = "../../../sdk/aws-types" } diff --git a/examples/examples/localstack/Cargo.toml b/examples/examples/localstack/Cargo.toml index 5ff860f63800..78abf3ca0b52 100644 --- a/examples/examples/localstack/Cargo.toml +++ b/examples/examples/localstack/Cargo.toml @@ -7,7 +7,7 @@ publish = false [dependencies] aws-config= { version = "1.5.6", path = "../../../sdk/aws-config", features = ["behavior-version-latest"] } -aws-sdk-s3= { version = "1.50.0", path = "../../../sdk/s3" } +aws-sdk-s3= { version = "1.51.0", path = "../../../sdk/s3" } [dependencies.tokio] version = "1" diff --git a/examples/examples/rds/Cargo.toml b/examples/examples/rds/Cargo.toml index 1043cfe36105..c5724c37a91b 100644 --- a/examples/examples/rds/Cargo.toml +++ b/examples/examples/rds/Cargo.toml @@ -16,7 +16,7 @@ secrecy = "0.8.0" tracing = "0.1.37" aws-config= { version = "1.5.6", path = "../../../sdk/aws-config", features = ["behavior-version-latest"] } aws-smithy-runtime-api= { version = "1.7.2", path = "../../../sdk/aws-smithy-runtime-api" } -aws-sdk-rds= { version = "1.55.0", path = "../../../sdk/rds" } +aws-sdk-rds= { version = "1.55.1", path = "../../../sdk/rds" } [dependencies.clap] version = "~4.4" diff --git a/examples/examples/s3/Cargo.toml b/examples/examples/s3/Cargo.toml index cc697bf388c4..903ff9f3f948 100644 --- a/examples/examples/s3/Cargo.toml +++ b/examples/examples/s3/Cargo.toml @@ -24,7 +24,7 @@ tracing = "0.1.37" serde_json = "1" chrono = "0.4.38" aws-config= { version = "1.5.6", path = "../../../sdk/aws-config", features = ["behavior-version-latest"] } -aws-sdk-s3= { version = "1.50.0", path = "../../../sdk/s3", features = ["rt-tokio"] } +aws-sdk-s3= { version = "1.51.0", path = "../../../sdk/s3", features = ["rt-tokio"] } aws-smithy-runtime= { version = "1.7.1", path = "../../../sdk/aws-smithy-runtime" } aws-smithy-runtime-api= { version = "1.7.2", path = "../../../sdk/aws-smithy-runtime-api", features = ["client"] } aws-smithy-types= { version = "1.2.6", path = "../../../sdk/aws-smithy-types", features = ["http-body-0-4-x"] } diff --git a/examples/examples/sdk-config/Cargo.toml b/examples/examples/sdk-config/Cargo.toml index 268bee5e96d7..65deaf5d7509 100644 --- a/examples/examples/sdk-config/Cargo.toml +++ b/examples/examples/sdk-config/Cargo.toml @@ -12,7 +12,7 @@ async_once = "0.2.6" tracing = "0.1.37" aws-config= { version = "1.5.6", path = "../../../sdk/aws-config", features = ["behavior-version-latest"] } aws-credential-types= { version = "1.2.1", path = "../../../sdk/aws-credential-types" } -aws-sdk-s3= { version = "1.50.0", path = "../../../sdk/s3" } +aws-sdk-s3= { version = "1.51.0", path = "../../../sdk/s3" } [dependencies.tokio] version = "1.20.1" diff --git a/examples/examples/sending-presigned-requests/Cargo.toml b/examples/examples/sending-presigned-requests/Cargo.toml index 00e220ffead5..d4645a2c940f 100644 --- a/examples/examples/sending-presigned-requests/Cargo.toml +++ b/examples/examples/sending-presigned-requests/Cargo.toml @@ -12,7 +12,7 @@ hyper = "0.14" reqwest = "0.11" hyper-tls = "0.5.0" aws-config= { version = "1.5.6", path = "../../../sdk/aws-config", features = ["behavior-version-latest"] } -aws-sdk-s3= { version = "1.50.0", path = "../../../sdk/s3" } +aws-sdk-s3= { version = "1.51.0", path = "../../../sdk/s3" } [dependencies.clap] version = "~4.4" diff --git a/examples/examples/testing/Cargo.toml b/examples/examples/testing/Cargo.toml index a4eda7d6f8bf..c0dcdbfe5061 100644 --- a/examples/examples/testing/Cargo.toml +++ b/examples/examples/testing/Cargo.toml @@ -17,7 +17,7 @@ mockall = "0.11.4" serde_json = "1" aws-config= { version = "1.5.6", path = "../../../sdk/aws-config", features = ["behavior-version-latest"] } aws-credential-types= { version = "1.2.1", path = "../../../sdk/aws-credential-types", features = ["hardcoded-credentials"] } -aws-sdk-s3= { version = "1.50.0", path = "../../../sdk/s3" } +aws-sdk-s3= { version = "1.51.0", path = "../../../sdk/s3" } aws-smithy-types= { version = "1.2.6", path = "../../../sdk/aws-smithy-types" } aws-smithy-runtime= { version = "1.7.1", path = "../../../sdk/aws-smithy-runtime", features = ["test-util"] } aws-smithy-runtime-api= { version = "1.7.2", path = "../../../sdk/aws-smithy-runtime-api", features = ["test-util"] } diff --git a/examples/lambda/calculator/Cargo.toml b/examples/lambda/calculator/Cargo.toml index e30fd8d5dcbb..3938bb46eaf3 100644 --- a/examples/lambda/calculator/Cargo.toml +++ b/examples/lambda/calculator/Cargo.toml @@ -16,7 +16,7 @@ serde = "1.0.164" aws-config= { version = "1.5.6", path = "../../../sdk/aws-config" } aws-sdk-ec2= { version = "1.73.0", path = "../../../sdk/ec2" } aws-sdk-lambda= { version = "1.46.0", path = "../../../sdk/lambda" } -aws-sdk-s3= { version = "1.50.0", path = "../../../sdk/s3" } +aws-sdk-s3= { version = "1.51.0", path = "../../../sdk/s3" } [dependencies.clap] version = "~4.4" diff --git a/index.md b/index.md index 0fce12c9c8eb..ca5f9063cbd2 100644 --- a/index.md +++ b/index.md @@ -56,6 +56,7 @@ The AWS SDK for Rust contains one crate for each AWS service, as well as [aws-co | AWS Device Farm | [aws-sdk-devicefarm](https://crates.io/crates/aws-sdk-devicefarm) ([docs](https://docs.rs/aws-sdk-devicefarm)) | | AWS Direct Connect | [aws-sdk-directconnect](https://crates.io/crates/aws-sdk-directconnect) ([docs](https://docs.rs/aws-sdk-directconnect)) | | AWS Directory Service | [aws-sdk-directory](https://crates.io/crates/aws-sdk-directory) ([docs](https://docs.rs/aws-sdk-directory)) | +| AWS Directory Service Data | [aws-sdk-directoryservicedata](https://crates.io/crates/aws-sdk-directoryservicedata) ([docs](https://docs.rs/aws-sdk-directoryservicedata)) | | AWS EC2 Instance Connect | [aws-sdk-ec2instanceconnect](https://crates.io/crates/aws-sdk-ec2instanceconnect) ([docs](https://docs.rs/aws-sdk-ec2instanceconnect)) | | AWS Elastic Beanstalk | [aws-sdk-elasticbeanstalk](https://crates.io/crates/aws-sdk-elasticbeanstalk) ([docs](https://docs.rs/aws-sdk-elasticbeanstalk)) | | AWS Elemental MediaConvert | [aws-sdk-mediaconvert](https://crates.io/crates/aws-sdk-mediaconvert) ([docs](https://docs.rs/aws-sdk-mediaconvert)) | diff --git a/sdk/costexplorer/Cargo.toml b/sdk/costexplorer/Cargo.toml index 6ab3ce83b047..756d4b2b0fcf 100644 --- a/sdk/costexplorer/Cargo.toml +++ b/sdk/costexplorer/Cargo.toml @@ -1,7 +1,7 @@ # Code generated by software.amazon.smithy.rust.codegen.smithy-rs. DO NOT EDIT. [package] name = "aws-sdk-costexplorer" -version = "1.45.0" +version = "1.46.0" authors = ["AWS Rust SDK TeamThe value that you want to search the filter values for.
The start date and end date for retrieving the dimension values. The start date is inclusive, but the end date is exclusive. For example, if start
is 2017-01-01
and end
is 2017-05-01
, then the cost and usage data is retrieved from 2017-01-01
up to and including 2017-04-30
but not including 2017-05-01
.
The name of the dimension. Each Dimension
is available for a different Context
. For more information, see Context
. LINK_ACCOUNT_NAME
and SERVICE_CODE
can only be used in CostCategoryRule.
The context for the call to GetDimensionValues
. This can be RESERVATIONS
or COST_AND_USAGE
. The default value is COST_AND_USAGE
. If the context is set to RESERVATIONS
, the resulting dimension values can be used in the GetReservationUtilization
operation. If the context is set to COST_AND_USAGE
, the resulting dimension values can be used in the GetCostAndUsage
operation.
If you set the context to COST_AND_USAGE
, you can use the following dimensions for searching:
AZ - The Availability Zone. An example is us-east-1a
.
BILLING_ENTITY - The Amazon Web Services seller that your account is with. Possible values are the following:
- Amazon Web Services(Amazon Web Services): The entity that sells Amazon Web Services.
- AISPL (Amazon Internet Services Pvt. Ltd.): The local Indian entity that's an acting reseller for Amazon Web Services in India.
- Amazon Web Services Marketplace: The entity that supports the sale of solutions that are built on Amazon Web Services by third-party software providers.
CACHE_ENGINE - The Amazon ElastiCache operating system. Examples are Windows or Linux.
DEPLOYMENT_OPTION - The scope of Amazon Relational Database Service deployments. Valid values are SingleAZ
and MultiAZ
.
DATABASE_ENGINE - The Amazon Relational Database Service database. Examples are Aurora or MySQL.
INSTANCE_TYPE - The type of Amazon EC2 instance. An example is m4.xlarge
.
INSTANCE_TYPE_FAMILY - A family of instance types optimized to fit different use cases. Examples are Compute Optimized
(for example, C4
, C5
, C6g
, and C7g
), Memory Optimization
(for example, R4
, R5n
, R5b
, and R6g
).
INVOICING_ENTITY - The name of the entity that issues the Amazon Web Services invoice.
LEGAL_ENTITY_NAME - The name of the organization that sells you Amazon Web Services services, such as Amazon Web Services.
LINKED_ACCOUNT - The description in the attribute map that includes the full name of the member account. The value field contains the Amazon Web Services ID of the member account.
OPERATING_SYSTEM - The operating system. Examples are Windows or Linux.
OPERATION - The action performed. Examples include RunInstance
and CreateBucket
.
PLATFORM - The Amazon EC2 operating system. Examples are Windows or Linux.
PURCHASE_TYPE - The reservation type of the purchase that this usage is related to. Examples include On-Demand Instances and Standard Reserved Instances.
RESERVATION_ID - The unique identifier for an Amazon Web Services Reservation Instance.
SAVINGS_PLAN_ARN - The unique identifier for your Savings Plans.
SAVINGS_PLANS_TYPE - Type of Savings Plans (EC2 Instance or Compute).
SERVICE - The Amazon Web Services service such as Amazon DynamoDB.
TENANCY - The tenancy of a resource. Examples are shared or dedicated.
USAGE_TYPE - The type of usage. An example is DataTransfer-In-Bytes. The response for the GetDimensionValues
operation includes a unit attribute. Examples include GB and Hrs.
USAGE_TYPE_GROUP - The grouping of common usage types. An example is Amazon EC2: CloudWatch – Alarms. The response for this operation includes a unit attribute.
REGION - The Amazon Web Services Region.
RECORD_TYPE - The different types of charges such as Reserved Instance (RI) fees, usage costs, tax refunds, and credits.
RESOURCE_ID - The unique identifier of the resource. ResourceId is an opt-in feature only available for last 14 days for EC2-Compute Service.
If you set the context to RESERVATIONS
, you can use the following dimensions for searching:
AZ - The Availability Zone. An example is us-east-1a
.
CACHE_ENGINE - The Amazon ElastiCache operating system. Examples are Windows or Linux.
DEPLOYMENT_OPTION - The scope of Amazon Relational Database Service deployments. Valid values are SingleAZ
and MultiAZ
.
INSTANCE_TYPE - The type of Amazon EC2 instance. An example is m4.xlarge
.
LINKED_ACCOUNT - The description in the attribute map that includes the full name of the member account. The value field contains the Amazon Web Services ID of the member account.
PLATFORM - The Amazon EC2 operating system. Examples are Windows or Linux.
REGION - The Amazon Web Services Region.
SCOPE (Utilization only) - The scope of a Reserved Instance (RI). Values are regional or a single Availability Zone.
TAG (Coverage only) - The tags that are associated with a Reserved Instance (RI).
TENANCY - The tenancy of a resource. Examples are shared or dedicated.
If you set the context to SAVINGS_PLANS
, you can use the following dimensions for searching:
SAVINGS_PLANS_TYPE - Type of Savings Plans (EC2 Instance or Compute)
PAYMENT_OPTION - The payment option for the given Savings Plans (for example, All Upfront)
REGION - The Amazon Web Services Region.
INSTANCE_TYPE_FAMILY - The family of instances (For example, m5
)
LINKED_ACCOUNT - The description in the attribute map that includes the full name of the member account. The value field contains the Amazon Web Services ID of the member account.
SAVINGS_PLAN_ARN - The unique identifier for your Savings Plans.
The context for the call to GetDimensionValues
. This can be RESERVATIONS
or COST_AND_USAGE
. The default value is COST_AND_USAGE
. If the context is set to RESERVATIONS
, the resulting dimension values can be used in the GetReservationUtilization
operation. If the context is set to COST_AND_USAGE
, the resulting dimension values can be used in the GetCostAndUsage
operation.
If you set the context to COST_AND_USAGE
, you can use the following dimensions for searching:
AZ - The Availability Zone. An example is us-east-1a
.
BILLING_ENTITY - The Amazon Web Services seller that your account is with. Possible values are the following:
- Amazon Web Services(Amazon Web Services): The entity that sells Amazon Web Servicesservices.
- AISPL (Amazon Internet Services Pvt. Ltd.): The local Indian entity that's an acting reseller for Amazon Web Servicesservices in India.
- Amazon Web Services Marketplace: The entity that supports the sale of solutions that are built on Amazon Web Services by third-party software providers.
CACHE_ENGINE - The Amazon ElastiCache operating system. Examples are Windows or Linux.
DEPLOYMENT_OPTION - The scope of Amazon Relational Database Service deployments. Valid values are SingleAZ
and MultiAZ
.
DATABASE_ENGINE - The Amazon Relational Database Service database. Examples are Aurora or MySQL.
INSTANCE_TYPE - The type of Amazon EC2 instance. An example is m4.xlarge
.
INSTANCE_TYPE_FAMILY - A family of instance types optimized to fit different use cases. Examples are Compute Optimized
(for example, C4
, C5
, C6g
, and C7g
), Memory Optimization
(for example, R4
, R5n
, R5b
, and R6g
).
INVOICING_ENTITY - The name of the entity that issues the Amazon Web Services invoice.
LEGAL_ENTITY_NAME - The name of the organization that sells you Amazon Web Services services, such as Amazon Web Services.
LINKED_ACCOUNT - The description in the attribute map that includes the full name of the member account. The value field contains the Amazon Web Services ID of the member account.
OPERATING_SYSTEM - The operating system. Examples are Windows or Linux.
OPERATION - The action performed. Examples include RunInstance
and CreateBucket
.
PLATFORM - The Amazon EC2 operating system. Examples are Windows or Linux.
PURCHASE_TYPE - The reservation type of the purchase that this usage is related to. Examples include On-Demand Instances and Standard Reserved Instances.
RESERVATION_ID - The unique identifier for an Amazon Web Services Reservation Instance.
SAVINGS_PLAN_ARN - The unique identifier for your Savings Plans.
SAVINGS_PLANS_TYPE - Type of Savings Plans (EC2 Instance or Compute).
SERVICE - The Amazon Web Services service such as Amazon DynamoDB.
TENANCY - The tenancy of a resource. Examples are shared or dedicated.
USAGE_TYPE - The type of usage. An example is DataTransfer-In-Bytes. The response for the GetDimensionValues
operation includes a unit attribute. Examples include GB and Hrs.
USAGE_TYPE_GROUP - The grouping of common usage types. An example is Amazon EC2: CloudWatch – Alarms. The response for this operation includes a unit attribute.
REGION - The Amazon Web Services Region.
RECORD_TYPE - The different types of charges such as Reserved Instance (RI) fees, usage costs, tax refunds, and credits.
RESOURCE_ID - The unique identifier of the resource. ResourceId is an opt-in feature only available for last 14 days for EC2-Compute Service.
If you set the context to RESERVATIONS
, you can use the following dimensions for searching:
AZ - The Availability Zone. An example is us-east-1a
.
CACHE_ENGINE - The Amazon ElastiCache operating system. Examples are Windows or Linux.
DEPLOYMENT_OPTION - The scope of Amazon Relational Database Service deployments. Valid values are SingleAZ
and MultiAZ
.
INSTANCE_TYPE - The type of Amazon EC2 instance. An example is m4.xlarge
.
LINKED_ACCOUNT - The description in the attribute map that includes the full name of the member account. The value field contains the Amazon Web Services ID of the member account.
PLATFORM - The Amazon EC2 operating system. Examples are Windows or Linux.
REGION - The Amazon Web Services Region.
SCOPE (Utilization only) - The scope of a Reserved Instance (RI). Values are regional or a single Availability Zone.
TAG (Coverage only) - The tags that are associated with a Reserved Instance (RI).
TENANCY - The tenancy of a resource. Examples are shared or dedicated.
If you set the context to SAVINGS_PLANS
, you can use the following dimensions for searching:
SAVINGS_PLANS_TYPE - Type of Savings Plans (EC2 Instance or Compute)
PAYMENT_OPTION - The payment option for the given Savings Plans (for example, All Upfront)
REGION - The Amazon Web Services Region.
INSTANCE_TYPE_FAMILY - The family of instances (For example, m5
)
LINKED_ACCOUNT - The description in the attribute map that includes the full name of the member account. The value field contains the Amazon Web Services ID of the member account.
SAVINGS_PLAN_ARN - The unique identifier for your Savings Plans.
Use Expression
to filter in various Cost Explorer APIs.
Not all Expression
types are supported in each API. Refer to the documentation for each specific API to see what is supported.
There are two patterns:
Simple dimension values.
There are three types of simple dimension values: CostCategories
, Tags
, and Dimensions
.
Specify the CostCategories
field to define a filter that acts on Cost Categories.
Specify the Tags
field to define a filter that acts on Cost Allocation Tags.
Specify the Dimensions
field to define a filter that acts on the DimensionValues
.
For each filter type, you can set the dimension name and values for the filters that you plan to use.
For example, you can filter for REGION==us-east-1 OR REGION==us-west-1
. For GetRightsizingRecommendation
, the Region is a full name (for example, REGION==US East (N. Virginia)
.
The corresponding Expression
for this example is as follows: { "Dimensions": { "Key": "REGION", "Values": \[ "us-east-1", "us-west-1" \] } }
As shown in the previous example, lists of dimension values are combined with OR
when applying the filter.
You can also set different match options to further control how the filter behaves. Not all APIs support match options. Refer to the documentation for each specific API to see what is supported.
For example, you can filter for linked account names that start with "a".
The corresponding Expression
for this example is as follows: { "Dimensions": { "Key": "LINKED_ACCOUNT_NAME", "MatchOptions": \[ "STARTS_WITH" \], "Values": \[ "a" \] } }
Compound Expression
types with logical operations.
You can use multiple Expression
types and the logical operators AND/OR/NOT
to create a list of one or more Expression
objects. By doing this, you can filter by more advanced options.
For example, you can filter by ((REGION == us-east-1 OR REGION == us-west-1) OR (TAG.Type == Type1)) AND (USAGE_TYPE != DataTransfer)
.
The corresponding Expression
for this example is as follows: { "And": \[ {"Or": \[ {"Dimensions": { "Key": "REGION", "Values": \[ "us-east-1", "us-west-1" \] }}, {"Tags": { "Key": "TagName", "Values": \["Value1"\] } } \]}, {"Not": {"Dimensions": { "Key": "USAGE_TYPE", "Values": \["DataTransfer"\] }}} \] }
Because each Expression
can have only one operator, the service returns an error if more than one is specified. The following example shows an Expression
object that creates an error: { "And": \[ ... \], "Dimensions": { "Key": "USAGE_TYPE", "Values": \[ "DataTransfer" \] } }
The following is an example of the corresponding error message: "Expression has more than one roots. Only one root operator is allowed for each expression: And, Or, Not, Dimensions, Tags, CostCategories"
For the GetRightsizingRecommendation
action, a combination of OR and NOT isn't supported. OR isn't supported between different dimensions, or dimensions and tags. NOT operators aren't supported. Dimensions are also limited to LINKED_ACCOUNT
, REGION
, or RIGHTSIZING_TYPE
.
For the GetReservationPurchaseRecommendation
action, only NOT is supported. AND and OR aren't supported. Dimensions are limited to LINKED_ACCOUNT
.
The value that you want to sort the data by.
The key represents cost and usage metrics. The following values are supported:
BlendedCost
UnblendedCost
AmortizedCost
NetAmortizedCost
NetUnblendedCost
UsageQuantity
NormalizedUsageAmount
The supported values for the SortOrder
key are ASCENDING
or DESCENDING
.
When you specify a SortBy
paramater, the context must be COST_AND_USAGE
. Further, when using SortBy
, NextPageToken
and SearchString
aren't supported.
This field is only used when SortBy is provided in the request. The maximum number of objects that are returned for this request. If MaxResults isn't specified with SortBy, the request returns 1000 results as the default value for this parameter.
For GetDimensionValues
, MaxResults has an upper limit of 1000.
AZ - The Availability Zone. An example is us-east-1a
.
BILLING_ENTITY - The Amazon Web Services seller that your account is with. Possible values are the following:
- ///- Amazon Web Services(Amazon Web Services): The entity that sells Amazon Web Services.
- ///- AISPL (Amazon Internet Services Pvt. Ltd.): The local Indian entity that's an acting reseller for Amazon Web Services in India.
+ ///- Amazon Web Services(Amazon Web Services): The entity that sells Amazon Web Servicesservices.
+ ///- AISPL (Amazon Internet Services Pvt. Ltd.): The local Indian entity that's an acting reseller for Amazon Web Servicesservices in India.
///- Amazon Web Services Marketplace: The entity that supports the sale of solutions that are built on Amazon Web Services by third-party software providers.
CACHE_ENGINE - The Amazon ElastiCache operating system. Examples are Windows or Linux.
AZ - The Availability Zone. An example is us-east-1a
.
BILLING_ENTITY - The Amazon Web Services seller that your account is with. Possible values are the following:
- ///- Amazon Web Services(Amazon Web Services): The entity that sells Amazon Web Services.
- ///- AISPL (Amazon Internet Services Pvt. Ltd.): The local Indian entity that's an acting reseller for Amazon Web Services in India.
+ ///- Amazon Web Services(Amazon Web Services): The entity that sells Amazon Web Servicesservices.
+ ///- AISPL (Amazon Internet Services Pvt. Ltd.): The local Indian entity that's an acting reseller for Amazon Web Servicesservices in India.
///- Amazon Web Services Marketplace: The entity that supports the sale of solutions that are built on Amazon Web Services by third-party software providers.
CACHE_ENGINE - The Amazon ElastiCache operating system. Examples are Windows or Linux.
AZ - The Availability Zone. An example is us-east-1a
.
BILLING_ENTITY - The Amazon Web Services seller that your account is with. Possible values are the following:
- ///- Amazon Web Services(Amazon Web Services): The entity that sells Amazon Web Services.
- ///- AISPL (Amazon Internet Services Pvt. Ltd.): The local Indian entity that's an acting reseller for Amazon Web Services in India.
+ ///- Amazon Web Services(Amazon Web Services): The entity that sells Amazon Web Servicesservices.
+ ///- AISPL (Amazon Internet Services Pvt. Ltd.): The local Indian entity that's an acting reseller for Amazon Web Servicesservices in India.
///- Amazon Web Services Marketplace: The entity that supports the sale of solutions that are built on Amazon Web Services by third-party software providers.
CACHE_ENGINE - The Amazon ElastiCache operating system. Examples are Windows or Linux.
AZ - The Availability Zone. An example is us-east-1a
.
BILLING_ENTITY - The Amazon Web Services seller that your account is with. Possible values are the following:
- ///- Amazon Web Services(Amazon Web Services): The entity that sells Amazon Web Services.
- ///- AISPL (Amazon Internet Services Pvt. Ltd.): The local Indian entity that's an acting reseller for Amazon Web Services in India.
+ ///- Amazon Web Services(Amazon Web Services): The entity that sells Amazon Web Servicesservices.
+ ///- AISPL (Amazon Internet Services Pvt. Ltd.): The local Indian entity that's an acting reseller for Amazon Web Servicesservices in India.
///- Amazon Web Services Marketplace: The entity that supports the sale of solutions that are built on Amazon Web Services by third-party software providers.
CACHE_ENGINE - The Amazon ElastiCache operating system. Examples are Windows or Linux.
AZ - The Availability Zone. An example is us-east-1a
.
BILLING_ENTITY - The Amazon Web Services seller that your account is with. Possible values are the following:
- ///- Amazon Web Services(Amazon Web Services): The entity that sells Amazon Web Services.
- ///- AISPL (Amazon Internet Services Pvt. Ltd.): The local Indian entity that's an acting reseller for Amazon Web Services in India.
+ ///- Amazon Web Services(Amazon Web Services): The entity that sells Amazon Web Servicesservices.
+ ///- AISPL (Amazon Internet Services Pvt. Ltd.): The local Indian entity that's an acting reseller for Amazon Web Servicesservices in India.
///- Amazon Web Services Marketplace: The entity that supports the sale of solutions that are built on Amazon Web Services by third-party software providers.
CACHE_ENGINE - The Amazon ElastiCache operating system. Examples are Windows or Linux.
AZ - The Availability Zone. An example is us-east-1a
.
BILLING_ENTITY - The Amazon Web Services seller that your account is with. Possible values are the following:
- ///- Amazon Web Services(Amazon Web Services): The entity that sells Amazon Web Services.
- ///- AISPL (Amazon Internet Services Pvt. Ltd.): The local Indian entity that's an acting reseller for Amazon Web Services in India.
+ ///- Amazon Web Services(Amazon Web Services): The entity that sells Amazon Web Servicesservices.
+ ///- AISPL (Amazon Internet Services Pvt. Ltd.): The local Indian entity that's an acting reseller for Amazon Web Servicesservices in India.
///- Amazon Web Services Marketplace: The entity that supports the sale of solutions that are built on Amazon Web Services by third-party software providers.
CACHE_ENGINE - The Amazon ElastiCache operating system. Examples are Windows or Linux.
AZ - The Availability Zone. An example is us-east-1a
.
BILLING_ENTITY - The Amazon Web Services seller that your account is with. Possible values are the following:
- ///- Amazon Web Services(Amazon Web Services): The entity that sells Amazon Web Services.
- ///- AISPL (Amazon Internet Services Pvt. Ltd.): The local Indian entity that's an acting reseller for Amazon Web Services in India.
+ ///- Amazon Web Services(Amazon Web Services): The entity that sells Amazon Web Servicesservices.
+ ///- AISPL (Amazon Internet Services Pvt. Ltd.): The local Indian entity that's an acting reseller for Amazon Web Servicesservices in India.
///- Amazon Web Services Marketplace: The entity that supports the sale of solutions that are built on Amazon Web Services by third-party software providers.
CACHE_ENGINE - The Amazon ElastiCache operating system. Examples are Windows or Linux.
AZ - The Availability Zone. An example is us-east-1a
.
BILLING_ENTITY - The Amazon Web Services seller that your account is with. Possible values are the following:
- ///- Amazon Web Services(Amazon Web Services): The entity that sells Amazon Web Services.
- ///- AISPL (Amazon Internet Services Pvt. Ltd.): The local Indian entity that's an acting reseller for Amazon Web Services in India.
+ ///- Amazon Web Services(Amazon Web Services): The entity that sells Amazon Web Servicesservices.
+ ///- AISPL (Amazon Internet Services Pvt. Ltd.): The local Indian entity that's an acting reseller for Amazon Web Servicesservices in India.
///- Amazon Web Services Marketplace: The entity that supports the sale of solutions that are built on Amazon Web Services by third-party software providers.
CACHE_ENGINE - The Amazon ElastiCache operating system. Examples are Windows or Linux.