diff --git a/CHANGELOG.md b/CHANGELOG.md index 4a12ca738f0..7e7c484210a 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -1,3 +1,18 @@ +Release v1.44.214 (2023-03-03) +=== + +### Service Client Updates +* `service/dynamodb`: Updates service API, documentation, waiters, paginators, and examples + * Documentation updates for DynamoDB. +* `service/ec2`: Updates service API and documentation + * This release adds support for a new boot mode for EC2 instances called 'UEFI Preferred'. +* `service/macie2`: Updates service documentation +* `service/mediaconvert`: Updates service API and documentation + * The AWS Elemental MediaConvert SDK has improved handling for different input and output color space combinations. +* `service/medialive`: Updates service API and documentation + * AWS Elemental MediaLive adds support for Nielsen watermark timezones. +* `service/transcribe`: Updates service API and documentation + Release v1.44.213 (2023-03-02) === diff --git a/aws/endpoints/defaults.go b/aws/endpoints/defaults.go index 357e30733e3..ad6cf3696df 100644 --- a/aws/endpoints/defaults.go +++ b/aws/endpoints/defaults.go @@ -18309,6 +18309,9 @@ var awsPartition = partition{ }, Deprecated: boxedTrue, }, + endpointKey{ + Region: "me-central-1", + }: endpoint{}, endpointKey{ Region: "me-south-1", }: endpoint{}, @@ -19820,9 +19823,6 @@ var awsPartition = partition{ endpointKey{ Region: "ap-southeast-2", }: endpoint{}, - endpointKey{ - Region: "api", - }: endpoint{}, endpointKey{ Region: "ca-central-1", }: endpoint{}, @@ -34104,12 +34104,22 @@ var awsusgovPartition = partition{ "mediaconvert": service{ Endpoints: serviceEndpoints{ endpointKey{ - Region: "us-gov-west-1", + Region: "fips-us-gov-west-1", }: endpoint{ Hostname: "mediaconvert.us-gov-west-1.amazonaws.com", CredentialScope: credentialScope{ Region: "us-gov-west-1", }, + Deprecated: boxedTrue, + }, + endpointKey{ + Region: "us-gov-west-1", + }: endpoint{}, + endpointKey{ + Region: "us-gov-west-1", + Variant: fipsVariant, + }: endpoint{ + Hostname: "mediaconvert.us-gov-west-1.amazonaws.com", }, }, }, diff --git a/aws/version.go b/aws/version.go index d1cd0acfa83..b716e0a2467 100644 --- a/aws/version.go +++ b/aws/version.go @@ -5,4 +5,4 @@ package aws const SDKName = "aws-sdk-go" // SDKVersion is the version of this SDK -const SDKVersion = "1.44.213" +const SDKVersion = "1.44.214" diff --git a/models/apis/dynamodb/2012-08-10/docs-2.json b/models/apis/dynamodb/2012-08-10/docs-2.json index 7d1c103fd31..0f00e9f8c14 100644 --- a/models/apis/dynamodb/2012-08-10/docs-2.json +++ b/models/apis/dynamodb/2012-08-10/docs-2.json @@ -6,27 +6,27 @@ "BatchGetItem": "
The BatchGetItem
operation returns the attributes of one or more items from one or more tables. You identify requested items by primary key.
A single operation can retrieve up to 16 MB of data, which can contain as many as 100 items. BatchGetItem
returns a partial result if the response size limit is exceeded, the table's provisioned throughput is exceeded, or an internal processing failure occurs. If a partial result is returned, the operation returns a value for UnprocessedKeys
. You can use this value to retry the operation starting with the next item to get.
If you request more than 100 items, BatchGetItem
returns a ValidationException
with the message \"Too many items requested for the BatchGetItem call.\"
For example, if you ask to retrieve 100 items, but each individual item is 300 KB in size, the system returns 52 items (so as not to exceed the 16 MB limit). It also returns an appropriate UnprocessedKeys
value so you can get the next page of results. If desired, your application can include its own logic to assemble the pages of results into one dataset.
If none of the items can be processed due to insufficient provisioned throughput on all of the tables in the request, then BatchGetItem
returns a ProvisionedThroughputExceededException
. If at least one of the items is successfully processed, then BatchGetItem
completes successfully, while returning the keys of the unread items in UnprocessedKeys
.
If DynamoDB returns any unprocessed items, you should retry the batch operation on those items. However, we strongly recommend that you use an exponential backoff algorithm. If you retry the batch operation immediately, the underlying read or write requests can still fail due to throttling on the individual tables. If you delay the batch operation using exponential backoff, the individual requests in the batch are much more likely to succeed.
For more information, see Batch Operations and Error Handling in the Amazon DynamoDB Developer Guide.
By default, BatchGetItem
performs eventually consistent reads on every table in the request. If you want strongly consistent reads instead, you can set ConsistentRead
to true
for any or all tables.
In order to minimize response latency, BatchGetItem
retrieves items in parallel.
When designing your application, keep in mind that DynamoDB does not return items in any particular order. To help parse the response by item, include the primary key values for the items in your request in the ProjectionExpression
parameter.
If a requested item does not exist, it is not returned in the result. Requests for nonexistent items consume the minimum read capacity units according to the type of read. For more information, see Working with Tables in the Amazon DynamoDB Developer Guide.
", "BatchWriteItem": "The BatchWriteItem
operation puts or deletes multiple items in one or more tables. A single call to BatchWriteItem
can transmit up to 16MB of data over the network, consisting of up to 25 item put or delete operations. While individual items can be up to 400 KB once stored, it's important to note that an item's representation might be greater than 400KB while being sent in DynamoDB's JSON format for the API call. For more details on this distinction, see Naming Rules and Data Types.
BatchWriteItem
cannot update items. If you perform a BatchWriteItem
operation on an existing item, that item's values will be overwritten by the operation and it will appear like it was updated. To update items, we recommend you use the UpdateItem
action.
The individual PutItem
and DeleteItem
operations specified in BatchWriteItem
are atomic; however BatchWriteItem
as a whole is not. If any requested operations fail because the table's provisioned throughput is exceeded or an internal processing failure occurs, the failed operations are returned in the UnprocessedItems
response parameter. You can investigate and optionally resend the requests. Typically, you would call BatchWriteItem
in a loop. Each iteration would check for unprocessed items and submit a new BatchWriteItem
request with those unprocessed items until all items have been processed.
If none of the items can be processed due to insufficient provisioned throughput on all of the tables in the request, then BatchWriteItem
returns a ProvisionedThroughputExceededException
.
If DynamoDB returns any unprocessed items, you should retry the batch operation on those items. However, we strongly recommend that you use an exponential backoff algorithm. If you retry the batch operation immediately, the underlying read or write requests can still fail due to throttling on the individual tables. If you delay the batch operation using exponential backoff, the individual requests in the batch are much more likely to succeed.
For more information, see Batch Operations and Error Handling in the Amazon DynamoDB Developer Guide.
With BatchWriteItem
, you can efficiently write or delete large amounts of data, such as from Amazon EMR, or copy data from another database into DynamoDB. In order to improve performance with these large-scale operations, BatchWriteItem
does not behave in the same way as individual PutItem
and DeleteItem
calls would. For example, you cannot specify conditions on individual put and delete requests, and BatchWriteItem
does not return deleted items in the response.
If you use a programming language that supports concurrency, you can use threads to write items in parallel. Your application must include the necessary logic to manage the threads. With languages that don't support threading, you must update or delete the specified items one at a time. In both situations, BatchWriteItem
performs the specified put and delete operations in parallel, giving you the power of the thread pool approach without having to introduce complexity into your application.
Parallel processing reduces latency, but each specified put and delete request consumes the same number of write capacity units whether it is processed in parallel or not. Delete operations on nonexistent items consume one write capacity unit.
If one or more of the following is true, DynamoDB rejects the entire batch write operation:
One or more tables specified in the BatchWriteItem
request does not exist.
Primary key attributes specified on an item in the request do not match those in the corresponding table's primary key schema.
You try to perform multiple operations on the same item in the same BatchWriteItem
request. For example, you cannot put and delete the same item in the same BatchWriteItem
request.
Your request contains at least two items with identical hash and range keys (which essentially is two put operations).
There are more than 25 requests in the batch.
Any individual item in a batch exceeds 400 KB.
The total request size exceeds 16 MB.
Creates a backup for an existing table.
Each time you create an on-demand backup, the entire table data is backed up. There is no limit to the number of on-demand backups that can be taken.
When you create an on-demand backup, a time marker of the request is cataloged, and the backup is created asynchronously, by applying all changes until the time of the request to the last full table snapshot. Backup requests are processed instantaneously and become available for restore within minutes.
You can call CreateBackup
at a maximum rate of 50 times per second.
All backups in DynamoDB work without consuming any provisioned throughput on the table.
If you submit a backup request on 2018-12-14 at 14:25:00, the backup is guaranteed to contain all data committed to the table up to 14:24:00, and data committed after 14:26:00 will not be. The backup might contain data modifications made between 14:24:00 and 14:26:00. On-demand backup does not support causal consistency.
Along with data, the following are also included on the backups:
Global secondary indexes (GSIs)
Local secondary indexes (LSIs)
Streams
Provisioned read and write capacity
Creates a global table from an existing table. A global table creates a replication relationship between two or more DynamoDB tables with the same table name in the provided Regions.
This operation only applies to Version 2017.11.29 of global tables.
If you want to add a new replica table to a global table, each of the following conditions must be true:
The table must have the same primary key as all of the other replicas.
The table must have the same name as all of the other replicas.
The table must have DynamoDB Streams enabled, with the stream containing both the new and the old images of the item.
None of the replica tables in the global table can contain any data.
If global secondary indexes are specified, then the following conditions must also be met:
The global secondary indexes must have the same name.
The global secondary indexes must have the same hash key and sort key (if present).
If local secondary indexes are specified, then the following conditions must also be met:
The local secondary indexes must have the same name.
The local secondary indexes must have the same hash key and sort key (if present).
Write capacity settings should be set consistently across your replica tables and secondary indexes. DynamoDB strongly recommends enabling auto scaling to manage the write capacity settings for all of your global tables replicas and indexes.
If you prefer to manage write capacity settings manually, you should provision equal replicated write capacity units to your replica tables. You should also provision equal replicated write capacity units to matching secondary indexes across your global table.
Creates a global table from an existing table. A global table creates a replication relationship between two or more DynamoDB tables with the same table name in the provided Regions.
This operation only applies to Version 2017.11.29 (Legacy) of global tables. We recommend using Version 2019.11.21 (Current) when creating new global tables, as it provides greater flexibility, higher efficiency and consumes less write capacity than 2017.11.29 (Legacy). To determine which version you are using, see Determining the version. To update existing global tables from version 2017.11.29 (Legacy) to version 2019.11.21 (Current), see Updating global tables.
If you want to add a new replica table to a global table, each of the following conditions must be true:
The table must have the same primary key as all of the other replicas.
The table must have the same name as all of the other replicas.
The table must have DynamoDB Streams enabled, with the stream containing both the new and the old images of the item.
None of the replica tables in the global table can contain any data.
If global secondary indexes are specified, then the following conditions must also be met:
The global secondary indexes must have the same name.
The global secondary indexes must have the same hash key and sort key (if present).
If local secondary indexes are specified, then the following conditions must also be met:
The local secondary indexes must have the same name.
The local secondary indexes must have the same hash key and sort key (if present).
Write capacity settings should be set consistently across your replica tables and secondary indexes. DynamoDB strongly recommends enabling auto scaling to manage the write capacity settings for all of your global tables replicas and indexes.
If you prefer to manage write capacity settings manually, you should provision equal replicated write capacity units to your replica tables. You should also provision equal replicated write capacity units to matching secondary indexes across your global table.
The CreateTable
operation adds a new table to your account. In an Amazon Web Services account, table names must be unique within each Region. That is, you can have two tables with same name if you create the tables in different Regions.
CreateTable
is an asynchronous operation. Upon receiving a CreateTable
request, DynamoDB immediately returns a response with a TableStatus
of CREATING
. After the table is created, DynamoDB sets the TableStatus
to ACTIVE
. You can perform read and write operations only on an ACTIVE
table.
You can optionally define secondary indexes on the new table, as part of the CreateTable
operation. If you want to create multiple tables with secondary indexes on them, you must create the tables sequentially. Only one table with secondary indexes can be in the CREATING
state at any given time.
You can use the DescribeTable
action to check the table status.
Deletes an existing backup of a table.
You can call DeleteBackup
at a maximum rate of 10 times per second.
Deletes a single item in a table by primary key. You can perform a conditional delete operation that deletes the item if it exists, or if it has an expected attribute value.
In addition to deleting an item, you can also return the item's attribute values in the same operation, using the ReturnValues
parameter.
Unless you specify conditions, the DeleteItem
is an idempotent operation; running it multiple times on the same item or attribute does not result in an error response.
Conditional deletes are useful for deleting items only if specific conditions are met. If those conditions are met, DynamoDB performs the delete. Otherwise, the item is not deleted.
", - "DeleteTable": "The DeleteTable
operation deletes a table and all of its items. After a DeleteTable
request, the specified table is in the DELETING
state until DynamoDB completes the deletion. If the table is in the ACTIVE
state, you can delete it. If a table is in CREATING
or UPDATING
states, then DynamoDB returns a ResourceInUseException
. If the specified table does not exist, DynamoDB returns a ResourceNotFoundException
. If table is already in the DELETING
state, no error is returned.
DynamoDB might continue to accept data read and write operations, such as GetItem
and PutItem
, on a table in the DELETING
state until the table deletion is complete.
When you delete a table, any indexes on that table are also deleted.
If you have DynamoDB Streams enabled on the table, then the corresponding stream on that table goes into the DISABLED
state, and the stream is automatically deleted after 24 hours.
Use the DescribeTable
action to check the status of the table.
The DeleteTable
operation deletes a table and all of its items. After a DeleteTable
request, the specified table is in the DELETING
state until DynamoDB completes the deletion. If the table is in the ACTIVE
state, you can delete it. If a table is in CREATING
or UPDATING
states, then DynamoDB returns a ResourceInUseException
. If the specified table does not exist, DynamoDB returns a ResourceNotFoundException
. If table is already in the DELETING
state, no error is returned.
This operation only applies to Version 2019.11.21 (Current) of global tables.
DynamoDB might continue to accept data read and write operations, such as GetItem
and PutItem
, on a table in the DELETING
state until the table deletion is complete.
When you delete a table, any indexes on that table are also deleted.
If you have DynamoDB Streams enabled on the table, then the corresponding stream on that table goes into the DISABLED
state, and the stream is automatically deleted after 24 hours.
Use the DescribeTable
action to check the status of the table.
Describes an existing backup of a table.
You can call DescribeBackup
at a maximum rate of 10 times per second.
Checks the status of continuous backups and point in time recovery on the specified table. Continuous backups are ENABLED
on all tables at table creation. If point in time recovery is enabled, PointInTimeRecoveryStatus
will be set to ENABLED.
After continuous backups and point in time recovery are enabled, you can restore to any point in time within EarliestRestorableDateTime
and LatestRestorableDateTime
.
LatestRestorableDateTime
is typically 5 minutes before the current time. You can restore your table to any point in time during the last 35 days.
You can call DescribeContinuousBackups
at a maximum rate of 10 times per second.
Returns information about contributor insights, for a given table or global secondary index.
", - "DescribeEndpoints": "Returns the regional endpoint information.
", + "DescribeContributorInsights": "Returns information about contributor insights for a given table or global secondary index.
", + "DescribeEndpoints": "Returns the regional endpoint information. This action must be included in your VPC endpoint policies, or access to the DescribeEndpoints API will be denied. For more information on policy permissions, please see Internetwork traffic privacy.
", "DescribeExport": "Describes an existing table export.
", - "DescribeGlobalTable": "Returns information about the specified global table.
This operation only applies to Version 2017.11.29 of global tables. If you are using global tables Version 2019.11.21 you can use DescribeTable instead.
Describes Region-specific settings for a global table.
This operation only applies to Version 2017.11.29 of global tables.
Returns information about the specified global table.
This operation only applies to Version 2017.11.29 (Legacy) of global tables. We recommend using Version 2019.11.21 (Current) when creating new global tables, as it provides greater flexibility, higher efficiency and consumes less write capacity than 2017.11.29 (Legacy). To determine which version you are using, see Determining the version. To update existing global tables from version 2017.11.29 (Legacy) to version 2019.11.21 (Current), see Updating global tables.
Describes Region-specific settings for a global table.
This operation only applies to Version 2017.11.29 (Legacy) of global tables. We recommend using Version 2019.11.21 (Current) when creating new global tables, as it provides greater flexibility, higher efficiency and consumes less write capacity than 2017.11.29 (Legacy). To determine which version you are using, see Determining the version. To update existing global tables from version 2017.11.29 (Legacy) to version 2019.11.21 (Current), see Updating global tables.
Represents the properties of the import.
", "DescribeKinesisStreamingDestination": "Returns information about the status of Kinesis streaming.
", "DescribeLimits": "Returns the current provisioned-capacity quotas for your Amazon Web Services account in a Region, both for the Region as a whole and for any one DynamoDB table that you create there.
When you establish an Amazon Web Services account, the account has initial quotas on the maximum read capacity units and write capacity units that you can provision across all of your DynamoDB tables in a given Region. Also, there are per-table quotas that apply when you create a table there. For more information, see Service, Account, and Table Quotas page in the Amazon DynamoDB Developer Guide.
Although you can increase these quotas by filing a case at Amazon Web Services Support Center, obtaining the increase is not instantaneous. The DescribeLimits
action lets you write code to compare the capacity you are currently using to those quotas imposed by your account so that you have enough time to apply for an increase before you hit a quota.
For example, you could use one of the Amazon Web Services SDKs to do the following:
Call DescribeLimits
for a particular Region to obtain your current account quotas on provisioned capacity there.
Create a variable to hold the aggregate read capacity units provisioned for all your tables in that Region, and one to hold the aggregate write capacity units. Zero them both.
Call ListTables
to obtain a list of all your DynamoDB tables.
For each table name listed by ListTables
, do the following:
Call DescribeTable
with the table name.
Use the data returned by DescribeTable
to add the read capacity units and write capacity units provisioned for the table itself to your variables.
If the table has one or more global secondary indexes (GSIs), loop over these GSIs and add their provisioned capacity values to your variables as well.
Report the account quotas for that Region returned by DescribeLimits
, along with the total current provisioned capacity levels you have calculated.
This will let you see whether you are getting close to your account-level quotas.
The per-table quotas apply only when you are creating a new table. They restrict the sum of the provisioned capacity of the new table itself and all its global secondary indexes.
For existing tables and their GSIs, DynamoDB doesn't let you increase provisioned capacity extremely rapidly, but the only quota that applies is that the aggregate provisioned capacity over all your tables and GSIs cannot exceed either of the per-account quotas.
DescribeLimits
should only be called periodically. You can expect throttling errors if you call it more than once in a minute.
The DescribeLimits
Request element has no content.
Returns information about the table, including the current status of the table, when it was created, the primary key schema, and any indexes on the table.
If you issue a DescribeTable
request immediately after a CreateTable
request, DynamoDB might return a ResourceNotFoundException
. This is because DescribeTable
uses an eventually consistent query, and the metadata for your table might not be available at that moment. Wait for a few seconds, and then try the DescribeTable
request again.
Describes auto scaling settings across replicas of the global table at once.
This operation only applies to Version 2019.11.21 of global tables.
Returns information about the table, including the current status of the table, when it was created, the primary key schema, and any indexes on the table.
This operation only applies to Version 2019.11.21 (Current) of global tables.
If you issue a DescribeTable
request immediately after a CreateTable
request, DynamoDB might return a ResourceNotFoundException
. This is because DescribeTable
uses an eventually consistent query, and the metadata for your table might not be available at that moment. Wait for a few seconds, and then try the DescribeTable
request again.
Describes auto scaling settings across replicas of the global table at once.
This operation only applies to Version 2019.11.21 (Current) of global tables.
Gives a description of the Time to Live (TTL) status on the specified table.
", "DisableKinesisStreamingDestination": "Stops replication from the DynamoDB table to the Kinesis data stream. This is done without deleting either of the resources.
", "EnableKinesisStreamingDestination": "Starts table data replication to the specified Kinesis data stream at a timestamp chosen during the enable workflow. If this operation doesn't return results immediately, use DescribeKinesisStreamingDestination to check if streaming to the Kinesis data stream is ACTIVE.
", - "ExecuteStatement": "This operation allows you to perform reads and singleton writes on data stored in DynamoDB, using PartiQL.
For PartiQL reads (SELECT
statement), if the total number of processed items exceeds the maximum dataset size limit of 1 MB, the read stops and results are returned to the user as a LastEvaluatedKey
value to continue the read in a subsequent operation. If the filter criteria in WHERE
clause does not match any data, the read will return an empty result set.
A single SELECT
statement response can return up to the maximum number of items (if using the Limit parameter) or a maximum of 1 MB of data (and then apply any filtering to the results using WHERE
clause). If LastEvaluatedKey
is present in the response, you need to paginate the result set.
This operation allows you to perform reads and singleton writes on data stored in DynamoDB, using PartiQL.
For PartiQL reads (SELECT
statement), if the total number of processed items exceeds the maximum dataset size limit of 1 MB, the read stops and results are returned to the user as a LastEvaluatedKey
value to continue the read in a subsequent operation. If the filter criteria in WHERE
clause does not match any data, the read will return an empty result set.
A single SELECT
statement response can return up to the maximum number of items (if using the Limit parameter) or a maximum of 1 MB of data (and then apply any filtering to the results using WHERE
clause). If LastEvaluatedKey
is present in the response, you need to paginate the result set. If NextToken
is present, you need to paginate the result set and include NextToken
.
This operation allows you to perform transactional reads or writes on data stored in DynamoDB, using PartiQL.
The entire transaction must consist of either read statements or write statements, you cannot mix both in one transaction. The EXISTS function is an exception and can be used to check the condition of specific attributes of the item in a similar manner to ConditionCheck
in the TransactWriteItems API.
Exports table data to an S3 bucket. The table must have point in time recovery enabled, and you can export data from any time within the point in time recovery window.
", "GetItem": "The GetItem
operation returns a set of attributes for the item with the given primary key. If there is no matching item, GetItem
does not return any data and there will be no Item
element in the response.
GetItem
provides an eventually consistent read by default. If your application requires a strongly consistent read, set ConsistentRead
to true
. Although a strongly consistent read might take more time than an eventually consistent read, it always returns the last updated value.
List backups associated with an Amazon Web Services account. To list backups for a given table, specify TableName
. ListBackups
returns a paginated list of results with at most 1 MB worth of items in a page. You can also specify a maximum number of entries to be returned in a page.
In the request, start time is inclusive, but end time is exclusive. Note that these boundaries are for the time at which the original backup was requested.
You can call ListBackups
a maximum of five times per second.
Returns a list of ContributorInsightsSummary for a table and all its global secondary indexes.
", "ListExports": "Lists completed exports within the past 90 days.
", - "ListGlobalTables": "Lists all global tables that have a replica in the specified Region.
This operation only applies to Version 2017.11.29 of global tables.
Lists all global tables that have a replica in the specified Region.
This operation only applies to Version 2017.11.29 (Legacy) of global tables. We recommend using Version 2019.11.21 (Current) when creating new global tables, as it provides greater flexibility, higher efficiency and consumes less write capacity than 2017.11.29 (Legacy). To determine which version you are using, see Determining the version. To update existing global tables from version 2017.11.29 (Legacy) to version 2019.11.21 (Current), see Updating global tables.
Lists completed imports within the past 90 days.
", "ListTables": "Returns an array of table names associated with the current account and endpoint. The output from ListTables
is paginated, with each page returning a maximum of 100 table names.
List all tags on an Amazon DynamoDB resource. You can call ListTagsOfResource up to 10 times per second, per account.
For an overview on tagging DynamoDB resources, see Tagging for DynamoDB in the Amazon DynamoDB Developer Guide.
", @@ -44,16 +44,16 @@ "RestoreTableToPointInTime": "Restores the specified table to the specified point in time within EarliestRestorableDateTime
and LatestRestorableDateTime
. You can restore your table to any point in time during the last 35 days. Any number of users can execute up to 4 concurrent restores (any type of restore) in a given account.
When you restore using point in time recovery, DynamoDB restores your table data to the state based on the selected date and time (day:hour:minute:second) to a new table.
Along with data, the following are also included on the new restored table using point in time recovery:
Global secondary indexes (GSIs)
Local secondary indexes (LSIs)
Provisioned read and write capacity
Encryption settings
All these settings come from the current settings of the source table at the time of restore.
You must manually set up the following on the restored table:
Auto scaling policies
IAM policies
Amazon CloudWatch metrics and alarms
Tags
Stream settings
Time to Live (TTL) settings
Point in time recovery settings
The Scan
operation returns one or more items and item attributes by accessing every item in a table or a secondary index. To have DynamoDB return fewer items, you can provide a FilterExpression
operation.
If the total number of scanned items exceeds the maximum dataset size limit of 1 MB, the scan stops and results are returned to the user as a LastEvaluatedKey
value to continue the scan in a subsequent operation. The results also include the number of items exceeding the limit. A scan can result in no table data meeting the filter criteria.
A single Scan
operation reads up to the maximum number of items set (if using the Limit
parameter) or a maximum of 1 MB of data and then apply any filtering to the results using FilterExpression
. If LastEvaluatedKey
is present in the response, you need to paginate the result set. For more information, see Paginating the Results in the Amazon DynamoDB Developer Guide.
Scan
operations proceed sequentially; however, for faster performance on a large table or secondary index, applications can request a parallel Scan
operation by providing the Segment
and TotalSegments
parameters. For more information, see Parallel Scan in the Amazon DynamoDB Developer Guide.
Scan
uses eventually consistent reads when accessing the data in a table; therefore, the result set might not include the changes to data in the table immediately before the operation began. If you need a consistent copy of the data, as of the time that the Scan
begins, you can set the ConsistentRead
parameter to true
.
Associate a set of tags with an Amazon DynamoDB resource. You can then activate these user-defined tags so that they appear on the Billing and Cost Management console for cost allocation tracking. You can call TagResource up to five times per second, per account.
For an overview on tagging DynamoDB resources, see Tagging for DynamoDB in the Amazon DynamoDB Developer Guide.
", - "TransactGetItems": " TransactGetItems
is a synchronous operation that atomically retrieves multiple items from one or more tables (but not from indexes) in a single account and Region. A TransactGetItems
call can contain up to 100 TransactGetItem
objects, each of which contains a Get
structure that specifies an item to retrieve from a table in the account and Region. A call to TransactGetItems
cannot retrieve items from tables in more than one Amazon Web Services account or Region. The aggregate size of the items in the transaction cannot exceed 4 MB.
DynamoDB rejects the entire TransactGetItems
request if any of the following is true:
A conflicting operation is in the process of updating an item to be read.
There is insufficient provisioned capacity for the transaction to be completed.
There is a user error, such as an invalid data format.
The aggregate size of the items in the transaction cannot exceed 4 MB.
TransactGetItems
is a synchronous operation that atomically retrieves multiple items from one or more tables (but not from indexes) in a single account and Region. A TransactGetItems
call can contain up to 100 TransactGetItem
objects, each of which contains a Get
structure that specifies an item to retrieve from a table in the account and Region. A call to TransactGetItems
cannot retrieve items from tables in more than one Amazon Web Services account or Region. The aggregate size of the items in the transaction cannot exceed 4 MB.
DynamoDB rejects the entire TransactGetItems
request if any of the following is true:
A conflicting operation is in the process of updating an item to be read.
There is insufficient provisioned capacity for the transaction to be completed.
There is a user error, such as an invalid data format.
The aggregate size of the items in the transaction exceeded 4 MB.
TransactWriteItems
is a synchronous write operation that groups up to 100 action requests. These actions can target items in different tables, but not in different Amazon Web Services accounts or Regions, and no two actions can target the same item. For example, you cannot both ConditionCheck
and Update
the same item. The aggregate size of the items in the transaction cannot exceed 4 MB.
The actions are completed atomically so that either all of them succeed, or all of them fail. They are defined by the following objects:
Put
— Initiates a PutItem
operation to write a new item. This structure specifies the primary key of the item to be written, the name of the table to write it in, an optional condition expression that must be satisfied for the write to succeed, a list of the item's attributes, and a field indicating whether to retrieve the item's attributes if the condition is not met.
Update
— Initiates an UpdateItem
operation to update an existing item. This structure specifies the primary key of the item to be updated, the name of the table where it resides, an optional condition expression that must be satisfied for the update to succeed, an expression that defines one or more attributes to be updated, and a field indicating whether to retrieve the item's attributes if the condition is not met.
Delete
— Initiates a DeleteItem
operation to delete an existing item. This structure specifies the primary key of the item to be deleted, the name of the table where it resides, an optional condition expression that must be satisfied for the deletion to succeed, and a field indicating whether to retrieve the item's attributes if the condition is not met.
ConditionCheck
— Applies a condition to an item that is not being modified by the transaction. This structure specifies the primary key of the item to be checked, the name of the table where it resides, a condition expression that must be satisfied for the transaction to succeed, and a field indicating whether to retrieve the item's attributes if the condition is not met.
DynamoDB rejects the entire TransactWriteItems
request if any of the following is true:
A condition in one of the condition expressions is not met.
An ongoing operation is in the process of updating the same item.
There is insufficient provisioned capacity for the transaction to be completed.
An item size becomes too large (bigger than 400 KB), a local secondary index (LSI) becomes too large, or a similar validation error occurs because of changes made by the transaction.
The aggregate size of the items in the transaction exceeds 4 MB.
There is a user error, such as an invalid data format.
Removes the association of tags from an Amazon DynamoDB resource. You can call UntagResource
up to five times per second, per account.
For an overview on tagging DynamoDB resources, see Tagging for DynamoDB in the Amazon DynamoDB Developer Guide.
", "UpdateContinuousBackups": " UpdateContinuousBackups
enables or disables point in time recovery for the specified table. A successful UpdateContinuousBackups
call returns the current ContinuousBackupsDescription
. Continuous backups are ENABLED
on all tables at table creation. If point in time recovery is enabled, PointInTimeRecoveryStatus
will be set to ENABLED.
Once continuous backups and point in time recovery are enabled, you can restore to any point in time within EarliestRestorableDateTime
and LatestRestorableDateTime
.
LatestRestorableDateTime
is typically 5 minutes before the current time. You can restore your table to any point in time during the last 35 days.
Updates the status for contributor insights for a specific table or index. CloudWatch Contributor Insights for DynamoDB graphs display the partition key and (if applicable) sort key of frequently accessed items and frequently throttled items in plaintext. If you require the use of Amazon Web Services Key Management Service (KMS) to encrypt this table’s partition key and sort key data with an Amazon Web Services managed key or customer managed key, you should not enable CloudWatch Contributor Insights for DynamoDB for this table.
", - "UpdateGlobalTable": "Adds or removes replicas in the specified global table. The global table must already exist to be able to use this operation. Any replica to be added must be empty, have the same name as the global table, have the same key schema, have DynamoDB Streams enabled, and have the same provisioned and maximum write capacity units.
Although you can use UpdateGlobalTable
to add replicas and remove replicas in a single request, for simplicity we recommend that you issue separate requests for adding or removing replicas.
If global secondary indexes are specified, then the following conditions must also be met:
The global secondary indexes must have the same name.
The global secondary indexes must have the same hash key and sort key (if present).
The global secondary indexes must have the same provisioned and maximum write capacity units.
Updates settings for a global table.
", + "UpdateGlobalTable": "Adds or removes replicas in the specified global table. The global table must already exist to be able to use this operation. Any replica to be added must be empty, have the same name as the global table, have the same key schema, have DynamoDB Streams enabled, and have the same provisioned and maximum write capacity units.
This operation only applies to Version 2017.11.29 (Legacy) of global tables. We recommend using Version 2019.11.21 (Current) when creating new global tables, as it provides greater flexibility, higher efficiency and consumes less write capacity than 2017.11.29 (Legacy). To determine which version you are using, see Determining the version. To update existing global tables from version 2017.11.29 (Legacy) to version 2019.11.21 (Current), see Updating global tables.
This operation only applies to Version 2017.11.29 of global tables. If you are using global tables Version 2019.11.21 you can use DescribeTable instead.
Although you can use UpdateGlobalTable
to add replicas and remove replicas in a single request, for simplicity we recommend that you issue separate requests for adding or removing replicas.
If global secondary indexes are specified, then the following conditions must also be met:
The global secondary indexes must have the same name.
The global secondary indexes must have the same hash key and sort key (if present).
The global secondary indexes must have the same provisioned and maximum write capacity units.
Updates settings for a global table.
This operation only applies to Version 2017.11.29 (Legacy) of global tables. We recommend using Version 2019.11.21 (Current) when creating new global tables, as it provides greater flexibility, higher efficiency and consumes less write capacity than 2017.11.29 (Legacy). To determine which version you are using, see Determining the version. To update existing global tables from version 2017.11.29 (Legacy) to version 2019.11.21 (Current), see Updating global tables.
Edits an existing item's attributes, or adds a new item to the table if it does not already exist. You can put, delete, or add attribute values. You can also perform a conditional update on an existing item (insert a new attribute name-value pair if it doesn't exist, or replace an existing name-value pair if it has certain expected attribute values).
You can also return the item's attribute values in the same UpdateItem
operation using the ReturnValues
parameter.
Modifies the provisioned throughput settings, global secondary indexes, or DynamoDB Streams settings for a given table.
You can only perform one of the following operations at once:
Modify the provisioned throughput settings of the table.
Remove a global secondary index from the table.
Create a new global secondary index on the table. After the index begins backfilling, you can use UpdateTable
to perform other operations.
UpdateTable
is an asynchronous operation; while it is executing, the table status changes from ACTIVE
to UPDATING
. While it is UPDATING
, you cannot issue another UpdateTable
request. When the table returns to the ACTIVE
state, the UpdateTable
operation is complete.
Updates auto scaling settings on your global tables at once.
This operation only applies to Version 2019.11.21 of global tables.
Modifies the provisioned throughput settings, global secondary indexes, or DynamoDB Streams settings for a given table.
This operation only applies to Version 2019.11.21 (Current) of global tables.
You can only perform one of the following operations at once:
Modify the provisioned throughput settings of the table.
Remove a global secondary index from the table.
Create a new global secondary index on the table. After the index begins backfilling, you can use UpdateTable
to perform other operations.
UpdateTable
is an asynchronous operation; while it is executing, the table status changes from ACTIVE
to UPDATING
. While it is UPDATING
, you cannot issue another UpdateTable
request. When the table returns to the ACTIVE
state, the UpdateTable
operation is complete.
Updates auto scaling settings on your global tables at once.
This operation only applies to Version 2019.11.21 (Current) of global tables.
The UpdateTimeToLive
method enables or disables Time to Live (TTL) for the specified table. A successful UpdateTimeToLive
call returns the current TimeToLiveSpecification
. It can take up to one hour for the change to fully process. Any additional UpdateTimeToLive
calls for the same table during this one hour duration result in a ValidationException
.
TTL compares the current time in epoch time format to the time stored in the TTL attribute of an item. If the epoch time value stored in the attribute is less than the current time, the item is marked as expired and subsequently deleted.
The epoch time format is the number of seconds elapsed since 12:00:00 AM January 1, 1970 UTC.
DynamoDB deletes expired items on a best-effort basis to ensure availability of throughput for other data operations.
DynamoDB typically deletes expired items within two days of expiration. The exact duration within which an item gets deleted after expiration is specific to the nature of the workload. Items that have expired and not been deleted will still show up in reads, queries, and scans.
As items are deleted, they are removed from any local secondary index and global secondary index immediately in the same eventually consistent way as a standard delete operation.
For more information, see Time To Live in the Amazon DynamoDB Developer Guide.
" }, "shapes": { @@ -100,7 +100,7 @@ "ItemList$member": null, "ItemResponse$Item": "Map of attribute data consisting of the data type and attribute value.
", "PutItemOutput$Attributes": "The attribute values as they appeared before the PutItem
operation, but only if ReturnValues
is specified as ALL_OLD
in the request. Each element consists of an attribute name and an attribute value.
A map of attribute values as they appear before or after the UpdateItem
operation, as determined by the ReturnValues
parameter.
The Attributes
map is only present if ReturnValues
was specified as something other than NONE
in the request. Each element represents one attribute.
A map of attribute values as they appear before or after the UpdateItem
operation, as determined by the ReturnValues
parameter.
The Attributes
map is only present if the update was successful and ReturnValues
was specified as something other than NONE
in the request. Each element represents one attribute.
Set this value to get remaining results, if NextToken
was returned in the statement response.
Providing a ClientRequestToken
makes the call to TransactWriteItems
idempotent, meaning that multiple identical calls have the same effect as one single call.
Although multiple identical calls using the same client request token produce the same result on the server (no side effects), the responses to the calls might not be the same. If the ReturnConsumedCapacity>
parameter is set, then the initial TransactWriteItems
call returns the amount of write capacity units consumed in making the changes. Subsequent TransactWriteItems
calls with the same client token return the number of read capacity units consumed in reading the item.
A client request token is valid for 10 minutes after the first request that uses it is completed. After 10 minutes, any request with the same client token is treated as a new request. Do not resubmit the same request with the same client token for more than 10 minutes, or the result might not be idempotent.
If you submit a request with the same client token but a change in other parameters within the 10-minute idempotency window, DynamoDB returns an IdempotentParameterMismatch
exception.
Providing a ClientRequestToken
makes the call to TransactWriteItems
idempotent, meaning that multiple identical calls have the same effect as one single call.
Although multiple identical calls using the same client request token produce the same result on the server (no side effects), the responses to the calls might not be the same. If the ReturnConsumedCapacity
parameter is set, then the initial TransactWriteItems
call returns the amount of write capacity units consumed in making the changes. Subsequent TransactWriteItems
calls with the same client token return the number of read capacity units consumed in reading the item.
A client request token is valid for 10 minutes after the first request that uses it is completed. After 10 minutes, any request with the same client token is treated as a new request. Do not resubmit the same request with the same client token for more than 10 minutes, or the result might not be idempotent.
If you submit a request with the same client token but a change in other parameters within the 10-minute idempotency window, DynamoDB returns an IdempotentParameterMismatch
exception.
A condition that must be satisfied in order for a conditional update to succeed.
", + "ConditionCheck$ConditionExpression": "A condition that must be satisfied in order for a conditional update to succeed. For more information, see Condition expressions in the Amazon DynamoDB Developer Guide.
", "Delete$ConditionExpression": "A condition that must be satisfied in order for a conditional delete to succeed.
", "DeleteItemInput$ConditionExpression": "A condition that must be satisfied in order for a conditional DeleteItem
to succeed.
An expression can contain any of the following:
Functions: attribute_exists | attribute_not_exists | attribute_type | contains | begins_with | size
These function names are case-sensitive.
Comparison operators: = | <> | < | > | <= | >= | BETWEEN | IN
Logical operators: AND | OR | NOT
For more information about condition expressions, see Condition Expressions in the Amazon DynamoDB Developer Guide.
", "Put$ConditionExpression": "A condition that must be satisfied in order for a conditional update to succeed.
", @@ -580,13 +580,13 @@ "base": "The capacity units consumed by an operation. The data returned includes the total provisioned throughput consumed, along with statistics for the table and any indexes involved in the operation. ConsumedCapacity
is only returned if the request asked for it. For more information, see Provisioned Throughput in the Amazon DynamoDB Developer Guide.
The capacity units consumed by the DeleteItem
operation. The data returned includes the total provisioned throughput consumed, along with statistics for the table and any indexes involved in the operation. ConsumedCapacity
is only returned if the ReturnConsumedCapacity
parameter was specified. For more information, see Provisioned Mode in the Amazon DynamoDB Developer Guide.
The capacity units consumed by the DeleteItem
operation. The data returned includes the total provisioned throughput consumed, along with statistics for the table and any indexes involved in the operation. ConsumedCapacity
is only returned if the ReturnConsumedCapacity
parameter was specified. For more information, see Provisioned Throughput in the Amazon DynamoDB Developer Guide.
The capacity units consumed by the GetItem
operation. The data returned includes the total provisioned throughput consumed, along with statistics for the table and any indexes involved in the operation. ConsumedCapacity
is only returned if the ReturnConsumedCapacity
parameter was specified. For more information, see Read/Write Capacity Mode in the Amazon DynamoDB Developer Guide.
The capacity units consumed by the PutItem
operation. The data returned includes the total provisioned throughput consumed, along with statistics for the table and any indexes involved in the operation. ConsumedCapacity
is only returned if the ReturnConsumedCapacity
parameter was specified. For more information, see Read/Write Capacity Mode in the Amazon DynamoDB Developer Guide.
The capacity units consumed by the GetItem
operation. The data returned includes the total provisioned throughput consumed, along with statistics for the table and any indexes involved in the operation. ConsumedCapacity
is only returned if the ReturnConsumedCapacity
parameter was specified. For more information, see Provisioned Throughput in the Amazon DynamoDB Developer Guide.
The capacity units consumed by the PutItem
operation. The data returned includes the total provisioned throughput consumed, along with statistics for the table and any indexes involved in the operation. ConsumedCapacity
is only returned if the ReturnConsumedCapacity
parameter was specified. For more information, see Provisioned Throughput in the Amazon DynamoDB Developer Guide.
The capacity units consumed by the Query
operation. The data returned includes the total provisioned throughput consumed, along with statistics for the table and any indexes involved in the operation. ConsumedCapacity
is only returned if the ReturnConsumedCapacity
parameter was specified. For more information, see Provisioned Throughput in the Amazon DynamoDB Developer Guide.
The capacity units consumed by the Scan
operation. The data returned includes the total provisioned throughput consumed, along with statistics for the table and any indexes involved in the operation. ConsumedCapacity
is only returned if the ReturnConsumedCapacity
parameter was specified. For more information, see Provisioned Throughput in the Amazon DynamoDB Developer Guide.
The capacity units consumed by the UpdateItem
operation. The data returned includes the total provisioned throughput consumed, along with statistics for the table and any indexes involved in the operation. ConsumedCapacity
is only returned if the ReturnConsumedCapacity
parameter was specified. For more information, see Provisioned Throughput in the Amazon DynamoDB Developer Guide.
The capacity units consumed by the Scan
operation. The data returned includes the total provisioned throughput consumed, along with statistics for the table and any indexes involved in the operation. ConsumedCapacity
is only returned if the ReturnConsumedCapacity
parameter was specified. For more information, see Provisioned Throughput in the Amazon DynamoDB Developer Guide.
The capacity units consumed by the UpdateItem
operation. The data returned includes the total provisioned throughput consumed, along with statistics for the table and any indexes involved in the operation. ConsumedCapacity
is only returned if the ReturnConsumedCapacity
parameter was specified. For more information, see Provisioned Throughput in the Amazon DynamoDB Developer Guide.
One or more substitution tokens for attribute names in an expression.
", + "ConditionCheck$ExpressionAttributeNames": "One or more substitution tokens for attribute names in an expression. For more information, see Expression attribute names in the Amazon DynamoDB Developer Guide.
", "Delete$ExpressionAttributeNames": "One or more substitution tokens for attribute names in an expression.
", "DeleteItemInput$ExpressionAttributeNames": "One or more substitution tokens for attribute names in an expression. The following are some use cases for using ExpressionAttributeNames
:
To access an attribute whose name conflicts with a DynamoDB reserved word.
To create a placeholder for repeating occurrences of an attribute name in an expression.
To prevent special characters in an attribute name from being misinterpreted in an expression.
Use the # character in an expression to dereference an attribute name. For example, consider the following attribute name:
Percentile
The name of this attribute conflicts with a reserved word, so it cannot be used directly in an expression. (For the complete list of reserved words, see Reserved Words in the Amazon DynamoDB Developer Guide). To work around this, you could specify the following for ExpressionAttributeNames
:
{\"#P\":\"Percentile\"}
You could then use this substitution in an expression, as in this example:
#P = :val
Tokens that begin with the : character are expression attribute values, which are placeholders for the actual value at runtime.
For more information on expression attribute names, see Specifying Item Attributes in the Amazon DynamoDB Developer Guide.
", "Get$ExpressionAttributeNames": "One or more substitution tokens for attribute names in the ProjectionExpression parameter.
", @@ -1188,7 +1188,7 @@ "ExpressionAttributeValueMap": { "base": null, "refs": { - "ConditionCheck$ExpressionAttributeValues": "One or more values that can be substituted in an expression.
", + "ConditionCheck$ExpressionAttributeValues": "One or more values that can be substituted in an expression. For more information, see Condition expressions in the Amazon DynamoDB Developer Guide.
", "Delete$ExpressionAttributeValues": "One or more values that can be substituted in an expression.
", "DeleteItemInput$ExpressionAttributeValues": "One or more values that can be substituted in an expression.
Use the : (colon) character in an expression to dereference an attribute value. For example, suppose that you wanted to check whether the value of the ProductStatus attribute was one of the following:
Available | Backordered | Discontinued
You would first need to specify ExpressionAttributeValues
as follows:
{ \":avail\":{\"S\":\"Available\"}, \":back\":{\"S\":\"Backordered\"}, \":disc\":{\"S\":\"Discontinued\"} }
You could then use these values in an expression, such as this:
ProductStatus IN (:avail, :back, :disc)
For more information on expression attribute values, see Condition Expressions in the Amazon DynamoDB Developer Guide.
", "Put$ExpressionAttributeValues": "One or more values that can be substituted in an expression.
", @@ -1643,7 +1643,7 @@ "refs": { "ConditionCheck$Key": "The primary key of the item to be checked. Each element consists of an attribute name and a value for that attribute.
", "Delete$Key": "The primary key of the item to be deleted. Each element consists of an attribute name and a value for that attribute.
", - "DeleteItemInput$Key": "A map of attribute names to AttributeValue
objects, representing the primary key of the item to delete.
For the primary key, you must provide all of the attributes. For example, with a simple primary key, you only need to provide a value for the partition key. For a composite primary key, you must provide values for both the partition key and the sort key.
", + "DeleteItemInput$Key": "A map of attribute names to AttributeValue
objects, representing the primary key of the item to delete.
For the primary key, you must provide all of the key attributes. For example, with a simple primary key, you only need to provide a value for the partition key. For a composite primary key, you must provide values for both the partition key and the sort key.
", "DeleteRequest$Key": "A map of attribute name to attribute values, representing the primary key of the item to delete. All of the table's primary key attributes must be specified, and their data types must match those of the table's key schema.
", "ExecuteStatementOutput$LastEvaluatedKey": "The primary key of the item where the operation stopped, inclusive of the previous result set. Use this value to start a new operation, excluding this value in the new request. If LastEvaluatedKey
is empty, then the \"last page\" of results has been processed and there is no more data to be retrieved. If LastEvaluatedKey
is not empty, it does not necessarily mean that there is more data in the result set. The only way to know when you have reached the end of the result set is when LastEvaluatedKey
is empty.
A map of attribute names to AttributeValue
objects that specifies the primary key of the item to retrieve.
A list of replica update actions (create, delete, or update) for the table.
This property only applies to Version 2019.11.21 of global tables.
A list of replica update actions (create, delete, or update) for the table.
This property only applies to Version 2019.11.21 (Current) of global tables.
Use ReturnValues
if you want to get the item attributes as they appeared before they were deleted. For DeleteItem
, the valid values are:
NONE
- If ReturnValues
is not specified, or if its value is NONE
, then nothing is returned. (This setting is the default for ReturnValues
.)
ALL_OLD
- The content of the old item is returned.
There is no additional cost associated with requesting a return value aside from the small network and processing overhead of receiving a larger response. No read capacity units are consumed.
The ReturnValues
parameter is used by several DynamoDB operations; however, DeleteItem
does not recognize any values other than NONE
or ALL_OLD
.
Use ReturnValues
if you want to get the item attributes as they appeared before they were updated with the PutItem
request. For PutItem
, the valid values are:
NONE
- If ReturnValues
is not specified, or if its value is NONE
, then nothing is returned. (This setting is the default for ReturnValues
.)
ALL_OLD
- If PutItem
overwrote an attribute name-value pair, then the content of the old item is returned.
The values returned are strongly consistent.
There is no additional cost associated with requesting a return value aside from the small network and processing overhead of receiving a larger response. No read capacity units are consumed.
The ReturnValues
parameter is used by several DynamoDB operations; however, PutItem
does not recognize any values other than NONE
or ALL_OLD
.
Use ReturnValues
if you want to get the item attributes as they appear before or after they are updated. For UpdateItem
, the valid values are:
NONE
- If ReturnValues
is not specified, or if its value is NONE
, then nothing is returned. (This setting is the default for ReturnValues
.)
ALL_OLD
- Returns all of the attributes of the item, as they appeared before the UpdateItem operation.
UPDATED_OLD
- Returns only the updated attributes, as they appeared before the UpdateItem operation.
ALL_NEW
- Returns all of the attributes of the item, as they appear after the UpdateItem operation.
UPDATED_NEW
- Returns only the updated attributes, as they appear after the UpdateItem operation.
There is no additional cost associated with requesting a return value aside from the small network and processing overhead of receiving a larger response. No read capacity units are consumed.
The values returned are strongly consistent.
" + "UpdateItemInput$ReturnValues": "Use ReturnValues
if you want to get the item attributes as they appear before or after they are successfully updated. For UpdateItem
, the valid values are:
NONE
- If ReturnValues
is not specified, or if its value is NONE
, then nothing is returned. (This setting is the default for ReturnValues
.)
ALL_OLD
- Returns all of the attributes of the item, as they appeared before the UpdateItem operation.
UPDATED_OLD
- Returns only the updated attributes, as they appeared before the UpdateItem operation.
ALL_NEW
- Returns all of the attributes of the item, as they appear after the UpdateItem operation.
UPDATED_NEW
- Returns only the updated attributes, as they appear after the UpdateItem operation.
There is no additional cost associated with requesting a return value aside from the small network and processing overhead of receiving a larger response. No read capacity units are consumed.
The values returned are strongly consistent.
" } }, "ReturnValuesOnConditionCheckFailure": { @@ -2608,8 +2608,8 @@ "Select": { "base": null, "refs": { - "QueryInput$Select": "The attributes to be returned in the result. You can retrieve all item attributes, specific item attributes, the count of matching items, or in the case of an index, some or all of the attributes projected into the index.
ALL_ATTRIBUTES
- Returns all of the item attributes from the specified table or index. If you query a local secondary index, then for each matching item in the index, DynamoDB fetches the entire item from the parent table. If the index is configured to project all item attributes, then all of the data can be obtained from the local secondary index, and no fetching is required.
ALL_PROJECTED_ATTRIBUTES
- Allowed only when querying an index. Retrieves all attributes that have been projected into the index. If the index is configured to project all attributes, this return value is equivalent to specifying ALL_ATTRIBUTES
.
COUNT
- Returns the number of matching items, rather than the matching items themselves.
SPECIFIC_ATTRIBUTES
- Returns only the attributes listed in ProjectionExpression
. This return value is equivalent to specifying ProjectionExpression
without specifying any value for Select
.
If you query or scan a local secondary index and request only attributes that are projected into that index, the operation will read only the index and not the table. If any of the requested attributes are not projected into the local secondary index, DynamoDB fetches each of these attributes from the parent table. This extra fetching incurs additional throughput cost and latency.
If you query or scan a global secondary index, you can only request attributes that are projected into the index. Global secondary index queries cannot fetch attributes from the parent table.
If neither Select
nor ProjectionExpression
are specified, DynamoDB defaults to ALL_ATTRIBUTES
when accessing a table, and ALL_PROJECTED_ATTRIBUTES
when accessing an index. You cannot use both Select
and ProjectionExpression
together in a single request, unless the value for Select
is SPECIFIC_ATTRIBUTES
. (This usage is equivalent to specifying ProjectionExpression
without any value for Select
.)
If you use the ProjectionExpression
parameter, then the value for Select
can only be SPECIFIC_ATTRIBUTES
. Any other value for Select
will return an error.
The attributes to be returned in the result. You can retrieve all item attributes, specific item attributes, the count of matching items, or in the case of an index, some or all of the attributes projected into the index.
ALL_ATTRIBUTES
- Returns all of the item attributes from the specified table or index. If you query a local secondary index, then for each matching item in the index, DynamoDB fetches the entire item from the parent table. If the index is configured to project all item attributes, then all of the data can be obtained from the local secondary index, and no fetching is required.
ALL_PROJECTED_ATTRIBUTES
- Allowed only when querying an index. Retrieves all attributes that have been projected into the index. If the index is configured to project all attributes, this return value is equivalent to specifying ALL_ATTRIBUTES
.
COUNT
- Returns the number of matching items, rather than the matching items themselves.
SPECIFIC_ATTRIBUTES
- Returns only the attributes listed in ProjectionExpression
. This return value is equivalent to specifying ProjectionExpression
without specifying any value for Select
.
If you query or scan a local secondary index and request only attributes that are projected into that index, the operation reads only the index and not the table. If any of the requested attributes are not projected into the local secondary index, DynamoDB fetches each of these attributes from the parent table. This extra fetching incurs additional throughput cost and latency.
If you query or scan a global secondary index, you can only request attributes that are projected into the index. Global secondary index queries cannot fetch attributes from the parent table.
If neither Select
nor ProjectionExpression
are specified, DynamoDB defaults to ALL_ATTRIBUTES
when accessing a table, and ALL_PROJECTED_ATTRIBUTES
when accessing an index. You cannot use both Select
and ProjectionExpression
together in a single request, unless the value for Select
is SPECIFIC_ATTRIBUTES
. (This usage is equivalent to specifying ProjectionExpression
without any value for Select
.)
If you use the ProjectionExpression
parameter, then the value for Select
can only be SPECIFIC_ATTRIBUTES
. Any other value for Select
will return an error.
The attributes to be returned in the result. You can retrieve all item attributes, specific item attributes, the count of matching items, or in the case of an index, some or all of the attributes projected into the index.
ALL_ATTRIBUTES
- Returns all of the item attributes from the specified table or index. If you query a local secondary index, then for each matching item in the index, DynamoDB fetches the entire item from the parent table. If the index is configured to project all item attributes, then all of the data can be obtained from the local secondary index, and no fetching is required.
ALL_PROJECTED_ATTRIBUTES
- Allowed only when querying an index. Retrieves all attributes that have been projected into the index. If the index is configured to project all attributes, this return value is equivalent to specifying ALL_ATTRIBUTES
.
COUNT
- Returns the number of matching items, rather than the matching items themselves. Note that this uses the same quantity of read capacity units as getting the items, and is subject to the same item size calculations.
SPECIFIC_ATTRIBUTES
- Returns only the attributes listed in ProjectionExpression
. This return value is equivalent to specifying ProjectionExpression
without specifying any value for Select
.
If you query or scan a local secondary index and request only attributes that are projected into that index, the operation will read only the index and not the table. If any of the requested attributes are not projected into the local secondary index, DynamoDB fetches each of these attributes from the parent table. This extra fetching incurs additional throughput cost and latency.
If you query or scan a global secondary index, you can only request attributes that are projected into the index. Global secondary index queries cannot fetch attributes from the parent table.
If neither Select
nor ProjectionExpression
are specified, DynamoDB defaults to ALL_ATTRIBUTES
when accessing a table, and ALL_PROJECTED_ATTRIBUTES
when accessing an index. You cannot use both Select
and ProjectionExpression
together in a single request, unless the value for Select
is SPECIFIC_ATTRIBUTES
. (This usage is equivalent to specifying ProjectionExpression
without any value for Select
.)
If you use the ProjectionExpression
parameter, then the value for Select
can only be SPECIFIC_ATTRIBUTES
. Any other value for Select
will return an error.
The attributes to be returned in the result. You can retrieve all item attributes, specific item attributes, the count of matching items, or in the case of an index, some or all of the attributes projected into the index.
ALL_ATTRIBUTES
- Returns all of the item attributes from the specified table or index. If you query a local secondary index, then for each matching item in the index, DynamoDB fetches the entire item from the parent table. If the index is configured to project all item attributes, then all of the data can be obtained from the local secondary index, and no fetching is required.
ALL_PROJECTED_ATTRIBUTES
- Allowed only when querying an index. Retrieves all attributes that have been projected into the index. If the index is configured to project all attributes, this return value is equivalent to specifying ALL_ATTRIBUTES
.
COUNT
- Returns the number of matching items, rather than the matching items themselves. Note that this uses the same quantity of read capacity units as getting the items, and is subject to the same item size calculations.
SPECIFIC_ATTRIBUTES
- Returns only the attributes listed in ProjectionExpression
. This return value is equivalent to specifying ProjectionExpression
without specifying any value for Select
.
If you query or scan a local secondary index and request only attributes that are projected into that index, the operation reads only the index and not the table. If any of the requested attributes are not projected into the local secondary index, DynamoDB fetches each of these attributes from the parent table. This extra fetching incurs additional throughput cost and latency.
If you query or scan a global secondary index, you can only request attributes that are projected into the index. Global secondary index queries cannot fetch attributes from the parent table.
If neither Select
nor ProjectionExpression
are specified, DynamoDB defaults to ALL_ATTRIBUTES
when accessing a table, and ALL_PROJECTED_ATTRIBUTES
when accessing an index. You cannot use both Select
and ProjectionExpression
together in a single request, unless the value for Select
is SPECIFIC_ATTRIBUTES
. (This usage is equivalent to specifying ProjectionExpression
without any value for Select
.)
If you use the ProjectionExpression
parameter, then the value for Select
can only be SPECIFIC_ATTRIBUTES
. Any other value for Select
will return an error.
The entire transaction request was canceled.
DynamoDB cancels a TransactWriteItems
request under the following circumstances:
A condition in one of the condition expressions is not met.
A table in the TransactWriteItems
request is in a different account or region.
More than one action in the TransactWriteItems
operation targets the same item.
There is insufficient provisioned capacity for the transaction to be completed.
An item size becomes too large (larger than 400 KB), or a local secondary index (LSI) becomes too large, or a similar validation error occurs because of changes made by the transaction.
There is a user error, such as an invalid data format.
DynamoDB cancels a TransactGetItems
request under the following circumstances:
There is an ongoing TransactGetItems
operation that conflicts with a concurrent PutItem
, UpdateItem
, DeleteItem
or TransactWriteItems
request. In this case the TransactGetItems
operation fails with a TransactionCanceledException
.
A table in the TransactGetItems
request is in a different account or region.
There is insufficient provisioned capacity for the transaction to be completed.
There is a user error, such as an invalid data format.
If using Java, DynamoDB lists the cancellation reasons on the CancellationReasons
property. This property is not set for other languages. Transaction cancellation reasons are ordered in the order of requested items, if an item has no error it will have None
code and Null
message.
Cancellation reason codes and possible error messages:
No Errors:
Code: None
Message: null
Conditional Check Failed:
Code: ConditionalCheckFailed
Message: The conditional request failed.
Item Collection Size Limit Exceeded:
Code: ItemCollectionSizeLimitExceeded
Message: Collection size exceeded.
Transaction Conflict:
Code: TransactionConflict
Message: Transaction is ongoing for the item.
Provisioned Throughput Exceeded:
Code: ProvisionedThroughputExceeded
Messages:
The level of configured provisioned throughput for the table was exceeded. Consider increasing your provisioning level with the UpdateTable API.
This Message is received when provisioned throughput is exceeded is on a provisioned DynamoDB table.
The level of configured provisioned throughput for one or more global secondary indexes of the table was exceeded. Consider increasing your provisioning level for the under-provisioned global secondary indexes with the UpdateTable API.
This message is returned when provisioned throughput is exceeded is on a provisioned GSI.
Throttling Error:
Code: ThrottlingError
Messages:
Throughput exceeds the current capacity of your table or index. DynamoDB is automatically scaling your table or index so please try again shortly. If exceptions persist, check if you have a hot key: https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/bp-partition-key-design.html.
This message is returned when writes get throttled on an On-Demand table as DynamoDB is automatically scaling the table.
Throughput exceeds the current capacity for one or more global secondary indexes. DynamoDB is automatically scaling your index so please try again shortly.
This message is returned when when writes get throttled on an On-Demand GSI as DynamoDB is automatically scaling the GSI.
Validation Error:
Code: ValidationError
Messages:
One or more parameter values were invalid.
The update expression attempted to update the secondary index key beyond allowed size limits.
The update expression attempted to update the secondary index key to unsupported type.
An operand in the update expression has an incorrect data type.
Item size to update has exceeded the maximum allowed size.
Number overflow. Attempting to store a number with magnitude larger than supported range.
Type mismatch for attribute to update.
Nesting Levels have exceeded supported limits.
The document path provided in the update expression is invalid for update.
The provided expression refers to an attribute that does not exist in the item.
The entire transaction request was canceled.
DynamoDB cancels a TransactWriteItems
request under the following circumstances:
A condition in one of the condition expressions is not met.
A table in the TransactWriteItems
request is in a different account or region.
More than one action in the TransactWriteItems
operation targets the same item.
There is insufficient provisioned capacity for the transaction to be completed.
An item size becomes too large (larger than 400 KB), or a local secondary index (LSI) becomes too large, or a similar validation error occurs because of changes made by the transaction.
There is a user error, such as an invalid data format.
DynamoDB cancels a TransactGetItems
request under the following circumstances:
There is an ongoing TransactGetItems
operation that conflicts with a concurrent PutItem
, UpdateItem
, DeleteItem
or TransactWriteItems
request. In this case the TransactGetItems
operation fails with a TransactionCanceledException
.
A table in the TransactGetItems
request is in a different account or region.
There is insufficient provisioned capacity for the transaction to be completed.
There is a user error, such as an invalid data format.
If using Java, DynamoDB lists the cancellation reasons on the CancellationReasons
property. This property is not set for other languages. Transaction cancellation reasons are ordered in the order of requested items, if an item has no error it will have None
code and Null
message.
Cancellation reason codes and possible error messages:
No Errors:
Code: None
Message: null
Conditional Check Failed:
Code: ConditionalCheckFailed
Message: The conditional request failed.
Item Collection Size Limit Exceeded:
Code: ItemCollectionSizeLimitExceeded
Message: Collection size exceeded.
Transaction Conflict:
Code: TransactionConflict
Message: Transaction is ongoing for the item.
Provisioned Throughput Exceeded:
Code: ProvisionedThroughputExceeded
Messages:
The level of configured provisioned throughput for the table was exceeded. Consider increasing your provisioning level with the UpdateTable API.
This Message is received when provisioned throughput is exceeded is on a provisioned DynamoDB table.
The level of configured provisioned throughput for one or more global secondary indexes of the table was exceeded. Consider increasing your provisioning level for the under-provisioned global secondary indexes with the UpdateTable API.
This message is returned when provisioned throughput is exceeded is on a provisioned GSI.
Throttling Error:
Code: ThrottlingError
Messages:
Throughput exceeds the current capacity of your table or index. DynamoDB is automatically scaling your table or index so please try again shortly. If exceptions persist, check if you have a hot key: https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/bp-partition-key-design.html.
This message is returned when writes get throttled on an On-Demand table as DynamoDB is automatically scaling the table.
Throughput exceeds the current capacity for one or more global secondary indexes. DynamoDB is automatically scaling your index so please try again shortly.
This message is returned when writes get throttled on an On-Demand GSI as DynamoDB is automatically scaling the GSI.
Validation Error:
Code: ValidationError
Messages:
One or more parameter values were invalid.
The update expression attempted to update the secondary index key beyond allowed size limits.
The update expression attempted to update the secondary index key to unsupported type.
An operand in the update expression has an incorrect data type.
Item size to update has exceeded the maximum allowed size.
Number overflow. Attempting to store a number with magnitude larger than supported range.
Type mismatch for attribute to update.
Nesting Levels have exceeded supported limits.
The document path provided in the update expression is invalid for update.
The provided expression refers to an attribute that does not exist in the item.
The transaction with the given request token is already in progress.
", + "base": "The transaction with the given request token is already in progress.
Recommended Settings
This is a general recommendation for handling the TransactionInProgressException
. These settings help ensure that the client retries will trigger completion of the ongoing TransactWriteItems
request.
Set clientExecutionTimeout
to a value that allows at least one retry to be processed after 5 seconds have elapsed since the first attempt for the TransactWriteItems
operation.
Set socketTimeout
to a value a little lower than the requestTimeout
setting.
requestTimeout
should be set based on the time taken for the individual retries of a single HTTP request for your use case, but setting it to 1 second or higher should work well to reduce chances of retries and TransactionInProgressException
errors.
Use exponential backoff when retrying and tune backoff if needed.
Assuming default retry policy, example timeout settings based on the guidelines above are as follows:
Example timeline:
0-1000 first attempt
1000-1500 first sleep/delay (default retry policy uses 500 ms as base delay for 4xx errors)
1500-2500 second attempt
2500-3500 second sleep/delay (500 * 2, exponential backoff)
3500-4500 third attempt
4500-6500 third sleep/delay (500 * 2^2)
6500-7500 fourth attempt (this can trigger inline recovery since 5 seconds have elapsed since the first attempt reached TC)
The boot mode of the image. For more information, see Boot modes in the Amazon EC2 User Guide.
", "ImportImageRequest$BootMode": "The boot mode of the virtual machine.
", "ImportImageTask$BootMode": "The boot mode of the virtual machine.
", - "Instance$BootMode": "The boot mode of the instance. For more information, see Boot modes in the Amazon EC2 User Guide.
", - "RegisterImageRequest$BootMode": "The boot mode of the AMI. For more information, see Boot modes in the Amazon EC2 User Guide.
" + "Instance$BootMode": "The boot mode that was specified by the AMI. If the value is uefi-preferred
, the AMI supports both UEFI and Legacy BIOS. The currentInstanceBootMode
parameter is the boot mode that is used to boot the instance at launch or start.
The operating system contained in the AMI must be configured to support the specified boot mode.
For more information, see Boot modes in the Amazon EC2 User Guide.
", + "RegisterImageRequest$BootMode": "The boot mode of the AMI. A value of uefi-preferred
indicates that the AMI supports both UEFI and Legacy BIOS.
The operating system contained in the AMI must be configured to support the specified boot mode.
For more information, see Boot modes in the Amazon EC2 User Guide.
" } }, "BoxedDouble": { @@ -10623,6 +10623,12 @@ "ModifyInstanceAttributeRequest$BlockDeviceMappings": "Modifies the DeleteOnTermination
attribute for volumes that are currently attached. The volume must be owned by the caller. If no value is specified for DeleteOnTermination
, the default is true
and the volume is deleted when the instance is terminated.
To add instance store volumes to an Amazon EBS-backed instance, you must add them when you launch the instance. For more information, see Update the block device mapping when launching an instance in the Amazon EC2 User Guide.
" } }, + "InstanceBootModeValues": { + "base": null, + "refs": { + "Instance$CurrentInstanceBootMode": "The boot mode that is used to boot the instance at launch or start. For more information, see Boot modes in the Amazon EC2 User Guide.
" + } + }, "InstanceCapacity": { "base": "Information about the number of instances that can be launched onto the Dedicated Host.
", "refs": { diff --git a/models/apis/ec2/2016-11-15/endpoint-tests-1.json b/models/apis/ec2/2016-11-15/endpoint-tests-1.json index f438a23ab83..49ea940f100 100644 --- a/models/apis/ec2/2016-11-15/endpoint-tests-1.json +++ b/models/apis/ec2/2016-11-15/endpoint-tests-1.json @@ -8,9 +8,9 @@ } }, "params": { - "Region": "af-south-1", "UseFIPS": false, - "UseDualStack": false + "UseDualStack": false, + "Region": "af-south-1" } }, { @@ -21,9 +21,9 @@ } }, "params": { - "Region": "ap-east-1", "UseFIPS": false, - "UseDualStack": false + "UseDualStack": false, + "Region": "ap-east-1" } }, { @@ -34,9 +34,9 @@ } }, "params": { - "Region": "ap-northeast-1", "UseFIPS": false, - "UseDualStack": false + "UseDualStack": false, + "Region": "ap-northeast-1" } }, { @@ -47,9 +47,9 @@ } }, "params": { - "Region": "ap-northeast-2", "UseFIPS": false, - "UseDualStack": false + "UseDualStack": false, + "Region": "ap-northeast-2" } }, { @@ -60,9 +60,9 @@ } }, "params": { - "Region": "ap-northeast-3", "UseFIPS": false, - "UseDualStack": false + "UseDualStack": false, + "Region": "ap-northeast-3" } }, { @@ -73,9 +73,9 @@ } }, "params": { - "Region": "ap-south-1", "UseFIPS": false, - "UseDualStack": false + "UseDualStack": false, + "Region": "ap-south-1" } }, { @@ -86,9 +86,9 @@ } }, "params": { - "Region": "ap-south-1", "UseFIPS": false, - "UseDualStack": true + "UseDualStack": true, + "Region": "ap-south-1" } }, { @@ -99,9 +99,9 @@ } }, "params": { - "Region": "ap-southeast-1", "UseFIPS": false, - "UseDualStack": false + "UseDualStack": false, + "Region": "ap-southeast-1" } }, { @@ -112,9 +112,9 @@ } }, "params": { - "Region": "ap-southeast-2", "UseFIPS": false, - "UseDualStack": false + "UseDualStack": false, + "Region": "ap-southeast-2" } }, { @@ -125,9 +125,9 @@ } }, "params": { - "Region": "ap-southeast-3", "UseFIPS": false, - "UseDualStack": false + "UseDualStack": false, + "Region": "ap-southeast-3" } }, { @@ -138,9 +138,9 @@ } }, "params": { - "Region": "ca-central-1", "UseFIPS": false, - "UseDualStack": false + "UseDualStack": false, + "Region": "ca-central-1" } }, { @@ -151,9 +151,9 @@ } }, "params": { - "Region": "ca-central-1", "UseFIPS": true, - "UseDualStack": false + "UseDualStack": false, + "Region": "ca-central-1" } }, { @@ -164,9 +164,9 @@ } }, "params": { - "Region": "eu-central-1", "UseFIPS": false, - "UseDualStack": false + "UseDualStack": false, + "Region": "eu-central-1" } }, { @@ -177,9 +177,9 @@ } }, "params": { - "Region": "eu-north-1", "UseFIPS": false, - "UseDualStack": false + "UseDualStack": false, + "Region": "eu-north-1" } }, { @@ -190,9 +190,9 @@ } }, "params": { - "Region": "eu-south-1", "UseFIPS": false, - "UseDualStack": false + "UseDualStack": false, + "Region": "eu-south-1" } }, { @@ -203,9 +203,9 @@ } }, "params": { - "Region": "eu-west-1", "UseFIPS": false, - "UseDualStack": false + "UseDualStack": false, + "Region": "eu-west-1" } }, { @@ -216,9 +216,9 @@ } }, "params": { - "Region": "eu-west-1", "UseFIPS": false, - "UseDualStack": true + "UseDualStack": true, + "Region": "eu-west-1" } }, { @@ -229,9 +229,9 @@ } }, "params": { - "Region": "eu-west-2", "UseFIPS": false, - "UseDualStack": false + "UseDualStack": false, + "Region": "eu-west-2" } }, { @@ -242,9 +242,9 @@ } }, "params": { - "Region": "eu-west-3", "UseFIPS": false, - "UseDualStack": false + "UseDualStack": false, + "Region": "eu-west-3" } }, { @@ -255,9 +255,9 @@ } }, "params": { - "Region": "me-south-1", "UseFIPS": false, - "UseDualStack": false + "UseDualStack": false, + "Region": "me-south-1" } }, { @@ -268,9 +268,9 @@ } }, "params": { - "Region": "sa-east-1", "UseFIPS": false, - "UseDualStack": false + "UseDualStack": false, + "Region": "sa-east-1" } }, { @@ -281,9 +281,9 @@ } }, "params": { - "Region": "sa-east-1", "UseFIPS": false, - "UseDualStack": true + "UseDualStack": true, + "Region": "sa-east-1" } }, { @@ -294,9 +294,9 @@ } }, "params": { - "Region": "us-east-1", "UseFIPS": false, - "UseDualStack": false + "UseDualStack": false, + "Region": "us-east-1" } }, { @@ -307,9 +307,9 @@ } }, "params": { - "Region": "us-east-1", "UseFIPS": true, - "UseDualStack": false + "UseDualStack": false, + "Region": "us-east-1" } }, { @@ -320,9 +320,9 @@ } }, "params": { - "Region": "us-east-1", "UseFIPS": false, - "UseDualStack": true + "UseDualStack": true, + "Region": "us-east-1" } }, { @@ -333,9 +333,9 @@ } }, "params": { - "Region": "us-east-2", "UseFIPS": false, - "UseDualStack": false + "UseDualStack": false, + "Region": "us-east-2" } }, { @@ -346,9 +346,9 @@ } }, "params": { - "Region": "us-east-2", "UseFIPS": true, - "UseDualStack": false + "UseDualStack": false, + "Region": "us-east-2" } }, { @@ -359,9 +359,9 @@ } }, "params": { - "Region": "us-east-2", "UseFIPS": false, - "UseDualStack": true + "UseDualStack": true, + "Region": "us-east-2" } }, { @@ -372,9 +372,9 @@ } }, "params": { - "Region": "us-west-1", "UseFIPS": false, - "UseDualStack": false + "UseDualStack": false, + "Region": "us-west-1" } }, { @@ -385,9 +385,9 @@ } }, "params": { - "Region": "us-west-1", "UseFIPS": true, - "UseDualStack": false + "UseDualStack": false, + "Region": "us-west-1" } }, { @@ -398,9 +398,9 @@ } }, "params": { - "Region": "us-west-2", "UseFIPS": false, - "UseDualStack": false + "UseDualStack": false, + "Region": "us-west-2" } }, { @@ -411,9 +411,9 @@ } }, "params": { - "Region": "us-west-2", "UseFIPS": true, - "UseDualStack": false + "UseDualStack": false, + "Region": "us-west-2" } }, { @@ -424,9 +424,9 @@ } }, "params": { - "Region": "us-west-2", "UseFIPS": false, - "UseDualStack": true + "UseDualStack": true, + "Region": "us-west-2" } }, { @@ -437,9 +437,9 @@ } }, "params": { - "Region": "us-east-1", "UseFIPS": true, - "UseDualStack": true + "UseDualStack": true, + "Region": "us-east-1" } }, { @@ -450,9 +450,9 @@ } }, "params": { - "Region": "cn-north-1", "UseFIPS": false, - "UseDualStack": false + "UseDualStack": false, + "Region": "cn-north-1" } }, { @@ -463,9 +463,9 @@ } }, "params": { - "Region": "cn-northwest-1", "UseFIPS": false, - "UseDualStack": false + "UseDualStack": false, + "Region": "cn-northwest-1" } }, { @@ -476,9 +476,9 @@ } }, "params": { - "Region": "cn-north-1", "UseFIPS": true, - "UseDualStack": true + "UseDualStack": true, + "Region": "cn-north-1" } }, { @@ -489,9 +489,9 @@ } }, "params": { - "Region": "cn-north-1", "UseFIPS": true, - "UseDualStack": false + "UseDualStack": false, + "Region": "cn-north-1" } }, { @@ -502,9 +502,9 @@ } }, "params": { - "Region": "cn-north-1", "UseFIPS": false, - "UseDualStack": true + "UseDualStack": true, + "Region": "cn-north-1" } }, { @@ -515,9 +515,9 @@ } }, "params": { - "Region": "us-gov-east-1", "UseFIPS": false, - "UseDualStack": false + "UseDualStack": false, + "Region": "us-gov-east-1" } }, { @@ -528,9 +528,9 @@ } }, "params": { - "Region": "us-gov-west-1", "UseFIPS": false, - "UseDualStack": false + "UseDualStack": false, + "Region": "us-gov-west-1" } }, { @@ -541,9 +541,9 @@ } }, "params": { - "Region": "us-gov-east-1", "UseFIPS": true, - "UseDualStack": true + "UseDualStack": true, + "Region": "us-gov-east-1" } }, { @@ -554,9 +554,9 @@ } }, "params": { - "Region": "us-gov-east-1", "UseFIPS": true, - "UseDualStack": false + "UseDualStack": false, + "Region": "us-gov-east-1" } }, { @@ -567,9 +567,9 @@ } }, "params": { - "Region": "us-gov-east-1", "UseFIPS": false, - "UseDualStack": true + "UseDualStack": true, + "Region": "us-gov-east-1" } }, { @@ -580,9 +580,9 @@ } }, "params": { - "Region": "us-iso-east-1", "UseFIPS": false, - "UseDualStack": false + "UseDualStack": false, + "Region": "us-iso-east-1" } }, { @@ -593,9 +593,9 @@ } }, "params": { - "Region": "us-iso-west-1", "UseFIPS": false, - "UseDualStack": false + "UseDualStack": false, + "Region": "us-iso-west-1" } }, { @@ -606,9 +606,9 @@ } }, "params": { - "Region": "us-iso-east-1", "UseFIPS": true, - "UseDualStack": false + "UseDualStack": false, + "Region": "us-iso-east-1" } }, { @@ -619,9 +619,9 @@ } }, "params": { - "Region": "us-isob-east-1", "UseFIPS": false, - "UseDualStack": false + "UseDualStack": false, + "Region": "us-isob-east-1" } }, { @@ -632,9 +632,9 @@ } }, "params": { - "Region": "us-isob-east-1", "UseFIPS": true, - "UseDualStack": false + "UseDualStack": false, + "Region": "us-isob-east-1" } }, { @@ -645,9 +645,9 @@ } }, "params": { - "Region": "us-east-1", "UseFIPS": false, "UseDualStack": false, + "Region": "us-east-1", "Endpoint": "https://example.com" } }, @@ -670,9 +670,9 @@ "error": "Invalid Configuration: FIPS and custom endpoint are not supported" }, "params": { - "Region": "us-east-1", "UseFIPS": true, "UseDualStack": false, + "Region": "us-east-1", "Endpoint": "https://example.com" } }, @@ -682,9 +682,9 @@ "error": "Invalid Configuration: Dualstack and custom endpoint are not supported" }, "params": { - "Region": "us-east-1", "UseFIPS": false, "UseDualStack": true, + "Region": "us-east-1", "Endpoint": "https://example.com" } } diff --git a/models/apis/macie2/2020-01-01/docs-2.json b/models/apis/macie2/2020-01-01/docs-2.json index 5c0ed9ee4ff..d5f34502bc2 100644 --- a/models/apis/macie2/2020-01-01/docs-2.json +++ b/models/apis/macie2/2020-01-01/docs-2.json @@ -150,8 +150,8 @@ "AllowsUnencryptedObjectUploads" : { "base" : null, "refs" : { - "BucketMetadata$AllowsUnencryptedObjectUploads" : "Specifies whether the bucket policy for the bucket requires server-side encryption of objects when objects are uploaded to the bucket. Possible values are:
FALSE - The bucket policy requires server-side encryption of new objects. PutObject requests must include a valid server-side encryption header.
TRUE - The bucket doesn't have a bucket policy or it has a bucket policy that doesn't require server-side encryption of new objects. If a bucket policy exists, it doesn't require PutObject requests to include a valid server-side encryption header.
UNKNOWN - Amazon Macie can't determine whether the bucket policy requires server-side encryption of new objects.
Valid server-side encryption headers are: x-amz-server-side-encryption with a value of AES256 or aws:kms, and x-amz-server-side-encryption-customer-algorithm with a value of AES256.
", - "S3Bucket$AllowsUnencryptedObjectUploads" : "Specifies whether the bucket policy for the bucket requires server-side encryption of objects when objects are uploaded to the bucket. Possible values are:
FALSE - The bucket policy requires server-side encryption of new objects. PutObject requests must include a valid server-side encryption header.
TRUE - The bucket doesn't have a bucket policy or it has a bucket policy that doesn't require server-side encryption of new objects. If a bucket policy exists, it doesn't require PutObject requests to include a valid server-side encryption header.
UNKNOWN - Amazon Macie can't determine whether the bucket policy requires server-side encryption of new objects.
Valid server-side encryption headers are: x-amz-server-side-encryption with a value of AES256 or aws:kms, and x-amz-server-side-encryption-customer-algorithm with a value of AES256.
" + "BucketMetadata$AllowsUnencryptedObjectUploads" : "Specifies whether the bucket policy for the bucket requires server-side encryption of objects when objects are added to the bucket. Possible values are:
FALSE - The bucket policy requires server-side encryption of new objects. PutObject requests must include a valid server-side encryption header.
TRUE - The bucket doesn't have a bucket policy or it has a bucket policy that doesn't require server-side encryption of new objects. If a bucket policy exists, it doesn't require PutObject requests to include a valid server-side encryption header.
UNKNOWN - Amazon Macie can't determine whether the bucket policy requires server-side encryption of new objects.
Valid server-side encryption headers are: x-amz-server-side-encryption with a value of AES256 or aws:kms, and x-amz-server-side-encryption-customer-algorithm with a value of AES256.
", + "S3Bucket$AllowsUnencryptedObjectUploads" : "Specifies whether the bucket policy for the bucket requires server-side encryption of objects when objects are added to the bucket. Possible values are:
FALSE - The bucket policy requires server-side encryption of new objects. PutObject requests must include a valid server-side encryption header.
TRUE - The bucket doesn't have a bucket policy or it has a bucket policy that doesn't require server-side encryption of new objects. If a bucket policy exists, it doesn't require PutObject requests to include a valid server-side encryption header.
UNKNOWN - Amazon Macie can't determine whether the bucket policy requires server-side encryption of new objects.
Valid server-side encryption headers are: x-amz-server-side-encryption with a value of AES256 or aws:kms, and x-amz-server-side-encryption-customer-algorithm with a value of AES256.
" } }, "ApiCallDetails" : { @@ -219,9 +219,9 @@ } }, "BucketCountByEncryptionType" : { - "base" : "Provides information about the number of S3 buckets that use certain types of server-side encryption by default or don't encrypt new objects by default. For detailed information about these settings, see Setting default server-side encryption behavior for Amazon S3 buckets in the Amazon Simple Storage Service User Guide.
", + "base" : "Provides information about the number of S3 buckets whose settings do or don't specify default server-side encryption behavior for objects that are added to the buckets. For detailed information about these settings, see Setting default server-side encryption behavior for Amazon S3 buckets in the Amazon Simple Storage Service User Guide.
", "refs" : { - "GetBucketStatisticsResponse$BucketCountByEncryptionType" : "The total number of buckets that use certain types of server-side encryption to encrypt new objects by default. This object also reports the total number of buckets that don't encrypt new objects by default.
" + "GetBucketStatisticsResponse$BucketCountByEncryptionType" : "The total number of buckets whose settings do or don't specify default server-side encryption behavior for objects that are added to the buckets.
" } }, "BucketCountBySharedAccessType" : { @@ -231,9 +231,9 @@ } }, "BucketCountPolicyAllowsUnencryptedObjectUploads" : { - "base" : "Provides information about the number of S3 buckets whose bucket policies do or don't require server-side encryption of objects when objects are uploaded to the buckets.
", + "base" : "Provides information about the number of S3 buckets whose bucket policies do or don't require server-side encryption of objects when objects are added to the buckets.
", "refs" : { - "GetBucketStatisticsResponse$BucketCountByObjectEncryptionRequirement" : "The total number of buckets whose bucket policies do or don't require server-side encryption of objects when objects are uploaded to the buckets.
" + "GetBucketStatisticsResponse$BucketCountByObjectEncryptionRequirement" : "The total number of buckets whose bucket policies do or don't require server-side encryption of objects when objects are added to the buckets.
" } }, "BucketCriteria" : { @@ -289,7 +289,7 @@ "BucketServerSideEncryption" : { "base" : "Provides information about the default server-side encryption settings for an S3 bucket. For detailed information about these settings, see Setting default server-side encryption behavior for Amazon S3 buckets in the Amazon Simple Storage Service User Guide.
", "refs" : { - "BucketMetadata$ServerSideEncryption" : "Specifies whether the bucket encrypts new objects by default and, if so, the type of server-side encryption that's used.
" + "BucketMetadata$ServerSideEncryption" : "The default server-side encryption settings for the bucket.
" } }, "BucketSortCriteria" : { @@ -587,9 +587,9 @@ "refs" : { } }, "EncryptionType" : { - "base" : "The type of server-side encryption that's used to encrypt an S3 object or objects in an S3 bucket. Possible values are:
", + "base" : "The server-side encryption algorithm that was used to encrypt an S3 object or is used by default to encrypt objects that are added to an S3 bucket. Possible values are:
", "refs" : { - "ServerSideEncryption$EncryptionType" : "The server-side encryption algorithm that's used when storing data in the bucket or object. If default encryption is disabled for the bucket or the object isn't encrypted using server-side encryption, this value is NONE.
" + "ServerSideEncryption$EncryptionType" : "The server-side encryption algorithm that's used when storing data in the bucket or object. If default encryption settings aren't configured for the bucket or the object isn't encrypted using server-side encryption, this value is NONE.
" } }, "ErrorCode" : { @@ -1488,10 +1488,10 @@ } }, "ServerSideEncryption" : { - "base" : "Provides information about the server-side encryption settings for an S3 bucket or S3 object.
", + "base" : "Provides information about the default server-side encryption settings for an S3 bucket or the encryption settings for an S3 object.
", "refs" : { - "S3Bucket$DefaultServerSideEncryption" : "The type of server-side encryption that's used by default to encrypt objects in the bucket.
", - "S3Object$ServerSideEncryption" : "The type of server-side encryption that's used to encrypt the object.
" + "S3Bucket$DefaultServerSideEncryption" : "The default server-side encryption settings for the bucket.
", + "S3Object$ServerSideEncryption" : "The type of server-side encryption that was used to encrypt the object.
" } }, "ServiceLimit" : { @@ -1551,7 +1551,7 @@ "SharedAccess" : { "base" : null, "refs" : { - "BucketMetadata$SharedAccess" : "Specifies whether the bucket is shared with another Amazon Web Services account, an Amazon CloudFront origin access identity (OAI), or a CloudFront origin access control (OAC). Possible values are:
EXTERNAL - The bucket is shared with one or more of the following or any combination of the following: an Amazon Web Services account that isn't part of your Amazon Macie organization, a CloudFront OAI, or a CloudFront OAC.
INTERNAL - The bucket is shared with one or more Amazon Web Services accounts that are part of your Amazon Macie organization. It isn't shared with a CloudFront OAI or OAC.
NOT_SHARED - The bucket isn't shared with another Amazon Web Services account, a CloudFront OAI, or a CloudFront OAC.
UNKNOWN - Amazon Macie wasn't able to evaluate the shared access settings for the bucket.
An Amazon Macie organization is a set of Macie accounts that are centrally managed as a group of related accounts through Organizations or by Macie invitation.
" + "BucketMetadata$SharedAccess" : "Specifies whether the bucket is shared with another Amazon Web Services account, an Amazon CloudFront origin access identity (OAI), or a CloudFront origin access control (OAC). Possible values are:
EXTERNAL - The bucket is shared with one or more of the following or any combination of the following: a CloudFront OAI, a CloudFront OAC, or an Amazon Web Services account that isn't part of your Amazon Macie organization.
INTERNAL - The bucket is shared with one or more Amazon Web Services accounts that are part of your Amazon Macie organization. It isn't shared with a CloudFront OAI or OAC.
NOT_SHARED - The bucket isn't shared with another Amazon Web Services account, a CloudFront OAI, or a CloudFront OAC.
UNKNOWN - Amazon Macie wasn't able to evaluate the shared access settings for the bucket.
An Amazon Macie organization is a set of Macie accounts that are centrally managed as a group of related accounts through Organizations or by Macie invitation.
" } }, "SimpleCriterionForJob" : { @@ -1685,7 +1685,7 @@ "Type" : { "base" : null, "refs" : { - "BucketServerSideEncryption$Type" : "The type of server-side encryption that's used by default when storing new objects in the bucket. Possible values are:
AES256 - New objects are encrypted with an Amazon S3 managed key. They use SSE-S3 encryption.
aws:kms - New objects are encrypted with an KMS key (kmsMasterKeyId), either an Amazon Web Services managed key or a customer managed key. They use SSE-KMS encryption.
NONE - New objects aren't encrypted by default. Default encryption is disabled for the bucket.
The server-side encryption algorithm that's used by default to encrypt objects that are added to the bucket. Possible values are:
AES256 - New objects are encrypted with an Amazon S3 managed key. They use SSE-S3 encryption.
aws:kms - New objects are encrypted with an KMS key (kmsMasterKeyId), either an Amazon Web Services managed key or a customer managed key. They use SSE-KMS encryption.
NONE - The bucket's default encryption settings don't specify server-side encryption behavior for new objects.
The total number of buckets that allow the general public to have read access to the bucket.
", "BucketCountByEffectivePermission$PubliclyWritable" : "The total number of buckets that allow the general public to have write access to the bucket.
", "BucketCountByEffectivePermission$Unknown" : "The total number of buckets that Amazon Macie wasn't able to evaluate permissions settings for. Macie can't determine whether these buckets are publicly accessible.
", - "BucketCountByEncryptionType$KmsManaged" : "The total number of buckets that use an KMS key to encrypt new objects by default, either an Amazon Web Services managed key or a customer managed key. These buckets use KMS encryption (SSE-KMS) by default.
", - "BucketCountByEncryptionType$S3Managed" : "The total number of buckets that use an Amazon S3 managed key to encrypt new objects by default. These buckets use Amazon S3 managed encryption (SSE-S3) by default.
", - "BucketCountByEncryptionType$Unencrypted" : "The total number of buckets that don't encrypt new objects by default. Default encryption is disabled for these buckets.
", + "BucketCountByEncryptionType$KmsManaged" : "The total number of buckets whose default encryption settings are configured to encrypt new objects with an Amazon Web Services managed KMS key or a customer managed KMS key. By default, these buckets encrypt new objects automatically using SSE-KMS encryption.
", + "BucketCountByEncryptionType$S3Managed" : "The total number of buckets whose default encryption settings are configured to encrypt new objects with an Amazon S3 managed key. By default, these buckets encrypt new objects automatically using SSE-S3 encryption.
", + "BucketCountByEncryptionType$Unencrypted" : "The total number of buckets that don't specify default server-side encryption behavior for new objects. Default encryption settings aren't configured for these buckets.
", "BucketCountByEncryptionType$Unknown" : "The total number of buckets that Amazon Macie doesn't have current encryption metadata for. Macie can't provide current data about the default encryption settings for these buckets.
", - "BucketCountBySharedAccessType$External" : "The total number of buckets that are shared with one or more of the following or any combination of the following: an Amazon Web Services account that isn't in the same Amazon Macie organization, an Amazon CloudFront OAI, or a CloudFront OAC.
", + "BucketCountBySharedAccessType$External" : "The total number of buckets that are shared with one or more of the following or any combination of the following: an Amazon CloudFront OAI, a CloudFront OAC, or an Amazon Web Services account that isn't in the same Amazon Macie organization.
", "BucketCountBySharedAccessType$Internal" : "The total number of buckets that are shared with one or more Amazon Web Services accounts in the same Amazon Macie organization. These buckets aren't shared with Amazon CloudFront OAIs or OACs.
", "BucketCountBySharedAccessType$NotShared" : "The total number of buckets that aren't shared with other Amazon Web Services accounts, Amazon CloudFront OAIs, or CloudFront OACs.
", "BucketCountBySharedAccessType$Unknown" : "The total number of buckets that Amazon Macie wasn't able to evaluate shared access settings for. Macie can't determine whether these buckets are shared with other Amazon Web Services accounts, Amazon CloudFront OAIs, or CloudFront OACs.
", @@ -2250,7 +2250,7 @@ "ObjectCountByEncryptionType$CustomerManaged" : "The total number of objects that are encrypted with a customer-provided key. The objects use customer-provided server-side encryption (SSE-C).
", "ObjectCountByEncryptionType$KmsManaged" : "The total number of objects that are encrypted with an KMS key, either an Amazon Web Services managed key or a customer managed key. The objects use KMS encryption (SSE-KMS).
", "ObjectCountByEncryptionType$S3Managed" : "The total number of objects that are encrypted with an Amazon S3 managed key. The objects use Amazon S3 managed encryption (SSE-S3).
", - "ObjectCountByEncryptionType$Unencrypted" : "The total number of objects that aren't encrypted or use client-side encryption.
", + "ObjectCountByEncryptionType$Unencrypted" : "The total number of objects that use client-side encryption or aren't encrypted.
", "ObjectCountByEncryptionType$Unknown" : "The total number of objects that Amazon Macie doesn't have current encryption metadata for. Macie can't provide current data about the encryption settings for these objects.
", "ObjectLevelStatistics$FileType" : "The total storage size (in bytes) or number of objects that Amazon Macie can't analyze because the objects don't have a file name extension for a supported file or storage format.
", "ObjectLevelStatistics$StorageClass" : "The total storage size (in bytes) or number of objects that Amazon Macie can't analyze because the objects use an unsupported storage class.
", @@ -2309,7 +2309,7 @@ "BucketMetadata$BucketName" : "The name of the bucket.
", "BucketMetadata$ErrorMessage" : "A brief description of the error (errorCode) that prevented Amazon Macie from retrieving and processing information about the bucket and the bucket's objects. This value is null if Macie was able to retrieve and process the information.
", "BucketMetadata$Region" : "The Amazon Web Services Region that hosts the bucket.
", - "BucketServerSideEncryption$KmsMasterKeyId" : "The Amazon Resource Name (ARN) or unique identifier (key ID) for the KMS key that's used by default to encrypt objects that are added to the bucket. This value is null if the bucket uses an Amazon S3 managed key to encrypt new objects or the bucket doesn't encrypt new objects by default.
", + "BucketServerSideEncryption$KmsMasterKeyId" : "The Amazon Resource Name (ARN) or unique identifier (key ID) for the KMS key that's used by default to encrypt objects that are added to the bucket. This value is null if the bucket is configured to use an Amazon S3 managed key to encrypt new objects.
", "BucketSortCriteria$AttributeName" : "The name of the bucket property to sort the results by. This value can be one of the following properties that Amazon Macie defines as bucket metadata: accountId, bucketName, classifiableObjectCount, classifiableSizeInBytes, objectCount, sensitivityScore, or sizeInBytes.
", "Cell$CellReference" : "The location of the cell, as an absolute cell reference, that contains the sensitive data, for example Sheet2!C5 for cell C5 on Sheet2 in a Microsoft Excel workbook. This value is null for CSV and TSV files.
", "Cell$ColumnName" : "The name of the column that contains the sensitive data, if available.
", @@ -2318,7 +2318,7 @@ "ClassificationDetails$JobId" : "The unique identifier for the classification job that produced the finding. This value is null if the origin of the finding (originType) is AUTOMATED_SENSITIVE_DATA_DISCOVERY.
", "ClassificationResult$MimeType" : "The type of content, as a MIME type, that the finding applies to. For example, application/gzip, for a GNU Gzip compressed archive file, or application/pdf, for an Adobe Portable Document Format file.
", "ClassificationResultStatus$Code" : "The status of the finding. Possible values are:
COMPLETE - Amazon Macie successfully completed its analysis of the S3 object that the finding applies to.
PARTIAL - Macie analyzed only a subset of the data in the S3 object that the finding applies to. For example, the object is an archive file that contains files in an unsupported format.
SKIPPED - Macie wasn't able to analyze the S3 object that the finding applies to. For example, the object is a file that uses an unsupported format.
A brief description of the status of the finding. This value is null if the status (code) of the finding is COMPLETE.
Amazon Macie uses this value to notify you of any errors, warnings, or considerations that might impact your analysis of the finding and the affected S3 object. Possible values are:
ARCHIVE_CONTAINS_UNPROCESSED_FILES - The object is an archive file and Macie extracted and analyzed only some or none of the files in the archive. To determine which files Macie analyzed, if any, refer to the corresponding sensitive data discovery result for the finding (ClassificationDetails.detailedResultsLocation).
ARCHIVE_EXCEEDS_SIZE_LIMIT - The object is an archive file whose total storage size exceeds the size quota for this type of archive.
ARCHIVE_NESTING_LEVEL_OVER_LIMIT - The object is an archive file whose nested depth exceeds the quota for the maximum number of nested levels that Macie analyzes for this type of archive.
ARCHIVE_TOTAL_BYTES_EXTRACTED_OVER_LIMIT - The object is an archive file that exceeds the quota for the maximum amount of data that Macie extracts and analyzes for this type of archive.
ARCHIVE_TOTAL_DOCUMENTS_PROCESSED_OVER_LIMIT - The object is an archive file that contains more than the maximum number of files that Macie extracts and analyzes for this type of archive.
FILE_EXCEEDS_SIZE_LIMIT - The storage size of the object exceeds the size quota for this type of file.
INVALID_ENCRYPTION - The object is encrypted using server-side encryption but Macie isn’t allowed to use the key. Macie can’t decrypt and analyze the object.
INVALID_KMS_KEY - The object is encrypted with an KMS key that was disabled or is being deleted. Macie can’t decrypt and analyze the object.
INVALID_OBJECT_STATE - The object doesn’t use a supported Amazon S3 storage class.
JSON_NESTING_LEVEL_OVER_LIMIT - The object contains JSON data and the nested depth of the data exceeds the quota for the number of nested levels that Macie analyzes for this type of file.
MALFORMED_FILE - The object is a malformed or corrupted file. An error occurred when Macie attempted to detect the file’s type or extract data from the file.
MALFORMED_OR_FILE_SIZE_EXCEEDS_LIMIT - The object is a Microsoft Office file that is malformed or exceeds the size quota for this type of file. If the file is malformed, an error occurred when Macie attempted to extract data from the file.
NO_SUCH_BUCKET_AVAILABLE - The object was in a bucket that was deleted shortly before or when Macie attempted to analyze the object.
OBJECT_VERSION_MISMATCH - The object was changed while Macie was analyzing it.
OOXML_UNCOMPRESSED_RATIO_EXCEEDS_LIMIT - The object is an Office Open XML file whose compression ratio exceeds the compression quota for this type of file.
OOXML_UNCOMPRESSED_SIZE_EXCEEDS_LIMIT - The object is an Office Open XML file that exceeds the size quota for this type of file.
PERMISSION_DENIED - Macie isn’t allowed to access the object. The object’s permissions settings prevent Macie from analyzing the object.
SOURCE_OBJECT_NO_LONGER_AVAILABLE - The object was deleted shortly before or when Macie attempted to analyze it.
TIME_CUT_OFF_REACHED - Macie started analyzing the object but additional analysis would exceed the time quota for analyzing an object.
UNABLE_TO_PARSE_FILE - The object is a file that contains structured data and an error occurred when Macie attempted to parse the data.
UNSUPPORTED_FILE_TYPE_EXCEPTION - The object is a file that uses an unsupported file or storage format.
For information about quotas, supported storage classes, and supported file and storage formats, see Quotas and Supported storage classes and formats in the Amazon Macie User Guide.
", + "ClassificationResultStatus$Reason" : "A brief description of the status of the finding. This value is null if the status (code) of the finding is COMPLETE.
Amazon Macie uses this value to notify you of any errors, warnings, or considerations that might impact your analysis of the finding and the affected S3 object. Possible values are:
ARCHIVE_CONTAINS_UNPROCESSED_FILES - The object is an archive file and Macie extracted and analyzed only some or none of the files in the archive. To determine which files Macie analyzed, if any, refer to the corresponding sensitive data discovery result for the finding (ClassificationDetails.detailedResultsLocation).
ARCHIVE_EXCEEDS_SIZE_LIMIT - The object is an archive file whose total storage size exceeds the size quota for this type of archive.
ARCHIVE_NESTING_LEVEL_OVER_LIMIT - The object is an archive file whose nested depth exceeds the quota for the maximum number of nested levels that Macie analyzes for this type of archive.
ARCHIVE_TOTAL_BYTES_EXTRACTED_OVER_LIMIT - The object is an archive file that exceeds the quota for the maximum amount of data that Macie extracts and analyzes for this type of archive.
ARCHIVE_TOTAL_DOCUMENTS_PROCESSED_OVER_LIMIT - The object is an archive file that contains more than the maximum number of files that Macie extracts and analyzes for this type of archive.
FILE_EXCEEDS_SIZE_LIMIT - The storage size of the object exceeds the size quota for this type of file.
INVALID_ENCRYPTION - The object is encrypted using server-side encryption but Macie isn't allowed to use the key. Macie can't decrypt and analyze the object.
INVALID_KMS_KEY - The object is encrypted with an KMS key that was disabled or is being deleted. Macie can't decrypt and analyze the object.
INVALID_OBJECT_STATE - The object doesn't use a supported Amazon S3 storage class.
JSON_NESTING_LEVEL_OVER_LIMIT - The object contains JSON data and the nested depth of the data exceeds the quota for the number of nested levels that Macie analyzes for this type of file.
MALFORMED_FILE - The object is a malformed or corrupted file. An error occurred when Macie attempted to detect the file's type or extract data from the file.
MALFORMED_OR_FILE_SIZE_EXCEEDS_LIMIT - The object is a Microsoft Office file that is malformed or exceeds the size quota for this type of file. If the file is malformed, an error occurred when Macie attempted to extract data from the file.
NO_SUCH_BUCKET_AVAILABLE - The object was in a bucket that was deleted shortly before or when Macie attempted to analyze the object.
OBJECT_VERSION_MISMATCH - The object was changed while Macie was analyzing it.
OOXML_UNCOMPRESSED_RATIO_EXCEEDS_LIMIT - The object is an Office Open XML file whose compression ratio exceeds the compression quota for this type of file.
OOXML_UNCOMPRESSED_SIZE_EXCEEDS_LIMIT - The object is an Office Open XML file that exceeds the size quota for this type of file.
PERMISSION_DENIED - Macie isn't allowed to access the object. The object's permissions settings prevent Macie from analyzing the object.
SOURCE_OBJECT_NO_LONGER_AVAILABLE - The object was deleted shortly before or when Macie attempted to analyze it.
TIME_CUT_OFF_REACHED - Macie started analyzing the object but additional analysis would exceed the time quota for analyzing an object.
UNABLE_TO_PARSE_FILE - The object is a file that contains structured data and an error occurred when Macie attempted to parse the data.
UNSUPPORTED_FILE_TYPE_EXCEPTION - The object is a file that uses an unsupported file or storage format.
For information about quotas, supported storage classes, and supported file and storage formats, see Quotas and Supported storage classes and formats in the Amazon Macie User Guide.
", "ConflictException$Message" : "The explanation of the error that occurred.
", "CreateAllowListRequest$ClientToken" : "A unique, case-sensitive token that you provide to ensure the idempotency of the request.
", "CreateClassificationJobRequest$ClientToken" : "A unique, case-sensitive token that you provide to ensure the idempotency of the request.
", @@ -2398,7 +2398,7 @@ "IamUser$AccountId" : "The unique identifier for the Amazon Web Services account that's associated with the IAM user who performed the action.
", "IamUser$Arn" : "The Amazon Resource Name (ARN) of the principal that performed the action. The last section of the ARN contains the name of the user who performed the action.
", "IamUser$PrincipalId" : "The unique identifier for the IAM user who performed the action.
", - "IamUser$UserName" : "The user name of the IAM user who performed the action.
", + "IamUser$UserName" : "The username of the IAM user who performed the action.
", "InternalServerException$Message" : "The explanation of the error that occurred.
", "Invitation$AccountId" : "The Amazon Web Services account ID for the account that sent the invitation.
", "Invitation$InvitationId" : "The unique identifier for the invitation.
", @@ -2463,8 +2463,8 @@ "SearchResourcesResponse$NextToken" : "The string to use in a subsequent request to get the next page of results in a paginated response. This value is null if there are no additional pages.
", "SearchResourcesTagCriterionPair$Key" : "The value for the tag key to use in the condition.
", "SearchResourcesTagCriterionPair$Value" : "The tag value to use in the condition.
", - "SensitivityInspectionTemplatesEntry$Id" : "The unique identifier for the sensitivity inspection template for the account.
", - "SensitivityInspectionTemplatesEntry$Name" : "The name of the sensitivity inspection template for the account: automated-sensitive-data-discovery.
", + "SensitivityInspectionTemplatesEntry$Id" : "The unique identifier for the sensitivity inspection template.
", + "SensitivityInspectionTemplatesEntry$Name" : "The name of the sensitivity inspection template: automated-sensitive-data-discovery.
", "ServerSideEncryption$KmsMasterKeyId" : "The Amazon Resource Name (ARN) or unique identifier (key ID) for the KMS key that's used to encrypt data in the bucket or the object. This value is null if an KMS key isn't used to encrypt the data.
", "ServiceQuotaExceededException$Message" : "The explanation of the error that occurred.
", "SessionIssuer$AccountId" : "The unique identifier for the Amazon Web Services account that owns the entity that was used to get the credentials.
", diff --git a/models/apis/macie2/2020-01-01/endpoint-tests-1.json b/models/apis/macie2/2020-01-01/endpoint-tests-1.json index 5954a322a0c..26119b4ea3c 100644 --- a/models/apis/macie2/2020-01-01/endpoint-tests-1.json +++ b/models/apis/macie2/2020-01-01/endpoint-tests-1.json @@ -8,9 +8,9 @@ } }, "params": { - "Region": "af-south-1", "UseDualStack": false, - "UseFIPS": false + "UseFIPS": false, + "Region": "af-south-1" } }, { @@ -21,9 +21,9 @@ } }, "params": { - "Region": "ap-east-1", "UseDualStack": false, - "UseFIPS": false + "UseFIPS": false, + "Region": "ap-east-1" } }, { @@ -34,9 +34,9 @@ } }, "params": { - "Region": "ap-northeast-1", "UseDualStack": false, - "UseFIPS": false + "UseFIPS": false, + "Region": "ap-northeast-1" } }, { @@ -47,9 +47,9 @@ } }, "params": { - "Region": "ap-northeast-2", "UseDualStack": false, - "UseFIPS": false + "UseFIPS": false, + "Region": "ap-northeast-2" } }, { @@ -60,9 +60,9 @@ } }, "params": { - "Region": "ap-northeast-3", "UseDualStack": false, - "UseFIPS": false + "UseFIPS": false, + "Region": "ap-northeast-3" } }, { @@ -73,9 +73,9 @@ } }, "params": { - "Region": "ap-south-1", "UseDualStack": false, - "UseFIPS": false + "UseFIPS": false, + "Region": "ap-south-1" } }, { @@ -86,9 +86,9 @@ } }, "params": { - "Region": "ap-southeast-1", "UseDualStack": false, - "UseFIPS": false + "UseFIPS": false, + "Region": "ap-southeast-1" } }, { @@ -99,9 +99,9 @@ } }, "params": { - "Region": "ap-southeast-2", "UseDualStack": false, - "UseFIPS": false + "UseFIPS": false, + "Region": "ap-southeast-2" } }, { @@ -112,9 +112,9 @@ } }, "params": { - "Region": "ca-central-1", "UseDualStack": false, - "UseFIPS": false + "UseFIPS": false, + "Region": "ca-central-1" } }, { @@ -125,9 +125,9 @@ } }, "params": { - "Region": "eu-central-1", "UseDualStack": false, - "UseFIPS": false + "UseFIPS": false, + "Region": "eu-central-1" } }, { @@ -138,9 +138,9 @@ } }, "params": { - "Region": "eu-north-1", "UseDualStack": false, - "UseFIPS": false + "UseFIPS": false, + "Region": "eu-north-1" } }, { @@ -151,9 +151,9 @@ } }, "params": { - "Region": "eu-south-1", "UseDualStack": false, - "UseFIPS": false + "UseFIPS": false, + "Region": "eu-south-1" } }, { @@ -164,9 +164,9 @@ } }, "params": { - "Region": "eu-west-1", "UseDualStack": false, - "UseFIPS": false + "UseFIPS": false, + "Region": "eu-west-1" } }, { @@ -177,9 +177,9 @@ } }, "params": { - "Region": "eu-west-2", "UseDualStack": false, - "UseFIPS": false + "UseFIPS": false, + "Region": "eu-west-2" } }, { @@ -190,9 +190,9 @@ } }, "params": { - "Region": "eu-west-3", "UseDualStack": false, - "UseFIPS": false + "UseFIPS": false, + "Region": "eu-west-3" } }, { @@ -203,9 +203,9 @@ } }, "params": { - "Region": "me-south-1", "UseDualStack": false, - "UseFIPS": false + "UseFIPS": false, + "Region": "me-south-1" } }, { @@ -216,9 +216,9 @@ } }, "params": { - "Region": "sa-east-1", "UseDualStack": false, - "UseFIPS": false + "UseFIPS": false, + "Region": "sa-east-1" } }, { @@ -229,9 +229,9 @@ } }, "params": { - "Region": "us-east-1", "UseDualStack": false, - "UseFIPS": false + "UseFIPS": false, + "Region": "us-east-1" } }, { @@ -242,9 +242,9 @@ } }, "params": { - "Region": "us-east-1", "UseDualStack": false, - "UseFIPS": true + "UseFIPS": true, + "Region": "us-east-1" } }, { @@ -255,9 +255,9 @@ } }, "params": { - "Region": "us-east-2", "UseDualStack": false, - "UseFIPS": false + "UseFIPS": false, + "Region": "us-east-2" } }, { @@ -268,9 +268,9 @@ } }, "params": { - "Region": "us-east-2", "UseDualStack": false, - "UseFIPS": true + "UseFIPS": true, + "Region": "us-east-2" } }, { @@ -281,9 +281,9 @@ } }, "params": { - "Region": "us-west-1", "UseDualStack": false, - "UseFIPS": false + "UseFIPS": false, + "Region": "us-west-1" } }, { @@ -294,9 +294,9 @@ } }, "params": { - "Region": "us-west-1", "UseDualStack": false, - "UseFIPS": true + "UseFIPS": true, + "Region": "us-west-1" } }, { @@ -307,9 +307,9 @@ } }, "params": { - "Region": "us-west-2", "UseDualStack": false, - "UseFIPS": false + "UseFIPS": false, + "Region": "us-west-2" } }, { @@ -320,9 +320,9 @@ } }, "params": { - "Region": "us-west-2", "UseDualStack": false, - "UseFIPS": true + "UseFIPS": true, + "Region": "us-west-2" } }, { @@ -333,9 +333,9 @@ } }, "params": { - "Region": "us-east-1", "UseDualStack": true, - "UseFIPS": true + "UseFIPS": true, + "Region": "us-east-1" } }, { @@ -346,9 +346,9 @@ } }, "params": { - "Region": "us-east-1", "UseDualStack": true, - "UseFIPS": false + "UseFIPS": false, + "Region": "us-east-1" } }, { @@ -359,9 +359,9 @@ } }, "params": { - "Region": "cn-north-1", "UseDualStack": true, - "UseFIPS": true + "UseFIPS": true, + "Region": "cn-north-1" } }, { @@ -372,9 +372,9 @@ } }, "params": { - "Region": "cn-north-1", "UseDualStack": false, - "UseFIPS": true + "UseFIPS": true, + "Region": "cn-north-1" } }, { @@ -385,9 +385,9 @@ } }, "params": { - "Region": "cn-north-1", "UseDualStack": true, - "UseFIPS": false + "UseFIPS": false, + "Region": "cn-north-1" } }, { @@ -398,9 +398,9 @@ } }, "params": { - "Region": "cn-north-1", "UseDualStack": false, - "UseFIPS": false + "UseFIPS": false, + "Region": "cn-north-1" } }, { @@ -411,9 +411,9 @@ } }, "params": { - "Region": "us-gov-east-1", "UseDualStack": true, - "UseFIPS": true + "UseFIPS": true, + "Region": "us-gov-east-1" } }, { @@ -424,9 +424,9 @@ } }, "params": { - "Region": "us-gov-east-1", "UseDualStack": false, - "UseFIPS": true + "UseFIPS": true, + "Region": "us-gov-east-1" } }, { @@ -437,9 +437,9 @@ } }, "params": { - "Region": "us-gov-east-1", "UseDualStack": true, - "UseFIPS": false + "UseFIPS": false, + "Region": "us-gov-east-1" } }, { @@ -450,9 +450,9 @@ } }, "params": { - "Region": "us-gov-east-1", "UseDualStack": false, - "UseFIPS": false + "UseFIPS": false, + "Region": "us-gov-east-1" } }, { @@ -463,9 +463,9 @@ } }, "params": { - "Region": "us-iso-east-1", "UseDualStack": false, - "UseFIPS": true + "UseFIPS": true, + "Region": "us-iso-east-1" } }, { @@ -476,9 +476,9 @@ } }, "params": { - "Region": "us-iso-east-1", "UseDualStack": false, - "UseFIPS": false + "UseFIPS": false, + "Region": "us-iso-east-1" } }, { @@ -489,9 +489,9 @@ } }, "params": { - "Region": "us-isob-east-1", "UseDualStack": false, - "UseFIPS": true + "UseFIPS": true, + "Region": "us-isob-east-1" } }, { @@ -502,9 +502,9 @@ } }, "params": { - "Region": "us-isob-east-1", "UseDualStack": false, - "UseFIPS": false + "UseFIPS": false, + "Region": "us-isob-east-1" } }, { @@ -515,9 +515,9 @@ } }, "params": { - "Region": "us-east-1", "UseDualStack": false, "UseFIPS": false, + "Region": "us-east-1", "Endpoint": "https://example.com" } }, @@ -540,9 +540,9 @@ "error": "Invalid Configuration: FIPS and custom endpoint are not supported" }, "params": { - "Region": "us-east-1", "UseDualStack": false, "UseFIPS": true, + "Region": "us-east-1", "Endpoint": "https://example.com" } }, @@ -552,9 +552,9 @@ "error": "Invalid Configuration: Dualstack and custom endpoint are not supported" }, "params": { - "Region": "us-east-1", "UseDualStack": true, "UseFIPS": false, + "Region": "us-east-1", "Endpoint": "https://example.com" } } diff --git a/models/apis/mediaconvert/2017-08-29/api-2.json b/models/apis/mediaconvert/2017-08-29/api-2.json index fe2b26a127a..11cf07cd6e5 100644 --- a/models/apis/mediaconvert/2017-08-29/api-2.json +++ b/models/apis/mediaconvert/2017-08-29/api-2.json @@ -2895,7 +2895,8 @@ "HDR10", "HLG_2020", "P3DCI", - "P3D65_SDR" + "P3D65_SDR", + "P3D65_HDR" ] }, "ColorSpaceConversion": { @@ -2907,7 +2908,8 @@ "FORCE_HDR10", "FORCE_HLG_2020", "FORCE_P3DCI", - "FORCE_P3D65_SDR" + "FORCE_P3D65_SDR", + "FORCE_P3D65_HDR" ] }, "ColorSpaceUsage": { diff --git a/models/apis/mediaconvert/2017-08-29/docs-2.json b/models/apis/mediaconvert/2017-08-29/docs-2.json index 3a7811372da..c7f78a7a6e3 100644 --- a/models/apis/mediaconvert/2017-08-29/docs-2.json +++ b/models/apis/mediaconvert/2017-08-29/docs-2.json @@ -449,9 +449,9 @@ } }, "BandwidthReductionFilter": { - "base": "The Bandwidth reduction filter increases the video quality of your output relative to its bitrate. Use to lower the bitrate of your constant quality QVBR output, with little or no perceptual decrease in quality. Or, use to increase the video quality of outputs with other rate control modes relative to the bitrate that you specify. Bandwidth reduction increases further when your input is low quality or noisy.Outputs that use this feature incur pro-tier pricing.When you include Bandwidth reduction filter, you cannot include the Noise reducer preprocessor.", + "base": "The Bandwidth reduction filter increases the video quality of your output relative to its bitrate. Use to lower the bitrate of your constant quality QVBR output, with little or no perceptual decrease in quality. Or, use to increase the video quality of outputs with other rate control modes relative to the bitrate that you specify. Bandwidth reduction increases further when your input is low quality or noisy. Outputs that use this feature incur pro-tier pricing. When you include Bandwidth reduction filter, you cannot include the Noise reducer preprocessor.", "refs": { - "H264Settings$BandwidthReductionFilter": "The Bandwidth reduction filter increases the video quality of your output relative to its bitrate. Use to lower the bitrate of your constant quality QVBR output, with little or no perceptual decrease in quality. Or, use to increase the video quality of outputs with other rate control modes relative to the bitrate that you specify. Bandwidth reduction increases further when your input is low quality or noisy.Outputs that use this feature incur pro-tier pricing.When you include Bandwidth reduction filter, you cannot include the Noise reducer preprocessor." + "H264Settings$BandwidthReductionFilter": "The Bandwidth reduction filter increases the video quality of your output relative to its bitrate. Use to lower the bitrate of your constant quality QVBR output, with little or no perceptual decrease in quality. Or, use to increase the video quality of outputs with other rate control modes relative to the bitrate that you specify. Bandwidth reduction increases further when your input is low quality or noisy. Outputs that use this feature incur pro-tier pricing. When you include Bandwidth reduction filter, you cannot include the Noise reducer preprocessor." } }, "BandwidthReductionFilterSharpening": { @@ -461,9 +461,9 @@ } }, "BandwidthReductionFilterStrength": { - "base": "Specify the strength of the Bandwidth reduction filter. For most workflows, we recommend that you choose Auto. Your output bandwidth will be reduced by at least 8 percent with no perceptual decrease in video quality. If your output bandwidth isn't constrained, set Filter strength to Low or Medium. Low results in minimal to no impact in perceptual quality. For more bandwidth reduction, choose High. The filter helps equalize quality between all scenes and increases video softness. We recommend that you choose High for low bitrate outputs.", + "base": "Specify the strength of the Bandwidth reduction filter. For most workflows, we recommend that you choose Auto to reduce the bandwidth of your output with little to no perceptual decrease in video quality. For high quality and high bitrate outputs, choose Low. For the most bandwidth reduction, choose High. We recommend that you choose High for low bitrate outputs. Note that High may incur a slight increase in the softness of your output.", "refs": { - "BandwidthReductionFilter$Strength": "Specify the strength of the Bandwidth reduction filter. For most workflows, we recommend that you choose Auto. Your output bandwidth will be reduced by at least 8 percent with no perceptual decrease in video quality. If your output bandwidth isn't constrained, set Filter strength to Low or Medium. Low results in minimal to no impact in perceptual quality. For more bandwidth reduction, choose High. The filter helps equalize quality between all scenes and increases video softness. We recommend that you choose High for low bitrate outputs." + "BandwidthReductionFilter$Strength": "Specify the strength of the Bandwidth reduction filter. For most workflows, we recommend that you choose Auto to reduce the bandwidth of your output with little to no perceptual decrease in video quality. For high quality and high bitrate outputs, choose Low. For the most bandwidth reduction, choose High. We recommend that you choose High for low bitrate outputs. Note that High may incur a slight increase in the softness of your output." } }, "BillingTagsSource": { @@ -828,15 +828,15 @@ } }, "ColorSpace": { - "base": "If your input video has accurate color space metadata, or if you don't know about color space, leave this set to the default value Follow. The service will automatically detect your input color space. If your input video has metadata indicating the wrong color space, specify the accurate color space here. If your input video is HDR 10 and the SMPTE ST 2086 Mastering Display Color Volume static metadata isn't present in your video stream, or if that metadata is present but not accurate, choose Force HDR 10 here and specify correct values in the input HDR 10 metadata settings. For more information about MediaConvert HDR jobs, see https://docs.aws.amazon.com/console/mediaconvert/hdr. Select P3D65 (SDR) to set the input color space metadata to the following:\n * Color primaries: Display P3\n * Transfer characteristics: SMPTE 428M\n * Matrix coefficients: BT.709", + "base": "If your input video has accurate color space metadata, or if you don't know about color space: Keep the default value, Follow. MediaConvert will automatically detect your input color space. If your input video has metadata indicating the wrong color space, or has missing metadata: Specify the accurate color space here. If your input video is HDR 10 and the SMPTE ST 2086 Mastering Display Color Volume static metadata isn't present in your video stream, or if that metadata is present but not accurate: Choose Force HDR 10. Specify correct values in the input HDR 10 metadata settings. For more information about HDR jobs, see https://docs.aws.amazon.com/console/mediaconvert/hdr. When you specify an input color space, MediaConvert uses the following color space metadata, which includes color primaries, transfer characteristics, and matrix coefficients:\n * HDR 10: BT.2020, PQ, BT.2020 non-constant\n * HLG 2020: BT.2020, HLG, BT.2020 non-constant\n * P3DCI (Theater): DCIP3, SMPTE 428M, BT.709\n * P3D65 (SDR): Display P3, sRGB, BT.709\n * P3D65 (HDR): Display P3, PQ, BT.709", "refs": { - "VideoSelector$ColorSpace": "If your input video has accurate color space metadata, or if you don't know about color space, leave this set to the default value Follow. The service will automatically detect your input color space. If your input video has metadata indicating the wrong color space, specify the accurate color space here. If your input video is HDR 10 and the SMPTE ST 2086 Mastering Display Color Volume static metadata isn't present in your video stream, or if that metadata is present but not accurate, choose Force HDR 10 here and specify correct values in the input HDR 10 metadata settings. For more information about MediaConvert HDR jobs, see https://docs.aws.amazon.com/console/mediaconvert/hdr. Select P3D65 (SDR) to set the input color space metadata to the following:\n * Color primaries: Display P3\n * Transfer characteristics: SMPTE 428M\n * Matrix coefficients: BT.709" + "VideoSelector$ColorSpace": "If your input video has accurate color space metadata, or if you don't know about color space: Keep the default value, Follow. MediaConvert will automatically detect your input color space. If your input video has metadata indicating the wrong color space, or has missing metadata: Specify the accurate color space here. If your input video is HDR 10 and the SMPTE ST 2086 Mastering Display Color Volume static metadata isn't present in your video stream, or if that metadata is present but not accurate: Choose Force HDR 10. Specify correct values in the input HDR 10 metadata settings. For more information about HDR jobs, see https://docs.aws.amazon.com/console/mediaconvert/hdr. When you specify an input color space, MediaConvert uses the following color space metadata, which includes color primaries, transfer characteristics, and matrix coefficients:\n * HDR 10: BT.2020, PQ, BT.2020 non-constant\n * HLG 2020: BT.2020, HLG, BT.2020 non-constant\n * P3DCI (Theater): DCIP3, SMPTE 428M, BT.709\n * P3D65 (SDR): Display P3, sRGB, BT.709\n * P3D65 (HDR): Display P3, PQ, BT.709" } }, "ColorSpaceConversion": { - "base": "Specify the color space you want for this output. The service supports conversion between HDR formats, between SDR formats, from SDR to HDR, and from HDR to SDR. SDR to HDR conversion doesn't upgrade the dynamic range. The converted video has an HDR format, but visually appears the same as an unconverted output. HDR to SDR conversion uses Elemental tone mapping technology to approximate the outcome of manually regrading from HDR to SDR. Select Force P3D65 (SDR) to set the output color space metadata to the following: * Color primaries: Display P3 * Transfer characteristics: SMPTE 428M * Matrix coefficients: BT.709", + "base": "Specify the color space you want for this output. The service supports conversion between HDR formats, between SDR formats, from SDR to HDR, and from HDR to SDR. SDR to HDR conversion doesn't upgrade the dynamic range. The converted video has an HDR format, but visually appears the same as an unconverted output. HDR to SDR conversion uses tone mapping to approximate the outcome of manually regrading from HDR to SDR. When you specify an output color space, MediaConvert uses the following color space metadata, which includes color primaries, transfer characteristics, and matrix coefficients:\n * HDR 10: BT.2020, PQ, BT.2020 non-constant\n * HLG 2020: BT.2020, HLG, BT.2020 non-constant\n * P3DCI (Theater): DCIP3, SMPTE 428M, BT.709\n * P3D65 (SDR): Display P3, sRGB, BT.709\n * P3D65 (HDR): Display P3, PQ, BT.709", "refs": { - "ColorCorrector$ColorSpaceConversion": "Specify the color space you want for this output. The service supports conversion between HDR formats, between SDR formats, from SDR to HDR, and from HDR to SDR. SDR to HDR conversion doesn't upgrade the dynamic range. The converted video has an HDR format, but visually appears the same as an unconverted output. HDR to SDR conversion uses Elemental tone mapping technology to approximate the outcome of manually regrading from HDR to SDR. Select Force P3D65 (SDR) to set the output color space metadata to the following: * Color primaries: Display P3 * Transfer characteristics: SMPTE 428M * Matrix coefficients: BT.709" + "ColorCorrector$ColorSpaceConversion": "Specify the color space you want for this output. The service supports conversion between HDR formats, between SDR formats, from SDR to HDR, and from HDR to SDR. SDR to HDR conversion doesn't upgrade the dynamic range. The converted video has an HDR format, but visually appears the same as an unconverted output. HDR to SDR conversion uses tone mapping to approximate the outcome of manually regrading from HDR to SDR. When you specify an output color space, MediaConvert uses the following color space metadata, which includes color primaries, transfer characteristics, and matrix coefficients:\n * HDR 10: BT.2020, PQ, BT.2020 non-constant\n * HLG 2020: BT.2020, HLG, BT.2020 non-constant\n * P3DCI (Theater): DCIP3, SMPTE 428M, BT.709\n * P3D65 (SDR): Display P3, sRGB, BT.709\n * P3D65 (HDR): Display P3, PQ, BT.709" } }, "ColorSpaceUsage": { diff --git a/models/apis/mediaconvert/2017-08-29/endpoint-tests-1.json b/models/apis/mediaconvert/2017-08-29/endpoint-tests-1.json index 630f3dce768..cec29153d73 100644 --- a/models/apis/mediaconvert/2017-08-29/endpoint-tests-1.json +++ b/models/apis/mediaconvert/2017-08-29/endpoint-tests-1.json @@ -8,9 +8,9 @@ } }, "params": { - "Region": "ap-northeast-1", + "UseDualStack": false, "UseFIPS": false, - "UseDualStack": false + "Region": "ap-northeast-1" } }, { @@ -21,9 +21,9 @@ } }, "params": { - "Region": "ap-northeast-2", + "UseDualStack": false, "UseFIPS": false, - "UseDualStack": false + "Region": "ap-northeast-2" } }, { @@ -34,9 +34,9 @@ } }, "params": { - "Region": "ap-south-1", + "UseDualStack": false, "UseFIPS": false, - "UseDualStack": false + "Region": "ap-south-1" } }, { @@ -47,9 +47,9 @@ } }, "params": { - "Region": "ap-southeast-1", + "UseDualStack": false, "UseFIPS": false, - "UseDualStack": false + "Region": "ap-southeast-1" } }, { @@ -60,9 +60,9 @@ } }, "params": { - "Region": "ap-southeast-2", + "UseDualStack": false, "UseFIPS": false, - "UseDualStack": false + "Region": "ap-southeast-2" } }, { @@ -73,9 +73,9 @@ } }, "params": { - "Region": "ca-central-1", + "UseDualStack": false, "UseFIPS": false, - "UseDualStack": false + "Region": "ca-central-1" } }, { @@ -86,9 +86,9 @@ } }, "params": { - "Region": "ca-central-1", + "UseDualStack": false, "UseFIPS": true, - "UseDualStack": false + "Region": "ca-central-1" } }, { @@ -99,9 +99,9 @@ } }, "params": { - "Region": "eu-central-1", + "UseDualStack": false, "UseFIPS": false, - "UseDualStack": false + "Region": "eu-central-1" } }, { @@ -112,9 +112,9 @@ } }, "params": { - "Region": "eu-north-1", + "UseDualStack": false, "UseFIPS": false, - "UseDualStack": false + "Region": "eu-north-1" } }, { @@ -125,9 +125,9 @@ } }, "params": { - "Region": "eu-west-1", + "UseDualStack": false, "UseFIPS": false, - "UseDualStack": false + "Region": "eu-west-1" } }, { @@ -138,9 +138,9 @@ } }, "params": { - "Region": "eu-west-2", + "UseDualStack": false, "UseFIPS": false, - "UseDualStack": false + "Region": "eu-west-2" } }, { @@ -151,9 +151,9 @@ } }, "params": { - "Region": "eu-west-3", + "UseDualStack": false, "UseFIPS": false, - "UseDualStack": false + "Region": "eu-west-3" } }, { @@ -164,9 +164,9 @@ } }, "params": { - "Region": "sa-east-1", + "UseDualStack": false, "UseFIPS": false, - "UseDualStack": false + "Region": "sa-east-1" } }, { @@ -177,9 +177,9 @@ } }, "params": { - "Region": "us-east-1", + "UseDualStack": false, "UseFIPS": false, - "UseDualStack": false + "Region": "us-east-1" } }, { @@ -190,9 +190,9 @@ } }, "params": { - "Region": "us-east-1", + "UseDualStack": false, "UseFIPS": true, - "UseDualStack": false + "Region": "us-east-1" } }, { @@ -203,9 +203,9 @@ } }, "params": { - "Region": "us-east-2", + "UseDualStack": false, "UseFIPS": false, - "UseDualStack": false + "Region": "us-east-2" } }, { @@ -216,9 +216,9 @@ } }, "params": { - "Region": "us-east-2", + "UseDualStack": false, "UseFIPS": true, - "UseDualStack": false + "Region": "us-east-2" } }, { @@ -229,9 +229,9 @@ } }, "params": { - "Region": "us-west-1", + "UseDualStack": false, "UseFIPS": false, - "UseDualStack": false + "Region": "us-west-1" } }, { @@ -242,9 +242,9 @@ } }, "params": { - "Region": "us-west-1", + "UseDualStack": false, "UseFIPS": true, - "UseDualStack": false + "Region": "us-west-1" } }, { @@ -255,9 +255,9 @@ } }, "params": { - "Region": "us-west-2", + "UseDualStack": false, "UseFIPS": false, - "UseDualStack": false + "Region": "us-west-2" } }, { @@ -268,9 +268,9 @@ } }, "params": { - "Region": "us-west-2", + "UseDualStack": false, "UseFIPS": true, - "UseDualStack": false + "Region": "us-west-2" } }, { @@ -281,9 +281,9 @@ } }, "params": { - "Region": "us-east-1", + "UseDualStack": true, "UseFIPS": true, - "UseDualStack": true + "Region": "us-east-1" } }, { @@ -294,9 +294,9 @@ } }, "params": { - "Region": "us-east-1", + "UseDualStack": true, "UseFIPS": false, - "UseDualStack": true + "Region": "us-east-1" } }, { @@ -307,9 +307,9 @@ } }, "params": { - "Region": "cn-northwest-1", + "UseDualStack": false, "UseFIPS": false, - "UseDualStack": false + "Region": "cn-northwest-1" } }, { @@ -320,9 +320,9 @@ } }, "params": { - "Region": "cn-north-1", + "UseDualStack": true, "UseFIPS": true, - "UseDualStack": true + "Region": "cn-north-1" } }, { @@ -333,9 +333,9 @@ } }, "params": { - "Region": "cn-north-1", + "UseDualStack": false, "UseFIPS": true, - "UseDualStack": false + "Region": "cn-north-1" } }, { @@ -346,9 +346,9 @@ } }, "params": { - "Region": "cn-north-1", + "UseDualStack": true, "UseFIPS": false, - "UseDualStack": true + "Region": "cn-north-1" } }, { @@ -359,9 +359,9 @@ } }, "params": { - "Region": "cn-north-1", + "UseDualStack": false, "UseFIPS": false, - "UseDualStack": false + "Region": "cn-north-1" } }, { @@ -372,9 +372,9 @@ } }, "params": { - "Region": "us-gov-west-1", + "UseDualStack": false, "UseFIPS": false, - "UseDualStack": false + "Region": "us-gov-west-1" } }, { @@ -385,9 +385,9 @@ } }, "params": { - "Region": "us-gov-west-1", + "UseDualStack": false, "UseFIPS": true, - "UseDualStack": false + "Region": "us-gov-west-1" } }, { @@ -398,9 +398,9 @@ } }, "params": { - "Region": "us-gov-east-1", + "UseDualStack": true, "UseFIPS": true, - "UseDualStack": true + "Region": "us-gov-east-1" } }, { @@ -411,9 +411,9 @@ } }, "params": { - "Region": "us-gov-east-1", + "UseDualStack": false, "UseFIPS": true, - "UseDualStack": false + "Region": "us-gov-east-1" } }, { @@ -424,9 +424,9 @@ } }, "params": { - "Region": "us-gov-east-1", + "UseDualStack": true, "UseFIPS": false, - "UseDualStack": true + "Region": "us-gov-east-1" } }, { @@ -437,9 +437,9 @@ } }, "params": { - "Region": "us-gov-east-1", + "UseDualStack": false, "UseFIPS": false, - "UseDualStack": false + "Region": "us-gov-east-1" } }, { @@ -450,9 +450,9 @@ } }, "params": { - "Region": "us-iso-east-1", + "UseDualStack": false, "UseFIPS": true, - "UseDualStack": false + "Region": "us-iso-east-1" } }, { @@ -463,9 +463,9 @@ } }, "params": { - "Region": "us-iso-east-1", + "UseDualStack": false, "UseFIPS": false, - "UseDualStack": false + "Region": "us-iso-east-1" } }, { @@ -476,9 +476,9 @@ } }, "params": { - "Region": "us-isob-east-1", + "UseDualStack": false, "UseFIPS": true, - "UseDualStack": false + "Region": "us-isob-east-1" } }, { @@ -489,9 +489,9 @@ } }, "params": { - "Region": "us-isob-east-1", + "UseDualStack": false, "UseFIPS": false, - "UseDualStack": false + "Region": "us-isob-east-1" } }, { @@ -502,9 +502,9 @@ } }, "params": { - "Region": "us-east-1", - "UseFIPS": false, "UseDualStack": false, + "UseFIPS": false, + "Region": "us-east-1", "Endpoint": "https://example.com" } }, @@ -516,8 +516,8 @@ } }, "params": { - "UseFIPS": false, "UseDualStack": false, + "UseFIPS": false, "Endpoint": "https://example.com" } }, @@ -527,9 +527,9 @@ "error": "Invalid Configuration: FIPS and custom endpoint are not supported" }, "params": { - "Region": "us-east-1", - "UseFIPS": true, "UseDualStack": false, + "UseFIPS": true, + "Region": "us-east-1", "Endpoint": "https://example.com" } }, @@ -539,9 +539,9 @@ "error": "Invalid Configuration: Dualstack and custom endpoint are not supported" }, "params": { - "Region": "us-east-1", - "UseFIPS": false, "UseDualStack": true, + "UseFIPS": false, + "Region": "us-east-1", "Endpoint": "https://example.com" } } diff --git a/models/apis/medialive/2017-10-14/api-2.json b/models/apis/medialive/2017-10-14/api-2.json index 615a9fe2a01..acb486159d5 100644 --- a/models/apis/medialive/2017-10-14/api-2.json +++ b/models/apis/medialive/2017-10-14/api-2.json @@ -10483,6 +10483,10 @@ "Sid": { "shape": "__doubleMin1Max65535", "locationName": "sid" + }, + "Timezone": { + "shape": "NielsenWatermarkTimezones", + "locationName": "timezone" } }, "required": [ @@ -10497,6 +10501,21 @@ "ENABLED" ] }, + "NielsenWatermarkTimezones": { + "type": "string", + "enum": [ + "AMERICA_PUERTO_RICO", + "US_ALASKA", + "US_ARIZONA", + "US_CENTRAL", + "US_EASTERN", + "US_HAWAII", + "US_MOUNTAIN", + "US_PACIFIC", + "US_SAMOA", + "UTC" + ] + }, "NielsenWatermarksCbetStepaside": { "type": "string", "enum": [ diff --git a/models/apis/medialive/2017-10-14/docs-2.json b/models/apis/medialive/2017-10-14/docs-2.json index c46883d5456..277ffe381f8 100644 --- a/models/apis/medialive/2017-10-14/docs-2.json +++ b/models/apis/medialive/2017-10-14/docs-2.json @@ -391,7 +391,7 @@ "AvailSettings": { "base": "Avail Settings", "refs": { - "AvailConfiguration$AvailSettings": "Ad avail settings." + "AvailConfiguration$AvailSettings": "Controls how SCTE-35 messages create cues. Splice Insert mode treats all segmentation signals traditionally. With Time Signal APOS mode only Time Signal Placement Opportunity and Break messages create segment breaks. With ESAM mode, signals are forwarded to an ESAM server for possible update." } }, "BadGatewayException": { @@ -565,7 +565,7 @@ } }, "CaptionSelector": { - "base": "Output groups for this Live Event. Output groups contain information about where streams should be distributed.", + "base": "Caption Selector", "refs": { "__listOfCaptionSelector$member": null } @@ -2600,6 +2600,12 @@ "NielsenConfiguration$NielsenPcmToId3Tagging": "Enables Nielsen PCM to ID3 tagging" } }, + "NielsenWatermarkTimezones": { + "base": "Nielsen Watermark Timezones", + "refs": { + "NielsenNaesIiNw$Timezone": "Choose the timezone for the time stamps in the watermark. If not provided,\nthe timestamps will be in Coordinated Universal Time (UTC)" + } + }, "NielsenWatermarksCbetStepaside": { "base": "Nielsen Watermarks Cbet Stepaside", "refs": { @@ -3055,7 +3061,7 @@ } }, "Scte35SpliceInsert": { - "base": "Scte35 Splice Insert", + "base": "Typical configuration that applies breaks on splice inserts in addition to time signal placement opportunities, breaks, and advertisements.", "refs": { "AvailSettings$Scte35SpliceInsert": null } @@ -3079,7 +3085,7 @@ } }, "Scte35TimeSignalApos": { - "base": "Scte35 Time Signal Apos", + "base": "Atypical configuration that applies segment breaks only on SCTE-35 time signal placement opportunities and breaks.", "refs": { "AvailSettings$Scte35TimeSignalApos": null } @@ -3656,9 +3662,9 @@ "H264Settings$GopClosedCadence": "Frequency of closed GOPs. In streaming applications, it is recommended that this be set to 1 so a decoder joining mid-stream will receive an IDR frame as quickly as possible. Setting this value to 0 will break output segmenting.", "H265Settings$GopClosedCadence": "Frequency of closed GOPs. In streaming applications, it is recommended that this be set to 1 so a decoder joining mid-stream will receive an IDR frame as quickly as possible. Setting this value to 0 will break output segmenting.", "HlsAkamaiSettings$ConnectionRetryInterval": "Number of seconds to wait before retrying connection to the CDN if the connection is lost.", - "HlsAkamaiSettings$NumRetries": "Number of retry attempts that will be made before the Live Event is put into an error state.", + "HlsAkamaiSettings$NumRetries": "Number of retry attempts that will be made before the Live Event is put into an error state. Applies only if the CDN destination URI begins with \"s3\" or \"mediastore\". For other URIs, the value is always 3.", "HlsBasicPutSettings$ConnectionRetryInterval": "Number of seconds to wait before retrying connection to the CDN if the connection is lost.", - "HlsBasicPutSettings$NumRetries": "Number of retry attempts that will be made before the Live Event is put into an error state.", + "HlsBasicPutSettings$NumRetries": "Number of retry attempts that will be made before the Live Event is put into an error state. Applies only if the CDN destination URI begins with \"s3\" or \"mediastore\". For other URIs, the value is always 3.", "HlsGroupSettings$MinSegmentLength": "Minimum length of MPEG-2 Transport Stream segments in seconds. When set, minimum segment length is enforced by looking ahead and back within the specified range for a nearby avail and extending the segment size if needed.", "HlsGroupSettings$TimedMetadataId3Period": "Timed Metadata interval in seconds.", "HlsGroupSettings$TimestampDeltaMilliseconds": "Provides an extra millisecond delta offset to fine tune the timestamps.", @@ -3667,9 +3673,9 @@ "HlsInputSettings$Retries": "The number of consecutive times that attempts to read a manifest or segment must fail before the input is considered unavailable.", "HlsInputSettings$RetryInterval": "The number of seconds between retries when an attempt to read a manifest or segment fails.", "HlsMediaStoreSettings$ConnectionRetryInterval": "Number of seconds to wait before retrying connection to the CDN if the connection is lost.", - "HlsMediaStoreSettings$NumRetries": "Number of retry attempts that will be made before the Live Event is put into an error state.", + "HlsMediaStoreSettings$NumRetries": "Number of retry attempts that will be made before the Live Event is put into an error state. Applies only if the CDN destination URI begins with \"s3\" or \"mediastore\". For other URIs, the value is always 3.", "HlsWebdavSettings$ConnectionRetryInterval": "Number of seconds to wait before retrying connection to the CDN if the connection is lost.", - "HlsWebdavSettings$NumRetries": "Number of retry attempts that will be made before the Live Event is put into an error state.", + "HlsWebdavSettings$NumRetries": "Number of retry attempts that will be made before the Live Event is put into an error state. Applies only if the CDN destination URI begins with \"s3\" or \"mediastore\". For other URIs, the value is always 3.", "M2tsSettings$AudioFramesPerPes": "The number of audio frames to insert for each PES packet.", "M2tsSettings$Bitrate": "The output bitrate of the transport stream in bits per second. Setting to 0 lets the muxer automatically determine the appropriate bitrate.", "M3u8Settings$AudioFramesPerPes": "The number of audio frames to insert for each PES packet.", diff --git a/models/apis/medialive/2017-10-14/endpoint-rule-set-1.json b/models/apis/medialive/2017-10-14/endpoint-rule-set-1.json index 377ab242500..a354a5ca273 100644 --- a/models/apis/medialive/2017-10-14/endpoint-rule-set-1.json +++ b/models/apis/medialive/2017-10-14/endpoint-rule-set-1.json @@ -3,7 +3,7 @@ "parameters": { "Region": { "builtIn": "AWS::Region", - "required": true, + "required": false, "documentation": "The AWS region used to dispatch the request.", "type": "String" }, @@ -32,13 +32,12 @@ { "conditions": [ { - "fn": "aws.partition", + "fn": "isSet", "argv": [ { - "ref": "Region" + "ref": "Endpoint" } - ], - "assign": "PartitionResult" + ] } ], "type": "tree", @@ -46,14 +45,20 @@ { "conditions": [ { - "fn": "isSet", + "fn": "booleanEquals", "argv": [ { - "ref": "Endpoint" - } + "ref": "UseFIPS" + }, + true ] } ], + "error": "Invalid Configuration: FIPS and custom endpoint are not supported", + "type": "error" + }, + { + "conditions": [], "type": "tree", "rules": [ { @@ -62,67 +67,42 @@ "fn": "booleanEquals", "argv": [ { - "ref": "UseFIPS" + "ref": "UseDualStack" }, true ] } ], - "error": "Invalid Configuration: FIPS and custom endpoint are not supported", + "error": "Invalid Configuration: Dualstack and custom endpoint are not supported", "type": "error" }, { "conditions": [], - "type": "tree", - "rules": [ - { - "conditions": [ - { - "fn": "booleanEquals", - "argv": [ - { - "ref": "UseDualStack" - }, - true - ] - } - ], - "error": "Invalid Configuration: Dualstack and custom endpoint are not supported", - "type": "error" + "endpoint": { + "url": { + "ref": "Endpoint" }, - { - "conditions": [], - "endpoint": { - "url": { - "ref": "Endpoint" - }, - "properties": {}, - "headers": {} - }, - "type": "endpoint" - } - ] + "properties": {}, + "headers": {} + }, + "type": "endpoint" } ] - }, + } + ] + }, + { + "conditions": [], + "type": "tree", + "rules": [ { "conditions": [ { - "fn": "booleanEquals", - "argv": [ - { - "ref": "UseFIPS" - }, - true - ] - }, - { - "fn": "booleanEquals", + "fn": "isSet", "argv": [ { - "ref": "UseDualStack" - }, - true + "ref": "Region" + } ] } ], @@ -131,90 +111,215 @@ { "conditions": [ { - "fn": "booleanEquals", + "fn": "aws.partition", "argv": [ - true, { - "fn": "getAttr", + "ref": "Region" + } + ], + "assign": "PartitionResult" + } + ], + "type": "tree", + "rules": [ + { + "conditions": [ + { + "fn": "booleanEquals", + "argv": [ + { + "ref": "UseFIPS" + }, + true + ] + }, + { + "fn": "booleanEquals", "argv": [ { - "ref": "PartitionResult" + "ref": "UseDualStack" + }, + true + ] + } + ], + "type": "tree", + "rules": [ + { + "conditions": [ + { + "fn": "booleanEquals", + "argv": [ + true, + { + "fn": "getAttr", + "argv": [ + { + "ref": "PartitionResult" + }, + "supportsFIPS" + ] + } + ] }, - "supportsFIPS" + { + "fn": "booleanEquals", + "argv": [ + true, + { + "fn": "getAttr", + "argv": [ + { + "ref": "PartitionResult" + }, + "supportsDualStack" + ] + } + ] + } + ], + "type": "tree", + "rules": [ + { + "conditions": [], + "type": "tree", + "rules": [ + { + "conditions": [], + "endpoint": { + "url": "https://medialive-fips.{Region}.{PartitionResult#dualStackDnsSuffix}", + "properties": {}, + "headers": {} + }, + "type": "endpoint" + } + ] + } ] + }, + { + "conditions": [], + "error": "FIPS and DualStack are enabled, but this partition does not support one or both", + "type": "error" } ] }, { - "fn": "booleanEquals", - "argv": [ - true, + "conditions": [ { - "fn": "getAttr", + "fn": "booleanEquals", "argv": [ { - "ref": "PartitionResult" + "ref": "UseFIPS" }, - "supportsDualStack" + true ] } + ], + "type": "tree", + "rules": [ + { + "conditions": [ + { + "fn": "booleanEquals", + "argv": [ + true, + { + "fn": "getAttr", + "argv": [ + { + "ref": "PartitionResult" + }, + "supportsFIPS" + ] + } + ] + } + ], + "type": "tree", + "rules": [ + { + "conditions": [], + "type": "tree", + "rules": [ + { + "conditions": [], + "endpoint": { + "url": "https://medialive-fips.{Region}.{PartitionResult#dnsSuffix}", + "properties": {}, + "headers": {} + }, + "type": "endpoint" + } + ] + } + ] + }, + { + "conditions": [], + "error": "FIPS is enabled but this partition does not support FIPS", + "type": "error" + } ] - } - ], - "type": "tree", - "rules": [ - { - "conditions": [], - "endpoint": { - "url": "https://medialive-fips.{Region}.{PartitionResult#dualStackDnsSuffix}", - "properties": {}, - "headers": {} - }, - "type": "endpoint" - } - ] - }, - { - "conditions": [], - "error": "FIPS and DualStack are enabled, but this partition does not support one or both", - "type": "error" - } - ] - }, - { - "conditions": [ - { - "fn": "booleanEquals", - "argv": [ - { - "ref": "UseFIPS" }, - true - ] - } - ], - "type": "tree", - "rules": [ - { - "conditions": [ { - "fn": "booleanEquals", - "argv": [ - true, + "conditions": [ { - "fn": "getAttr", + "fn": "booleanEquals", "argv": [ { - "ref": "PartitionResult" + "ref": "UseDualStack" }, - "supportsFIPS" + true + ] + } + ], + "type": "tree", + "rules": [ + { + "conditions": [ + { + "fn": "booleanEquals", + "argv": [ + true, + { + "fn": "getAttr", + "argv": [ + { + "ref": "PartitionResult" + }, + "supportsDualStack" + ] + } + ] + } + ], + "type": "tree", + "rules": [ + { + "conditions": [], + "type": "tree", + "rules": [ + { + "conditions": [], + "endpoint": { + "url": "https://medialive.{Region}.{PartitionResult#dualStackDnsSuffix}", + "properties": {}, + "headers": {} + }, + "type": "endpoint" + } + ] + } ] + }, + { + "conditions": [], + "error": "DualStack is enabled but this partition does not support DualStack", + "type": "error" } ] - } - ], - "type": "tree", - "rules": [ + }, { "conditions": [], "type": "tree", @@ -222,7 +327,7 @@ { "conditions": [], "endpoint": { - "url": "https://medialive-fips.{Region}.{PartitionResult#dnsSuffix}", + "url": "https://medialive.{Region}.{PartitionResult#dnsSuffix}", "properties": {}, "headers": {} }, @@ -231,74 +336,13 @@ ] } ] - }, - { - "conditions": [], - "error": "FIPS is enabled but this partition does not support FIPS", - "type": "error" - } - ] - }, - { - "conditions": [ - { - "fn": "booleanEquals", - "argv": [ - { - "ref": "UseDualStack" - }, - true - ] - } - ], - "type": "tree", - "rules": [ - { - "conditions": [ - { - "fn": "booleanEquals", - "argv": [ - true, - { - "fn": "getAttr", - "argv": [ - { - "ref": "PartitionResult" - }, - "supportsDualStack" - ] - } - ] - } - ], - "type": "tree", - "rules": [ - { - "conditions": [], - "endpoint": { - "url": "https://medialive.{Region}.{PartitionResult#dualStackDnsSuffix}", - "properties": {}, - "headers": {} - }, - "type": "endpoint" - } - ] - }, - { - "conditions": [], - "error": "DualStack is enabled but this partition does not support DualStack", - "type": "error" } ] }, { "conditions": [], - "endpoint": { - "url": "https://medialive.{Region}.{PartitionResult#dnsSuffix}", - "properties": {}, - "headers": {} - }, - "type": "endpoint" + "error": "Invalid Configuration: Missing Region", + "type": "error" } ] } diff --git a/models/apis/medialive/2017-10-14/endpoint-tests-1.json b/models/apis/medialive/2017-10-14/endpoint-tests-1.json index c0a96034e16..510effaf9d7 100644 --- a/models/apis/medialive/2017-10-14/endpoint-tests-1.json +++ b/models/apis/medialive/2017-10-14/endpoint-tests-1.json @@ -1,86 +1,393 @@ { "testCases": [ { - "documentation": "For region us-iso-east-1 with FIPS enabled and DualStack enabled", + "documentation": "For region ap-northeast-1 with FIPS disabled and DualStack disabled", "expect": { - "error": "FIPS and DualStack are enabled, but this partition does not support one or both" + "endpoint": { + "url": "https://medialive.ap-northeast-1.amazonaws.com" + } + }, + "params": { + "UseDualStack": false, + "UseFIPS": false, + "Region": "ap-northeast-1" + } + }, + { + "documentation": "For region ap-northeast-2 with FIPS disabled and DualStack disabled", + "expect": { + "endpoint": { + "url": "https://medialive.ap-northeast-2.amazonaws.com" + } + }, + "params": { + "UseDualStack": false, + "UseFIPS": false, + "Region": "ap-northeast-2" + } + }, + { + "documentation": "For region ap-south-1 with FIPS disabled and DualStack disabled", + "expect": { + "endpoint": { + "url": "https://medialive.ap-south-1.amazonaws.com" + } + }, + "params": { + "UseDualStack": false, + "UseFIPS": false, + "Region": "ap-south-1" + } + }, + { + "documentation": "For region ap-southeast-1 with FIPS disabled and DualStack disabled", + "expect": { + "endpoint": { + "url": "https://medialive.ap-southeast-1.amazonaws.com" + } + }, + "params": { + "UseDualStack": false, + "UseFIPS": false, + "Region": "ap-southeast-1" + } + }, + { + "documentation": "For region ap-southeast-2 with FIPS disabled and DualStack disabled", + "expect": { + "endpoint": { + "url": "https://medialive.ap-southeast-2.amazonaws.com" + } + }, + "params": { + "UseDualStack": false, + "UseFIPS": false, + "Region": "ap-southeast-2" + } + }, + { + "documentation": "For region eu-central-1 with FIPS disabled and DualStack disabled", + "expect": { + "endpoint": { + "url": "https://medialive.eu-central-1.amazonaws.com" + } + }, + "params": { + "UseDualStack": false, + "UseFIPS": false, + "Region": "eu-central-1" + } + }, + { + "documentation": "For region eu-north-1 with FIPS disabled and DualStack disabled", + "expect": { + "endpoint": { + "url": "https://medialive.eu-north-1.amazonaws.com" + } + }, + "params": { + "UseDualStack": false, + "UseFIPS": false, + "Region": "eu-north-1" + } + }, + { + "documentation": "For region eu-west-1 with FIPS disabled and DualStack disabled", + "expect": { + "endpoint": { + "url": "https://medialive.eu-west-1.amazonaws.com" + } + }, + "params": { + "UseDualStack": false, + "UseFIPS": false, + "Region": "eu-west-1" + } + }, + { + "documentation": "For region eu-west-2 with FIPS disabled and DualStack disabled", + "expect": { + "endpoint": { + "url": "https://medialive.eu-west-2.amazonaws.com" + } + }, + "params": { + "UseDualStack": false, + "UseFIPS": false, + "Region": "eu-west-2" + } + }, + { + "documentation": "For region eu-west-3 with FIPS disabled and DualStack disabled", + "expect": { + "endpoint": { + "url": "https://medialive.eu-west-3.amazonaws.com" + } + }, + "params": { + "UseDualStack": false, + "UseFIPS": false, + "Region": "eu-west-3" + } + }, + { + "documentation": "For region sa-east-1 with FIPS disabled and DualStack disabled", + "expect": { + "endpoint": { + "url": "https://medialive.sa-east-1.amazonaws.com" + } + }, + "params": { + "UseDualStack": false, + "UseFIPS": false, + "Region": "sa-east-1" + } + }, + { + "documentation": "For region us-east-1 with FIPS disabled and DualStack disabled", + "expect": { + "endpoint": { + "url": "https://medialive.us-east-1.amazonaws.com" + } + }, + "params": { + "UseDualStack": false, + "UseFIPS": false, + "Region": "us-east-1" + } + }, + { + "documentation": "For region us-east-1 with FIPS enabled and DualStack disabled", + "expect": { + "endpoint": { + "url": "https://medialive-fips.us-east-1.amazonaws.com" + } + }, + "params": { + "UseDualStack": false, + "UseFIPS": true, + "Region": "us-east-1" + } + }, + { + "documentation": "For region us-east-2 with FIPS disabled and DualStack disabled", + "expect": { + "endpoint": { + "url": "https://medialive.us-east-2.amazonaws.com" + } + }, + "params": { + "UseDualStack": false, + "UseFIPS": false, + "Region": "us-east-2" + } + }, + { + "documentation": "For region us-east-2 with FIPS enabled and DualStack disabled", + "expect": { + "endpoint": { + "url": "https://medialive-fips.us-east-2.amazonaws.com" + } + }, + "params": { + "UseDualStack": false, + "UseFIPS": true, + "Region": "us-east-2" + } + }, + { + "documentation": "For region us-west-2 with FIPS disabled and DualStack disabled", + "expect": { + "endpoint": { + "url": "https://medialive.us-west-2.amazonaws.com" + } + }, + "params": { + "UseDualStack": false, + "UseFIPS": false, + "Region": "us-west-2" + } + }, + { + "documentation": "For region us-west-2 with FIPS enabled and DualStack disabled", + "expect": { + "endpoint": { + "url": "https://medialive-fips.us-west-2.amazonaws.com" + } + }, + "params": { + "UseDualStack": false, + "UseFIPS": true, + "Region": "us-west-2" + } + }, + { + "documentation": "For region us-east-1 with FIPS enabled and DualStack enabled", + "expect": { + "endpoint": { + "url": "https://medialive-fips.us-east-1.api.aws" + } }, "params": { "UseDualStack": true, - "Region": "us-iso-east-1", - "UseFIPS": true + "UseFIPS": true, + "Region": "us-east-1" } }, { - "documentation": "For region us-iso-east-1 with FIPS enabled and DualStack disabled", + "documentation": "For region us-east-1 with FIPS disabled and DualStack enabled", "expect": { "endpoint": { - "url": "https://medialive-fips.us-iso-east-1.c2s.ic.gov" + "url": "https://medialive.us-east-1.api.aws" + } + }, + "params": { + "UseDualStack": true, + "UseFIPS": false, + "Region": "us-east-1" + } + }, + { + "documentation": "For region cn-north-1 with FIPS enabled and DualStack enabled", + "expect": { + "endpoint": { + "url": "https://medialive-fips.cn-north-1.api.amazonwebservices.com.cn" + } + }, + "params": { + "UseDualStack": true, + "UseFIPS": true, + "Region": "cn-north-1" + } + }, + { + "documentation": "For region cn-north-1 with FIPS enabled and DualStack disabled", + "expect": { + "endpoint": { + "url": "https://medialive-fips.cn-north-1.amazonaws.com.cn" } }, "params": { "UseDualStack": false, - "Region": "us-iso-east-1", - "UseFIPS": true + "UseFIPS": true, + "Region": "cn-north-1" } }, { - "documentation": "For region us-iso-east-1 with FIPS disabled and DualStack enabled", + "documentation": "For region cn-north-1 with FIPS disabled and DualStack enabled", "expect": { - "error": "DualStack is enabled but this partition does not support DualStack" + "endpoint": { + "url": "https://medialive.cn-north-1.api.amazonwebservices.com.cn" + } }, "params": { "UseDualStack": true, - "Region": "us-iso-east-1", - "UseFIPS": false + "UseFIPS": false, + "Region": "cn-north-1" } }, { - "documentation": "For region us-iso-east-1 with FIPS disabled and DualStack disabled", + "documentation": "For region cn-north-1 with FIPS disabled and DualStack disabled", "expect": { "endpoint": { - "url": "https://medialive.us-iso-east-1.c2s.ic.gov" + "url": "https://medialive.cn-north-1.amazonaws.com.cn" } }, "params": { "UseDualStack": false, - "Region": "us-iso-east-1", - "UseFIPS": false + "UseFIPS": false, + "Region": "cn-north-1" } }, { - "documentation": "For region us-isob-east-1 with FIPS enabled and DualStack enabled", + "documentation": "For region us-gov-east-1 with FIPS enabled and DualStack enabled", "expect": { - "error": "FIPS and DualStack are enabled, but this partition does not support one or both" + "endpoint": { + "url": "https://medialive-fips.us-gov-east-1.api.aws" + } }, "params": { "UseDualStack": true, - "Region": "us-isob-east-1", - "UseFIPS": true + "UseFIPS": true, + "Region": "us-gov-east-1" } }, { - "documentation": "For region us-isob-east-1 with FIPS enabled and DualStack disabled", + "documentation": "For region us-gov-east-1 with FIPS enabled and DualStack disabled", "expect": { "endpoint": { - "url": "https://medialive-fips.us-isob-east-1.sc2s.sgov.gov" + "url": "https://medialive-fips.us-gov-east-1.amazonaws.com" } }, "params": { "UseDualStack": false, - "Region": "us-isob-east-1", - "UseFIPS": true + "UseFIPS": true, + "Region": "us-gov-east-1" } }, { - "documentation": "For region us-isob-east-1 with FIPS disabled and DualStack enabled", + "documentation": "For region us-gov-east-1 with FIPS disabled and DualStack enabled", "expect": { - "error": "DualStack is enabled but this partition does not support DualStack" + "endpoint": { + "url": "https://medialive.us-gov-east-1.api.aws" + } }, "params": { "UseDualStack": true, - "Region": "us-isob-east-1", - "UseFIPS": false + "UseFIPS": false, + "Region": "us-gov-east-1" + } + }, + { + "documentation": "For region us-gov-east-1 with FIPS disabled and DualStack disabled", + "expect": { + "endpoint": { + "url": "https://medialive.us-gov-east-1.amazonaws.com" + } + }, + "params": { + "UseDualStack": false, + "UseFIPS": false, + "Region": "us-gov-east-1" + } + }, + { + "documentation": "For region us-iso-east-1 with FIPS disabled and DualStack disabled", + "expect": { + "endpoint": { + "url": "https://medialive.us-iso-east-1.c2s.ic.gov" + } + }, + "params": { + "UseDualStack": false, + "UseFIPS": false, + "Region": "us-iso-east-1" + } + }, + { + "documentation": "For region us-iso-east-1 with FIPS enabled and DualStack disabled", + "expect": { + "endpoint": { + "url": "https://medialive-fips.us-iso-east-1.c2s.ic.gov" + } + }, + "params": { + "UseDualStack": false, + "UseFIPS": true, + "Region": "us-iso-east-1" + } + }, + { + "documentation": "For region us-isob-east-1 with FIPS enabled and DualStack disabled", + "expect": { + "endpoint": { + "url": "https://medialive-fips.us-isob-east-1.sc2s.sgov.gov" + } + }, + "params": { + "UseDualStack": false, + "UseFIPS": true, + "Region": "us-isob-east-1" } }, { @@ -92,12 +399,12 @@ }, "params": { "UseDualStack": false, - "Region": "us-isob-east-1", - "UseFIPS": false + "UseFIPS": false, + "Region": "us-isob-east-1" } }, { - "documentation": "For custom endpoint with fips disabled and dualstack disabled", + "documentation": "For custom endpoint with region set and fips disabled and dualstack disabled", "expect": { "endpoint": { "url": "https://example.com" @@ -105,7 +412,20 @@ }, "params": { "UseDualStack": false, + "UseFIPS": false, "Region": "us-east-1", + "Endpoint": "https://example.com" + } + }, + { + "documentation": "For custom endpoint with region not set and fips disabled and dualstack disabled", + "expect": { + "endpoint": { + "url": "https://example.com" + } + }, + "params": { + "UseDualStack": false, "UseFIPS": false, "Endpoint": "https://example.com" } @@ -117,8 +437,8 @@ }, "params": { "UseDualStack": false, - "Region": "us-east-1", "UseFIPS": true, + "Region": "us-east-1", "Endpoint": "https://example.com" } }, @@ -129,8 +449,8 @@ }, "params": { "UseDualStack": true, - "Region": "us-east-1", "UseFIPS": false, + "Region": "us-east-1", "Endpoint": "https://example.com" } } diff --git a/models/apis/transcribe/2017-10-26/api-2.json b/models/apis/transcribe/2017-10-26/api-2.json index 84997fa59e3..dfa077dd358 100644 --- a/models/apis/transcribe/2017-10-26/api-2.json +++ b/models/apis/transcribe/2017-10-26/api-2.json @@ -822,7 +822,8 @@ "LanguageCode":{"shape":"LanguageCode"}, "Words":{"shape":"Words"}, "VocabularyFilterFileUri":{"shape":"Uri"}, - "Tags":{"shape":"TagList"} + "Tags":{"shape":"TagList"}, + "DataAccessRoleArn":{"shape":"DataAccessRoleArn"} } }, "CreateVocabularyFilterResponse":{ @@ -844,7 +845,8 @@ "LanguageCode":{"shape":"LanguageCode"}, "Phrases":{"shape":"Phrases"}, "VocabularyFileUri":{"shape":"Uri"}, - "Tags":{"shape":"TagList"} + "Tags":{"shape":"TagList"}, + "DataAccessRoleArn":{"shape":"DataAccessRoleArn"} } }, "CreateVocabularyResponse":{ @@ -1997,7 +1999,8 @@ "members":{ "VocabularyFilterName":{"shape":"VocabularyFilterName"}, "Words":{"shape":"Words"}, - "VocabularyFilterFileUri":{"shape":"Uri"} + "VocabularyFilterFileUri":{"shape":"Uri"}, + "DataAccessRoleArn":{"shape":"DataAccessRoleArn"} } }, "UpdateVocabularyFilterResponse":{ @@ -2018,7 +2021,8 @@ "VocabularyName":{"shape":"VocabularyName"}, "LanguageCode":{"shape":"LanguageCode"}, "Phrases":{"shape":"Phrases"}, - "VocabularyFileUri":{"shape":"Uri"} + "VocabularyFileUri":{"shape":"Uri"}, + "DataAccessRoleArn":{"shape":"DataAccessRoleArn"} } }, "UpdateVocabularyResponse":{ diff --git a/models/apis/transcribe/2017-10-26/docs-2.json b/models/apis/transcribe/2017-10-26/docs-2.json index e7b7e760c50..0a40cf2d096 100644 --- a/models/apis/transcribe/2017-10-26/docs-2.json +++ b/models/apis/transcribe/2017-10-26/docs-2.json @@ -2,9 +2,9 @@ "version": "2.0", "service": "Amazon Transcribe offers three main types of batch transcription: Standard, Medical, and Call Analytics.
Standard transcriptions are the most common option. Refer to for details.
Medical transcriptions are tailored to medical professionals and incorporate medical terms. A common use case for this service is transcribing doctor-patient dialogue into after-visit notes. Refer to for details.
Call Analytics transcriptions are designed for use with call center audio on two different channels; if you're looking for insight into customer service calls, use this option. Refer to for details.
Creates a new Call Analytics category.
All categories are automatically applied to your Call Analytics transcriptions. Note that in order to apply categories to your transcriptions, you must create them before submitting your transcription request, as categories cannot be applied retroactively.
When creating a new category, you can use the InputType
parameter to label the category as a batch category (POST_CALL
) or a streaming category (REAL_TIME
). Batch categories can only be applied to batch transcriptions and streaming categories can only be applied to streaming transcriptions. If you do not include InputType
, your category is created as a batch category by default.
Call Analytics categories are composed of rules. For each category, you must create between 1 and 20 rules. Rules can include these parameters: , , , and .
To update an existing category, see .
To learn more about Call Analytics categories, see Creating categories for batch transcriptions and Creating categories for streaming transcriptions.
", + "CreateCallAnalyticsCategory": "Creates a new Call Analytics category.
All categories are automatically applied to your Call Analytics transcriptions. Note that in order to apply categories to your transcriptions, you must create them before submitting your transcription request, as categories cannot be applied retroactively.
When creating a new category, you can use the InputType
parameter to label the category as a POST_CALL
or a REAL_TIME
category. POST_CALL
categories can only be applied to post-call transcriptions and REAL_TIME
categories can only be applied to real-time transcriptions. If you do not include InputType
, your category is created as a POST_CALL
category by default.
Call Analytics categories are composed of rules. For each category, you must create between 1 and 20 rules. Rules can include these parameters: , , , and .
To update an existing category, see .
To learn more about Call Analytics categories, see Creating categories for post-call transcriptions and Creating categories for real-time transcriptions.
", "CreateLanguageModel": "Creates a new custom language model.
When creating a new custom language model, you must specify:
If you want a Wideband (audio sample rates over 16,000 Hz) or Narrowband (audio sample rates under 16,000 Hz) base model
The location of your training and tuning files (this must be an Amazon S3 URI)
The language of your model
A unique name for your model
Creates a new custom medical vocabulary.
Before creating a new custom medical vocabulary, you must first upload a text file that contains your new entries, phrases, and terms into an Amazon S3 bucket. Note that this differs from , where you can include a list of terms within your request using the Phrases
flag; CreateMedicalVocabulary
does not support the Phrases
flag.
Each language has a character set that contains all allowed characters for that specific language. If you use unsupported characters, your custom vocabulary request fails. Refer to Character Sets for Custom Vocabularies to get the character set for your language.
For more information, see Custom vocabularies.
", + "CreateMedicalVocabulary": "Creates a new custom medical vocabulary.
Before creating a new custom medical vocabulary, you must first upload a text file that contains your vocabulary table into an Amazon S3 bucket. Note that this differs from , where you can include a list of terms within your request using the Phrases
flag; CreateMedicalVocabulary
does not support the Phrases
flag and only accepts vocabularies in table format.
Each language has a character set that contains all allowed characters for that specific language. If you use unsupported characters, your custom vocabulary request fails. Refer to Character Sets for Custom Vocabularies to get the character set for your language.
For more information, see Custom vocabularies.
", "CreateVocabulary": "Creates a new custom vocabulary.
When creating a new custom vocabulary, you can either upload a text file that contains your new entries, phrases, and terms into an Amazon S3 bucket and include the URI in your request. Or you can include a list of terms directly in your request using the Phrases
flag.
Each language has a character set that contains all allowed characters for that specific language. If you use unsupported characters, your custom vocabulary request fails. Refer to Character Sets for Custom Vocabularies to get the character set for your language.
For more information, see Custom vocabularies.
", "CreateVocabularyFilter": "Creates a new custom vocabulary filter.
You can use custom vocabulary filters to mask, delete, or flag specific words from your transcript. Custom vocabulary filters are commonly used to mask profanity in transcripts.
Each language has a character set that contains all allowed characters for that specific language. If you use unsupported characters, your custom vocabulary filter request fails. Refer to Character Sets for Custom Vocabularies to get the character set for your language.
For more information, see Vocabulary filtering.
", "DeleteCallAnalyticsCategory": "Deletes a Call Analytics category. To use this operation, specify the name of the category you want to delete using CategoryName
. Category names are case sensitive.
Provides a list of transcription jobs that match the specified criteria. If no criteria are specified, all transcription jobs are returned.
To get detailed information about a specific transcription job, use the operation.
", "ListVocabularies": "Provides a list of custom vocabularies that match the specified criteria. If no criteria are specified, all custom vocabularies are returned.
To get detailed information about a specific custom vocabulary, use the operation.
", "ListVocabularyFilters": "Provides a list of custom vocabulary filters that match the specified criteria. If no criteria are specified, all custom vocabularies are returned.
To get detailed information about a specific custom vocabulary filter, use the operation.
", - "StartCallAnalyticsJob": "Transcribes the audio from a customer service call and applies any additional Request Parameters you choose to include in your request.
In addition to many standard transcription features, Call Analytics provides you with call characteristics, call summarization, speaker sentiment, and optional redaction of your text transcript and your audio file. You can also apply custom categories to flag specified conditions. To learn more about these features and insights, refer to Analyzing call center audio with Call Analytics.
If you want to apply categories to your Call Analytics job, you must create them before submitting your job request. Categories cannot be retroactively applied to a job. To create a new category, use the operation. To learn more about Call Analytics categories, see Creating categories for batch transcriptions and Creating categories for streaming transcriptions.
To make a StartCallAnalyticsJob
request, you must first upload your media file into an Amazon S3 bucket; you can then specify the Amazon S3 location of the file using the Media
parameter.
You must include the following parameters in your StartCallAnalyticsJob
request:
region
: The Amazon Web Services Region where you are making your request. For a list of Amazon Web Services Regions supported with Amazon Transcribe, refer to Amazon Transcribe endpoints and quotas.
CallAnalyticsJobName
: A custom name that you create for your transcription job that's unique within your Amazon Web Services account.
DataAccessRoleArn
: The Amazon Resource Name (ARN) of an IAM role that has permissions to access the Amazon S3 bucket that contains your input files.
Media
(MediaFileUri
or RedactedMediaFileUri
): The Amazon S3 location of your media file.
With Call Analytics, you can redact the audio contained in your media file by including RedactedMediaFileUri
, instead of MediaFileUri
, to specify the location of your input audio. If you choose to redact your audio, you can find your redacted media at the location specified in the RedactedMediaFileUri
field of your response.
Transcribes the audio from a customer service call and applies any additional Request Parameters you choose to include in your request.
In addition to many standard transcription features, Call Analytics provides you with call characteristics, call summarization, speaker sentiment, and optional redaction of your text transcript and your audio file. You can also apply custom categories to flag specified conditions. To learn more about these features and insights, refer to Analyzing call center audio with Call Analytics.
If you want to apply categories to your Call Analytics job, you must create them before submitting your job request. Categories cannot be retroactively applied to a job. To create a new category, use the operation. To learn more about Call Analytics categories, see Creating categories for post-call transcriptions and Creating categories for real-time transcriptions.
To make a StartCallAnalyticsJob
request, you must first upload your media file into an Amazon S3 bucket; you can then specify the Amazon S3 location of the file using the Media
parameter.
Note that job queuing is enabled by default for Call Analytics jobs.
You must include the following parameters in your StartCallAnalyticsJob
request:
region
: The Amazon Web Services Region where you are making your request. For a list of Amazon Web Services Regions supported with Amazon Transcribe, refer to Amazon Transcribe endpoints and quotas.
CallAnalyticsJobName
: A custom name that you create for your transcription job that's unique within your Amazon Web Services account.
DataAccessRoleArn
: The Amazon Resource Name (ARN) of an IAM role that has permissions to access the Amazon S3 bucket that contains your input files.
Media
(MediaFileUri
or RedactedMediaFileUri
): The Amazon S3 location of your media file.
With Call Analytics, you can redact the audio contained in your media file by including RedactedMediaFileUri
, instead of MediaFileUri
, to specify the location of your input audio. If you choose to redact your audio, you can find your redacted media at the location specified in the RedactedMediaFileUri
field of your response.
Transcribes the audio from a medical dictation or conversation and applies any additional Request Parameters you choose to include in your request.
In addition to many standard transcription features, Amazon Transcribe Medical provides you with a robust medical vocabulary and, optionally, content identification, which adds flags to personal health information (PHI). To learn more about these features, refer to How Amazon Transcribe Medical works.
To make a StartMedicalTranscriptionJob
request, you must first upload your media file into an Amazon S3 bucket; you can then specify the S3 location of the file using the Media
parameter.
You must include the following parameters in your StartMedicalTranscriptionJob
request:
region
: The Amazon Web Services Region where you are making your request. For a list of Amazon Web Services Regions supported with Amazon Transcribe, refer to Amazon Transcribe endpoints and quotas.
MedicalTranscriptionJobName
: A custom name you create for your transcription job that is unique within your Amazon Web Services account.
Media
(MediaFileUri
): The Amazon S3 location of your media file.
LanguageCode
: This must be en-US
.
OutputBucketName
: The Amazon S3 bucket where you want your transcript stored. If you want your output stored in a sub-folder of this bucket, you must also include OutputKey
.
Specialty
: This must be PRIMARYCARE
.
Type
: Choose whether your audio is a conversation or a dictation.
Transcribes the audio from a media file and applies any additional Request Parameters you choose to include in your request.
To make a StartTranscriptionJob
request, you must first upload your media file into an Amazon S3 bucket; you can then specify the Amazon S3 location of the file using the Media
parameter.
You must include the following parameters in your StartTranscriptionJob
request:
region
: The Amazon Web Services Region where you are making your request. For a list of Amazon Web Services Regions supported with Amazon Transcribe, refer to Amazon Transcribe endpoints and quotas.
TranscriptionJobName
: A custom name you create for your transcription job that is unique within your Amazon Web Services account.
Media
(MediaFileUri
): The Amazon S3 location of your media file.
One of LanguageCode
, IdentifyLanguage
, or IdentifyMultipleLanguages
: If you know the language of your media file, specify it using the LanguageCode
parameter; you can find all valid language codes in the Supported languages table. If you don't know the languages spoken in your media, use either IdentifyLanguage
or IdentifyMultipleLanguages
and let Amazon Transcribe identify the languages for you.
Adds one or more custom tags, each in the form of a key:value pair, to the specified resource.
To learn more about using tags with Amazon Transcribe, refer to Tagging resources.
", @@ -69,7 +69,7 @@ "base": null, "refs": { "InterruptionFilter$Negate": "Set to TRUE
to flag speech that does not contain interruptions. Set to FALSE
to flag speech that contains interruptions.
Makes it possible to enable job queuing when your concurrent request limit is exceeded. When AllowDeferredExecution
is set to true
, transcription job requests are placed in a queue until the number of jobs falls below the concurrent request limit. If AllowDeferredExecution
is set to false
and the number of transcription job requests exceed the concurrent request limit, you get a LimitExceededException
error.
Note that job queuing is enabled by default for Call Analytics jobs.
If you include AllowDeferredExecution
in your request, you must also include DataAccessRoleArn
.
Makes it possible to enable job queuing when your concurrent request limit is exceeded. When AllowDeferredExecution
is set to true
, transcription job requests are placed in a queue until the number of jobs falls below the concurrent request limit. If AllowDeferredExecution
is set to false
and the number of transcription job requests exceed the concurrent request limit, you get a LimitExceededException
error.
If you include AllowDeferredExecution
in your request, you must also include DataAccessRoleArn
.
Shows if a more current base model is available for use with the specified custom language model.
If false
, your custom language model is using the most up-to-date base model.
If true
, there is a newer base model available than the one your language model is using.
Note that to update a base model, you must recreate the custom language model using the new base model. Base model upgrades for existing custom language models are not supported.
", "MedicalTranscriptionSetting$ShowSpeakerLabels": "Enables speaker partitioning (diarization) in your transcription output. Speaker partitioning labels the speech from individual speakers in your media file.
If you enable ShowSpeakerLabels
in your request, you must also include MaxSpeakerLabels
.
You can't include ShowSpeakerLabels
and ChannelIdentification
in the same request. Including both parameters returns a BadRequestException
.
For more information, see Partitioning speakers (diarization).
", "MedicalTranscriptionSetting$ChannelIdentification": "Enables channel identification in multi-channel audio.
Channel identification transcribes the audio on each channel independently, then appends the output for each channel into one transcript.
If you have multi-channel audio and do not enable channel identification, your audio is transcribed in a continuous manner and your transcript does not separate the speech by channel.
You can't include both ShowSpeakerLabels
and ChannelIdentification
in the same request. Including both parameters returns a BadRequestException
.
For more information, see Transcribing multi-channel audio.
", @@ -254,9 +254,13 @@ "base": null, "refs": { "CallAnalyticsJob$DataAccessRoleArn": "The Amazon Resource Name (ARN) you included in your request.
", + "CreateVocabularyFilterRequest$DataAccessRoleArn": "The Amazon Resource Name (ARN) of an IAM role that has permissions to access the Amazon S3 bucket that contains your input files (in this case, your custom vocabulary filter). If the role that you specify doesn’t have the appropriate permissions to access the specified Amazon S3 location, your request fails.
IAM role ARNs have the format arn:partition:iam::account:role/role-name-with-path
. For example: arn:aws:iam::111122223333:role/Admin
.
For more information, see IAM ARNs.
", + "CreateVocabularyRequest$DataAccessRoleArn": "The Amazon Resource Name (ARN) of an IAM role that has permissions to access the Amazon S3 bucket that contains your input files (in this case, your custom vocabulary). If the role that you specify doesn’t have the appropriate permissions to access the specified Amazon S3 location, your request fails.
IAM role ARNs have the format arn:partition:iam::account:role/role-name-with-path
. For example: arn:aws:iam::111122223333:role/Admin
.
For more information, see IAM ARNs.
", "InputDataConfig$DataAccessRoleArn": "The Amazon Resource Name (ARN) of an IAM role that has permissions to access the Amazon S3 bucket that contains your input files. If the role that you specify doesn’t have the appropriate permissions to access the specified Amazon S3 location, your request fails.
IAM role ARNs have the format arn:partition:iam::account:role/role-name-with-path
. For example: arn:aws:iam::111122223333:role/Admin
.
For more information, see IAM ARNs.
", "JobExecutionSettings$DataAccessRoleArn": "The Amazon Resource Name (ARN) of an IAM role that has permissions to access the Amazon S3 bucket that contains your input files. If the role that you specify doesn’t have the appropriate permissions to access the specified Amazon S3 location, your request fails.
IAM role ARNs have the format arn:partition:iam::account:role/role-name-with-path
. For example: arn:aws:iam::111122223333:role/Admin
. For more information, see IAM ARNs.
Note that if you include DataAccessRoleArn
in your request, you must also include AllowDeferredExecution
.
The Amazon Resource Name (ARN) of an IAM role that has permissions to access the Amazon S3 bucket that contains your input files. If the role that you specify doesn’t have the appropriate permissions to access the specified Amazon S3 location, your request fails.
IAM role ARNs have the format arn:partition:iam::account:role/role-name-with-path
. For example: arn:aws:iam::111122223333:role/Admin
.
For more information, see IAM ARNs.
" + "StartCallAnalyticsJobRequest$DataAccessRoleArn": "The Amazon Resource Name (ARN) of an IAM role that has permissions to access the Amazon S3 bucket that contains your input files. If the role that you specify doesn’t have the appropriate permissions to access the specified Amazon S3 location, your request fails.
IAM role ARNs have the format arn:partition:iam::account:role/role-name-with-path
. For example: arn:aws:iam::111122223333:role/Admin
.
For more information, see IAM ARNs.
", + "UpdateVocabularyFilterRequest$DataAccessRoleArn": "The Amazon Resource Name (ARN) of an IAM role that has permissions to access the Amazon S3 bucket that contains your input files (in this case, your custom vocabulary filter). If the role that you specify doesn’t have the appropriate permissions to access the specified Amazon S3 location, your request fails.
IAM role ARNs have the format arn:partition:iam::account:role/role-name-with-path
. For example: arn:aws:iam::111122223333:role/Admin
.
For more information, see IAM ARNs.
", + "UpdateVocabularyRequest$DataAccessRoleArn": "The Amazon Resource Name (ARN) of an IAM role that has permissions to access the Amazon S3 bucket that contains your input files (in this case, your custom vocabulary). If the role that you specify doesn’t have the appropriate permissions to access the specified Amazon S3 location, your request fails.
IAM role ARNs have the format arn:partition:iam::account:role/role-name-with-path
. For example: arn:aws:iam::111122223333:role/Admin
.
For more information, see IAM ARNs.
" } }, "DateTime": { @@ -470,8 +474,8 @@ "base": null, "refs": { "CategoryProperties$InputType": "The input type associated with the specified category. POST_CALL
refers to a category that is applied to batch transcriptions; REAL_TIME
refers to a category that is applied to streaming transcriptions.
Choose whether you want to create a streaming or a batch category for your Call Analytics transcription.
Specifying POST_CALL
assigns your category to batch transcriptions; categories with this input type cannot be applied to streaming (real-time) transcriptions.
Specifying REAL_TIME
assigns your category to streaming transcriptions; categories with this input type cannot be applied to batch (post-call) transcriptions.
If you do not include InputType
, your category is created as a batch category by default.
Choose whether you want to update a streaming or a batch Call Analytics category. The input type you specify must match the input type specified when the category was created. For example, if you created a category with the POST_CALL
input type, you must use POST_CALL
as the input type when updating this category.
Choose whether you want to create a real-time or a post-call category for your Call Analytics transcription.
Specifying POST_CALL
assigns your category to post-call transcriptions; categories with this input type cannot be applied to streaming (real-time) transcriptions.
Specifying REAL_TIME
assigns your category to streaming transcriptions; categories with this input type cannot be applied to post-call transcriptions.
If you do not include InputType
, your category is created as a post-call category by default.
Choose whether you want to update a real-time or a post-call category. The input type you specify must match the input type specified when the category was created. For example, if you created a category with the POST_CALL
input type, you must use POST_CALL
as the input type when updating this category.
Flag the presence or absence of interruptions in your Call Analytics transcription output.
Rules using InterruptionFilter
are designed to match:
Instances where an agent interrupts a customer
Instances where a customer interrupts an agent
Either participant interrupting the other
A lack of interruptions
See Rule criteria for batch categories for usage examples.
", + "base": "Flag the presence or absence of interruptions in your Call Analytics transcription output.
Rules using InterruptionFilter
are designed to match:
Instances where an agent interrupts a customer
Instances where a customer interrupts an agent
Either participant interrupting the other
A lack of interruptions
See Rule criteria for post-call categories for usage examples.
", "refs": { "Rule$InterruptionFilter": "Flag the presence or absence of interruptions in your Call Analytics transcription output. Refer to for more detail.
" } @@ -502,9 +506,9 @@ "KMSKeyId": { "base": null, "refs": { - "StartCallAnalyticsJobRequest$OutputEncryptionKMSKeyId": "The KMS key you want to use to encrypt your Call Analytics output.
If using a key located in the current Amazon Web Services account, you can specify your KMS key in one of four ways:
Use the KMS key ID itself. For example, 1234abcd-12ab-34cd-56ef-1234567890ab
.
Use an alias for the KMS key ID. For example, alias/ExampleAlias
.
Use the Amazon Resource Name (ARN) for the KMS key ID. For example, arn:aws:kms:region:account-ID:key/1234abcd-12ab-34cd-56ef-1234567890ab
.
Use the ARN for the KMS key alias. For example, arn:aws:kms:region:account-ID:alias/ExampleAlias
.
If using a key located in a different Amazon Web Services account than the current Amazon Web Services account, you can specify your KMS key in one of two ways:
Use the ARN for the KMS key ID. For example, arn:aws:kms:region:account-ID:key/1234abcd-12ab-34cd-56ef-1234567890ab
.
Use the ARN for the KMS key alias. For example, arn:aws:kms:region:account-ID:alias/ExampleAlias
.
If you don't specify an encryption key, your output is encrypted with the default Amazon S3 key (SSE-S3).
If you specify a KMS key to encrypt your output, you must also specify an output location using the OutputLocation
parameter.
Note that the user making the request must have permission to use the specified KMS key.
", - "StartMedicalTranscriptionJobRequest$OutputEncryptionKMSKeyId": "The KMS key you want to use to encrypt your medical transcription output.
If using a key located in the current Amazon Web Services account, you can specify your KMS key in one of four ways:
Use the KMS key ID itself. For example, 1234abcd-12ab-34cd-56ef-1234567890ab
.
Use an alias for the KMS key ID. For example, alias/ExampleAlias
.
Use the Amazon Resource Name (ARN) for the KMS key ID. For example, arn:aws:kms:region:account-ID:key/1234abcd-12ab-34cd-56ef-1234567890ab
.
Use the ARN for the KMS key alias. For example, arn:aws:kms:region:account-ID:alias/ExampleAlias
.
If using a key located in a different Amazon Web Services account than the current Amazon Web Services account, you can specify your KMS key in one of two ways:
Use the ARN for the KMS key ID. For example, arn:aws:kms:region:account-ID:key/1234abcd-12ab-34cd-56ef-1234567890ab
.
Use the ARN for the KMS key alias. For example, arn:aws:kms:region:account-ID:alias/ExampleAlias
.
If you don't specify an encryption key, your output is encrypted with the default Amazon S3 key (SSE-S3).
If you specify a KMS key to encrypt your output, you must also specify an output location using the OutputLocation
parameter.
Note that the user making the request must have permission to use the specified KMS key.
", - "StartTranscriptionJobRequest$OutputEncryptionKMSKeyId": "The KMS key you want to use to encrypt your transcription output.
If using a key located in the current Amazon Web Services account, you can specify your KMS key in one of four ways:
Use the KMS key ID itself. For example, 1234abcd-12ab-34cd-56ef-1234567890ab
.
Use an alias for the KMS key ID. For example, alias/ExampleAlias
.
Use the Amazon Resource Name (ARN) for the KMS key ID. For example, arn:aws:kms:region:account-ID:key/1234abcd-12ab-34cd-56ef-1234567890ab
.
Use the ARN for the KMS key alias. For example, arn:aws:kms:region:account-ID:alias/ExampleAlias
.
If using a key located in a different Amazon Web Services account than the current Amazon Web Services account, you can specify your KMS key in one of two ways:
Use the ARN for the KMS key ID. For example, arn:aws:kms:region:account-ID:key/1234abcd-12ab-34cd-56ef-1234567890ab
.
Use the ARN for the KMS key alias. For example, arn:aws:kms:region:account-ID:alias/ExampleAlias
.
If you don't specify an encryption key, your output is encrypted with the default Amazon S3 key (SSE-S3).
If you specify a KMS key to encrypt your output, you must also specify an output location using the OutputLocation
parameter.
Note that the user making the request must have permission to use the specified KMS key.
" + "StartCallAnalyticsJobRequest$OutputEncryptionKMSKeyId": "The KMS key you want to use to encrypt your Call Analytics output.
If using a key located in the current Amazon Web Services account, you can specify your KMS key in one of four ways:
Use the KMS key ID itself. For example, 1234abcd-12ab-34cd-56ef-1234567890ab
.
Use an alias for the KMS key ID. For example, alias/ExampleAlias
.
Use the Amazon Resource Name (ARN) for the KMS key ID. For example, arn:aws:kms:region:account-ID:key/1234abcd-12ab-34cd-56ef-1234567890ab
.
Use the ARN for the KMS key alias. For example, arn:aws:kms:region:account-ID:alias/ExampleAlias
.
If using a key located in a different Amazon Web Services account than the current Amazon Web Services account, you can specify your KMS key in one of two ways:
Use the ARN for the KMS key ID. For example, arn:aws:kms:region:account-ID:key/1234abcd-12ab-34cd-56ef-1234567890ab
.
Use the ARN for the KMS key alias. For example, arn:aws:kms:region:account-ID:alias/ExampleAlias
.
If you don't specify an encryption key, your output is encrypted with the default Amazon S3 key (SSE-S3).
If you specify a KMS key to encrypt your output, you must also specify an output location using the OutputLocation
parameter.
Note that the role making the request must have permission to use the specified KMS key.
", + "StartMedicalTranscriptionJobRequest$OutputEncryptionKMSKeyId": "The KMS key you want to use to encrypt your medical transcription output.
If using a key located in the current Amazon Web Services account, you can specify your KMS key in one of four ways:
Use the KMS key ID itself. For example, 1234abcd-12ab-34cd-56ef-1234567890ab
.
Use an alias for the KMS key ID. For example, alias/ExampleAlias
.
Use the Amazon Resource Name (ARN) for the KMS key ID. For example, arn:aws:kms:region:account-ID:key/1234abcd-12ab-34cd-56ef-1234567890ab
.
Use the ARN for the KMS key alias. For example, arn:aws:kms:region:account-ID:alias/ExampleAlias
.
If using a key located in a different Amazon Web Services account than the current Amazon Web Services account, you can specify your KMS key in one of two ways:
Use the ARN for the KMS key ID. For example, arn:aws:kms:region:account-ID:key/1234abcd-12ab-34cd-56ef-1234567890ab
.
Use the ARN for the KMS key alias. For example, arn:aws:kms:region:account-ID:alias/ExampleAlias
.
If you don't specify an encryption key, your output is encrypted with the default Amazon S3 key (SSE-S3).
If you specify a KMS key to encrypt your output, you must also specify an output location using the OutputLocation
parameter.
Note that the role making the request must have permission to use the specified KMS key.
", + "StartTranscriptionJobRequest$OutputEncryptionKMSKeyId": "The KMS key you want to use to encrypt your transcription output.
If using a key located in the current Amazon Web Services account, you can specify your KMS key in one of four ways:
Use the KMS key ID itself. For example, 1234abcd-12ab-34cd-56ef-1234567890ab
.
Use an alias for the KMS key ID. For example, alias/ExampleAlias
.
Use the Amazon Resource Name (ARN) for the KMS key ID. For example, arn:aws:kms:region:account-ID:key/1234abcd-12ab-34cd-56ef-1234567890ab
.
Use the ARN for the KMS key alias. For example, arn:aws:kms:region:account-ID:alias/ExampleAlias
.
If using a key located in a different Amazon Web Services account than the current Amazon Web Services account, you can specify your KMS key in one of two ways:
Use the ARN for the KMS key ID. For example, arn:aws:kms:region:account-ID:key/1234abcd-12ab-34cd-56ef-1234567890ab
.
Use the ARN for the KMS key alias. For example, arn:aws:kms:region:account-ID:alias/ExampleAlias
.
If you don't specify an encryption key, your output is encrypted with the default Amazon S3 key (SSE-S3).
If you specify a KMS key to encrypt your output, you must also specify an output location using the OutputLocation
parameter.
Note that the role making the request must have permission to use the specified KMS key.
" } }, "LanguageCode": { @@ -567,7 +571,7 @@ } }, "LanguageModel": { - "base": "Provides information about a custom language model, including the base model name, when the model was created, the location of the files used to train the model, when the model was last modified, the name you chose for the model, its language, its processing state, and if there is an upgrade available for the base model.
", + "base": "Provides information about a custom language model, including:
The base model name
When the model was created
The location of the files used to train the model
When the model was last modified
The name you chose for the model
The model's language
The model's processing state
Any available upgrades for the base model
Provides information about the specified custom language model.
This parameter also shows if the base language model you used to create your custom language model has been updated. If Amazon Transcribe has updated the base model, you can create a new custom language model using the updated base model.
If you tried to create a new custom language model and the request wasn't successful, you can use this DescribeLanguageModel
to help identify the reason for this failure.
Flag the presence or absence of periods of silence in your Call Analytics transcription output.
Rules using NonTalkTimeFilter
are designed to match:
The presence of silence at specified periods throughout the call
The presence of speech at specified periods throughout the call
See Rule criteria for batch categories for usage examples.
", + "base": "Flag the presence or absence of periods of silence in your Call Analytics transcription output.
Rules using NonTalkTimeFilter
are designed to match:
The presence of silence at specified periods throughout the call
The presence of speech at specified periods throughout the call
See Rule criteria for post-call categories for usage examples.
", "refs": { "Rule$NonTalkTimeFilter": "Flag the presence or absence of periods of silence in your Call Analytics transcription output. Refer to for more detail.
" } @@ -941,7 +945,7 @@ } }, "Rule": { - "base": "A rule is a set of criteria that you can specify to flag an attribute in your Call Analytics output. Rules define a Call Analytics category.
Rules can include these parameters: , , , and .
To learn more about Call Analytics rules and categories, see Creating categories for batch transcriptions and Creating categories for streaming transcriptions.
To learn more about Call Analytics, see Analyzing call center audio with Call Analytics.
", + "base": "A rule is a set of criteria that you can specify to flag an attribute in your Call Analytics output. Rules define a Call Analytics category.
Rules can include these parameters: , , , and .
To learn more about Call Analytics rules and categories, see Creating categories for post-call transcriptions and Creating categories for real-time transcriptions.
To learn more about Call Analytics, see Analyzing call center audio with Call Analytics.
", "refs": { "RuleList$member": null } @@ -955,7 +959,7 @@ } }, "SentimentFilter": { - "base": "Flag the presence or absence of specific sentiments detected in your Call Analytics transcription output.
Rules using SentimentFilter
are designed to match:
The presence or absence of a positive sentiment felt by the customer, agent, or both at specified points in the call
The presence or absence of a negative sentiment felt by the customer, agent, or both at specified points in the call
The presence or absence of a neutral sentiment felt by the customer, agent, or both at specified points in the call
The presence or absence of a mixed sentiment felt by the customer, the agent, or both at specified points in the call
See Rule criteria for batch categories for usage examples.
", + "base": "Flag the presence or absence of specific sentiments detected in your Call Analytics transcription output.
Rules using SentimentFilter
are designed to match:
The presence or absence of a positive sentiment felt by the customer, agent, or both at specified points in the call
The presence or absence of a negative sentiment felt by the customer, agent, or both at specified points in the call
The presence or absence of a neutral sentiment felt by the customer, agent, or both at specified points in the call
The presence or absence of a mixed sentiment felt by the customer, the agent, or both at specified points in the call
See Rule criteria for post-call categories for usage examples.
", "refs": { "Rule$SentimentFilter": "Flag the presence or absence of specific sentiments in your Call Analytics transcription output. Refer to for more detail.
" } @@ -1148,7 +1152,7 @@ } }, "TranscriptFilter": { - "base": "Flag the presence or absence of specific words or phrases detected in your Call Analytics transcription output.
Rules using TranscriptFilter
are designed to match:
Custom words or phrases spoken by the agent, the customer, or both
Custom words or phrases not spoken by the agent, the customer, or either
Custom words or phrases that occur at a specific time frame
See Rule criteria for batch categories and Rule criteria for streaming categories for usage examples.
", + "base": "Flag the presence or absence of specific words or phrases detected in your Call Analytics transcription output.
Rules using TranscriptFilter
are designed to match:
Custom words or phrases spoken by the agent, the customer, or both
Custom words or phrases not spoken by the agent, the customer, or either
Custom words or phrases that occur at a specific time frame
See Rule criteria for post-call categories and Rule criteria for streaming categories for usage examples.
", "refs": { "Rule$TranscriptFilter": "Flag the presence or absence of specific words or phrases in your Call Analytics transcription output. Refer to for more detail.
" } diff --git a/models/apis/transcribe/2017-10-26/endpoint-rule-set-1.json b/models/apis/transcribe/2017-10-26/endpoint-rule-set-1.json index c80101a4c58..6481376ce5a 100644 --- a/models/apis/transcribe/2017-10-26/endpoint-rule-set-1.json +++ b/models/apis/transcribe/2017-10-26/endpoint-rule-set-1.json @@ -3,7 +3,7 @@ "parameters": { "Region": { "builtIn": "AWS::Region", - "required": true, + "required": false, "documentation": "The AWS region used to dispatch the request.", "type": "String" }, @@ -32,13 +32,12 @@ { "conditions": [ { - "fn": "aws.partition", + "fn": "isSet", "argv": [ { - "ref": "Region" + "ref": "Endpoint" } - ], - "assign": "PartitionResult" + ] } ], "type": "tree", @@ -46,14 +45,20 @@ { "conditions": [ { - "fn": "isSet", + "fn": "booleanEquals", "argv": [ { - "ref": "Endpoint" - } + "ref": "UseFIPS" + }, + true ] } ], + "error": "Invalid Configuration: FIPS and custom endpoint are not supported", + "type": "error" + }, + { + "conditions": [], "type": "tree", "rules": [ { @@ -62,67 +67,42 @@ "fn": "booleanEquals", "argv": [ { - "ref": "UseFIPS" + "ref": "UseDualStack" }, true ] } ], - "error": "Invalid Configuration: FIPS and custom endpoint are not supported", + "error": "Invalid Configuration: Dualstack and custom endpoint are not supported", "type": "error" }, { "conditions": [], - "type": "tree", - "rules": [ - { - "conditions": [ - { - "fn": "booleanEquals", - "argv": [ - { - "ref": "UseDualStack" - }, - true - ] - } - ], - "error": "Invalid Configuration: Dualstack and custom endpoint are not supported", - "type": "error" + "endpoint": { + "url": { + "ref": "Endpoint" }, - { - "conditions": [], - "endpoint": { - "url": { - "ref": "Endpoint" - }, - "properties": {}, - "headers": {} - }, - "type": "endpoint" - } - ] + "properties": {}, + "headers": {} + }, + "type": "endpoint" } ] - }, + } + ] + }, + { + "conditions": [], + "type": "tree", + "rules": [ { "conditions": [ { - "fn": "booleanEquals", - "argv": [ - { - "ref": "UseFIPS" - }, - true - ] - }, - { - "fn": "booleanEquals", + "fn": "isSet", "argv": [ { - "ref": "UseDualStack" - }, - true + "ref": "Region" + } ] } ], @@ -131,114 +111,283 @@ { "conditions": [ { - "fn": "booleanEquals", + "fn": "aws.partition", "argv": [ - true, { - "fn": "getAttr", + "ref": "Region" + } + ], + "assign": "PartitionResult" + } + ], + "type": "tree", + "rules": [ + { + "conditions": [ + { + "fn": "booleanEquals", + "argv": [ + { + "ref": "UseFIPS" + }, + true + ] + }, + { + "fn": "booleanEquals", "argv": [ { - "ref": "PartitionResult" + "ref": "UseDualStack" + }, + true + ] + } + ], + "type": "tree", + "rules": [ + { + "conditions": [ + { + "fn": "booleanEquals", + "argv": [ + true, + { + "fn": "getAttr", + "argv": [ + { + "ref": "PartitionResult" + }, + "supportsFIPS" + ] + } + ] }, - "supportsFIPS" + { + "fn": "booleanEquals", + "argv": [ + true, + { + "fn": "getAttr", + "argv": [ + { + "ref": "PartitionResult" + }, + "supportsDualStack" + ] + } + ] + } + ], + "type": "tree", + "rules": [ + { + "conditions": [], + "type": "tree", + "rules": [ + { + "conditions": [], + "endpoint": { + "url": "https://transcribe-fips.{Region}.{PartitionResult#dualStackDnsSuffix}", + "properties": {}, + "headers": {} + }, + "type": "endpoint" + } + ] + } ] + }, + { + "conditions": [], + "error": "FIPS and DualStack are enabled, but this partition does not support one or both", + "type": "error" } ] }, { - "fn": "booleanEquals", - "argv": [ - true, + "conditions": [ { - "fn": "getAttr", + "fn": "booleanEquals", "argv": [ { - "ref": "PartitionResult" + "ref": "UseFIPS" }, - "supportsDualStack" + true + ] + } + ], + "type": "tree", + "rules": [ + { + "conditions": [ + { + "fn": "booleanEquals", + "argv": [ + true, + { + "fn": "getAttr", + "argv": [ + { + "ref": "PartitionResult" + }, + "supportsFIPS" + ] + } + ] + } + ], + "type": "tree", + "rules": [ + { + "conditions": [], + "type": "tree", + "rules": [ + { + "conditions": [ + { + "fn": "stringEquals", + "argv": [ + "aws", + { + "fn": "getAttr", + "argv": [ + { + "ref": "PartitionResult" + }, + "name" + ] + } + ] + } + ], + "endpoint": { + "url": "https://fips.transcribe.{Region}.amazonaws.com", + "properties": {}, + "headers": {} + }, + "type": "endpoint" + }, + { + "conditions": [ + { + "fn": "stringEquals", + "argv": [ + "aws-us-gov", + { + "fn": "getAttr", + "argv": [ + { + "ref": "PartitionResult" + }, + "name" + ] + } + ] + } + ], + "endpoint": { + "url": "https://fips.transcribe.{Region}.amazonaws.com", + "properties": {}, + "headers": {} + }, + "type": "endpoint" + }, + { + "conditions": [], + "endpoint": { + "url": "https://transcribe-fips.{Region}.{PartitionResult#dnsSuffix}", + "properties": {}, + "headers": {} + }, + "type": "endpoint" + } + ] + } ] + }, + { + "conditions": [], + "error": "FIPS is enabled but this partition does not support FIPS", + "type": "error" } ] - } - ], - "type": "tree", - "rules": [ - { - "conditions": [], - "endpoint": { - "url": "https://transcribe-fips.{Region}.{PartitionResult#dualStackDnsSuffix}", - "properties": {}, - "headers": {} - }, - "type": "endpoint" - } - ] - }, - { - "conditions": [], - "error": "FIPS and DualStack are enabled, but this partition does not support one or both", - "type": "error" - } - ] - }, - { - "conditions": [ - { - "fn": "booleanEquals", - "argv": [ - { - "ref": "UseFIPS" }, - true - ] - } - ], - "type": "tree", - "rules": [ - { - "conditions": [ { - "fn": "booleanEquals", - "argv": [ - true, + "conditions": [ { - "fn": "getAttr", + "fn": "booleanEquals", "argv": [ { - "ref": "PartitionResult" + "ref": "UseDualStack" }, - "supportsFIPS" + true ] } - ] - } - ], - "type": "tree", - "rules": [ - { - "conditions": [], + ], "type": "tree", "rules": [ { "conditions": [ { - "fn": "stringEquals", + "fn": "booleanEquals", "argv": [ - "aws", + true, { "fn": "getAttr", "argv": [ { "ref": "PartitionResult" }, - "name" + "supportsDualStack" ] } ] } ], + "type": "tree", + "rules": [ + { + "conditions": [], + "type": "tree", + "rules": [ + { + "conditions": [], + "endpoint": { + "url": "https://transcribe.{Region}.{PartitionResult#dualStackDnsSuffix}", + "properties": {}, + "headers": {} + }, + "type": "endpoint" + } + ] + } + ] + }, + { + "conditions": [], + "error": "DualStack is enabled but this partition does not support DualStack", + "type": "error" + } + ] + }, + { + "conditions": [], + "type": "tree", + "rules": [ + { + "conditions": [ + { + "fn": "stringEquals", + "argv": [ + { + "ref": "Region" + }, + "cn-north-1" + ] + } + ], "endpoint": { - "url": "https://fips.transcribe.{Region}.{PartitionResult#dnsSuffix}", + "url": "https://cn.transcribe.cn-north-1.amazonaws.com.cn", "properties": {}, "headers": {} }, @@ -249,21 +398,15 @@ { "fn": "stringEquals", "argv": [ - "aws-us-gov", { - "fn": "getAttr", - "argv": [ - { - "ref": "PartitionResult" - }, - "name" - ] - } + "ref": "Region" + }, + "cn-northwest-1" ] } ], "endpoint": { - "url": "https://fips.transcribe.{Region}.{PartitionResult#dnsSuffix}", + "url": "https://cn.transcribe.cn-northwest-1.amazonaws.com.cn", "properties": {}, "headers": {} }, @@ -272,7 +415,7 @@ { "conditions": [], "endpoint": { - "url": "https://transcribe-fips.{Region}.{PartitionResult#dnsSuffix}", + "url": "https://transcribe.{Region}.{PartitionResult#dnsSuffix}", "properties": {}, "headers": {} }, @@ -281,118 +424,13 @@ ] } ] - }, - { - "conditions": [], - "error": "FIPS is enabled but this partition does not support FIPS", - "type": "error" - } - ] - }, - { - "conditions": [ - { - "fn": "booleanEquals", - "argv": [ - { - "ref": "UseDualStack" - }, - true - ] - } - ], - "type": "tree", - "rules": [ - { - "conditions": [ - { - "fn": "booleanEquals", - "argv": [ - true, - { - "fn": "getAttr", - "argv": [ - { - "ref": "PartitionResult" - }, - "supportsDualStack" - ] - } - ] - } - ], - "type": "tree", - "rules": [ - { - "conditions": [], - "endpoint": { - "url": "https://transcribe.{Region}.{PartitionResult#dualStackDnsSuffix}", - "properties": {}, - "headers": {} - }, - "type": "endpoint" - } - ] - }, - { - "conditions": [], - "error": "DualStack is enabled but this partition does not support DualStack", - "type": "error" } ] }, { "conditions": [], - "type": "tree", - "rules": [ - { - "conditions": [ - { - "fn": "stringEquals", - "argv": [ - { - "ref": "Region" - }, - "cn-north-1" - ] - } - ], - "endpoint": { - "url": "https://cn.transcribe.cn-north-1.amazonaws.com.cn", - "properties": {}, - "headers": {} - }, - "type": "endpoint" - }, - { - "conditions": [ - { - "fn": "stringEquals", - "argv": [ - { - "ref": "Region" - }, - "cn-northwest-1" - ] - } - ], - "endpoint": { - "url": "https://cn.transcribe.cn-northwest-1.amazonaws.com.cn", - "properties": {}, - "headers": {} - }, - "type": "endpoint" - }, - { - "conditions": [], - "endpoint": { - "url": "https://transcribe.{Region}.{PartitionResult#dnsSuffix}", - "properties": {}, - "headers": {} - }, - "type": "endpoint" - } - ] + "error": "Invalid Configuration: Missing Region", + "type": "error" } ] } diff --git a/models/apis/transcribe/2017-10-26/endpoint-tests-1.json b/models/apis/transcribe/2017-10-26/endpoint-tests-1.json index 0a836f083f8..f6b5504dd0a 100644 --- a/models/apis/transcribe/2017-10-26/endpoint-tests-1.json +++ b/models/apis/transcribe/2017-10-26/endpoint-tests-1.json @@ -1,666 +1,29 @@ { "testCases": [ { - "documentation": "For region ap-south-1 with FIPS enabled and DualStack enabled", - "expect": { - "endpoint": { - "url": "https://transcribe-fips.ap-south-1.api.aws" - } - }, - "params": { - "UseFIPS": true, - "Region": "ap-south-1", - "UseDualStack": true - } - }, - { - "documentation": "For region ap-south-1 with FIPS enabled and DualStack disabled", - "expect": { - "endpoint": { - "url": "https://fips.transcribe.ap-south-1.amazonaws.com" - } - }, - "params": { - "UseFIPS": true, - "Region": "ap-south-1", - "UseDualStack": false - } - }, - { - "documentation": "For region ap-south-1 with FIPS disabled and DualStack enabled", - "expect": { - "endpoint": { - "url": "https://transcribe.ap-south-1.api.aws" - } - }, - "params": { - "UseFIPS": false, - "Region": "ap-south-1", - "UseDualStack": true - } - }, - { - "documentation": "For region ap-south-1 with FIPS disabled and DualStack disabled", - "expect": { - "endpoint": { - "url": "https://transcribe.ap-south-1.amazonaws.com" - } - }, - "params": { - "UseFIPS": false, - "Region": "ap-south-1", - "UseDualStack": false - } - }, - { - "documentation": "For region us-gov-east-1 with FIPS enabled and DualStack enabled", - "expect": { - "endpoint": { - "url": "https://transcribe-fips.us-gov-east-1.api.aws" - } - }, - "params": { - "UseFIPS": true, - "Region": "us-gov-east-1", - "UseDualStack": true - } - }, - { - "documentation": "For region us-gov-east-1 with FIPS enabled and DualStack disabled", - "expect": { - "endpoint": { - "url": "https://fips.transcribe.us-gov-east-1.amazonaws.com" - } - }, - "params": { - "UseFIPS": true, - "Region": "us-gov-east-1", - "UseDualStack": false - } - }, - { - "documentation": "For region us-gov-east-1 with FIPS disabled and DualStack enabled", - "expect": { - "endpoint": { - "url": "https://transcribe.us-gov-east-1.api.aws" - } - }, - "params": { - "UseFIPS": false, - "Region": "us-gov-east-1", - "UseDualStack": true - } - }, - { - "documentation": "For region us-gov-east-1 with FIPS disabled and DualStack disabled", - "expect": { - "endpoint": { - "url": "https://transcribe.us-gov-east-1.amazonaws.com" - } - }, - "params": { - "UseFIPS": false, - "Region": "us-gov-east-1", - "UseDualStack": false - } - }, - { - "documentation": "For region ca-central-1 with FIPS enabled and DualStack enabled", - "expect": { - "endpoint": { - "url": "https://transcribe-fips.ca-central-1.api.aws" - } - }, - "params": { - "UseFIPS": true, - "Region": "ca-central-1", - "UseDualStack": true - } - }, - { - "documentation": "For region ca-central-1 with FIPS enabled and DualStack disabled", - "expect": { - "endpoint": { - "url": "https://fips.transcribe.ca-central-1.amazonaws.com" - } - }, - "params": { - "UseFIPS": true, - "Region": "ca-central-1", - "UseDualStack": false - } - }, - { - "documentation": "For region ca-central-1 with FIPS disabled and DualStack enabled", - "expect": { - "endpoint": { - "url": "https://transcribe.ca-central-1.api.aws" - } - }, - "params": { - "UseFIPS": false, - "Region": "ca-central-1", - "UseDualStack": true - } - }, - { - "documentation": "For region ca-central-1 with FIPS disabled and DualStack disabled", - "expect": { - "endpoint": { - "url": "https://transcribe.ca-central-1.amazonaws.com" - } - }, - "params": { - "UseFIPS": false, - "Region": "ca-central-1", - "UseDualStack": false - } - }, - { - "documentation": "For region eu-central-1 with FIPS enabled and DualStack enabled", - "expect": { - "endpoint": { - "url": "https://transcribe-fips.eu-central-1.api.aws" - } - }, - "params": { - "UseFIPS": true, - "Region": "eu-central-1", - "UseDualStack": true - } - }, - { - "documentation": "For region eu-central-1 with FIPS enabled and DualStack disabled", - "expect": { - "endpoint": { - "url": "https://fips.transcribe.eu-central-1.amazonaws.com" - } - }, - "params": { - "UseFIPS": true, - "Region": "eu-central-1", - "UseDualStack": false - } - }, - { - "documentation": "For region eu-central-1 with FIPS disabled and DualStack enabled", - "expect": { - "endpoint": { - "url": "https://transcribe.eu-central-1.api.aws" - } - }, - "params": { - "UseFIPS": false, - "Region": "eu-central-1", - "UseDualStack": true - } - }, - { - "documentation": "For region eu-central-1 with FIPS disabled and DualStack disabled", - "expect": { - "endpoint": { - "url": "https://transcribe.eu-central-1.amazonaws.com" - } - }, - "params": { - "UseFIPS": false, - "Region": "eu-central-1", - "UseDualStack": false - } - }, - { - "documentation": "For region us-west-1 with FIPS enabled and DualStack enabled", - "expect": { - "endpoint": { - "url": "https://transcribe-fips.us-west-1.api.aws" - } - }, - "params": { - "UseFIPS": true, - "Region": "us-west-1", - "UseDualStack": true - } - }, - { - "documentation": "For region us-west-1 with FIPS enabled and DualStack disabled", - "expect": { - "endpoint": { - "url": "https://fips.transcribe.us-west-1.amazonaws.com" - } - }, - "params": { - "UseFIPS": true, - "Region": "us-west-1", - "UseDualStack": false - } - }, - { - "documentation": "For region us-west-1 with FIPS disabled and DualStack enabled", - "expect": { - "endpoint": { - "url": "https://transcribe.us-west-1.api.aws" - } - }, - "params": { - "UseFIPS": false, - "Region": "us-west-1", - "UseDualStack": true - } - }, - { - "documentation": "For region us-west-1 with FIPS disabled and DualStack disabled", - "expect": { - "endpoint": { - "url": "https://transcribe.us-west-1.amazonaws.com" - } - }, - "params": { - "UseFIPS": false, - "Region": "us-west-1", - "UseDualStack": false - } - }, - { - "documentation": "For region us-west-2 with FIPS enabled and DualStack enabled", - "expect": { - "endpoint": { - "url": "https://transcribe-fips.us-west-2.api.aws" - } - }, - "params": { - "UseFIPS": true, - "Region": "us-west-2", - "UseDualStack": true - } - }, - { - "documentation": "For region us-west-2 with FIPS enabled and DualStack disabled", - "expect": { - "endpoint": { - "url": "https://fips.transcribe.us-west-2.amazonaws.com" - } - }, - "params": { - "UseFIPS": true, - "Region": "us-west-2", - "UseDualStack": false - } - }, - { - "documentation": "For region us-west-2 with FIPS disabled and DualStack enabled", - "expect": { - "endpoint": { - "url": "https://transcribe.us-west-2.api.aws" - } - }, - "params": { - "UseFIPS": false, - "Region": "us-west-2", - "UseDualStack": true - } - }, - { - "documentation": "For region us-west-2 with FIPS disabled and DualStack disabled", - "expect": { - "endpoint": { - "url": "https://transcribe.us-west-2.amazonaws.com" - } - }, - "params": { - "UseFIPS": false, - "Region": "us-west-2", - "UseDualStack": false - } - }, - { - "documentation": "For region af-south-1 with FIPS enabled and DualStack enabled", - "expect": { - "endpoint": { - "url": "https://transcribe-fips.af-south-1.api.aws" - } - }, - "params": { - "UseFIPS": true, - "Region": "af-south-1", - "UseDualStack": true - } - }, - { - "documentation": "For region af-south-1 with FIPS enabled and DualStack disabled", - "expect": { - "endpoint": { - "url": "https://fips.transcribe.af-south-1.amazonaws.com" - } - }, - "params": { - "UseFIPS": true, - "Region": "af-south-1", - "UseDualStack": false - } - }, - { - "documentation": "For region af-south-1 with FIPS disabled and DualStack enabled", - "expect": { - "endpoint": { - "url": "https://transcribe.af-south-1.api.aws" - } - }, - "params": { - "UseFIPS": false, - "Region": "af-south-1", - "UseDualStack": true - } - }, - { - "documentation": "For region af-south-1 with FIPS disabled and DualStack disabled", - "expect": { - "endpoint": { - "url": "https://transcribe.af-south-1.amazonaws.com" - } - }, - "params": { - "UseFIPS": false, - "Region": "af-south-1", - "UseDualStack": false - } - }, - { - "documentation": "For region eu-north-1 with FIPS enabled and DualStack enabled", - "expect": { - "endpoint": { - "url": "https://transcribe-fips.eu-north-1.api.aws" - } - }, - "params": { - "UseFIPS": true, - "Region": "eu-north-1", - "UseDualStack": true - } - }, - { - "documentation": "For region eu-north-1 with FIPS enabled and DualStack disabled", - "expect": { - "endpoint": { - "url": "https://fips.transcribe.eu-north-1.amazonaws.com" - } - }, - "params": { - "UseFIPS": true, - "Region": "eu-north-1", - "UseDualStack": false - } - }, - { - "documentation": "For region eu-north-1 with FIPS disabled and DualStack enabled", - "expect": { - "endpoint": { - "url": "https://transcribe.eu-north-1.api.aws" - } - }, - "params": { - "UseFIPS": false, - "Region": "eu-north-1", - "UseDualStack": true - } - }, - { - "documentation": "For region eu-north-1 with FIPS disabled and DualStack disabled", - "expect": { - "endpoint": { - "url": "https://transcribe.eu-north-1.amazonaws.com" - } - }, - "params": { - "UseFIPS": false, - "Region": "eu-north-1", - "UseDualStack": false - } - }, - { - "documentation": "For region eu-west-3 with FIPS enabled and DualStack enabled", - "expect": { - "endpoint": { - "url": "https://transcribe-fips.eu-west-3.api.aws" - } - }, - "params": { - "UseFIPS": true, - "Region": "eu-west-3", - "UseDualStack": true - } - }, - { - "documentation": "For region eu-west-3 with FIPS enabled and DualStack disabled", - "expect": { - "endpoint": { - "url": "https://fips.transcribe.eu-west-3.amazonaws.com" - } - }, - "params": { - "UseFIPS": true, - "Region": "eu-west-3", - "UseDualStack": false - } - }, - { - "documentation": "For region eu-west-3 with FIPS disabled and DualStack enabled", - "expect": { - "endpoint": { - "url": "https://transcribe.eu-west-3.api.aws" - } - }, - "params": { - "UseFIPS": false, - "Region": "eu-west-3", - "UseDualStack": true - } - }, - { - "documentation": "For region eu-west-3 with FIPS disabled and DualStack disabled", - "expect": { - "endpoint": { - "url": "https://transcribe.eu-west-3.amazonaws.com" - } - }, - "params": { - "UseFIPS": false, - "Region": "eu-west-3", - "UseDualStack": false - } - }, - { - "documentation": "For region eu-west-2 with FIPS enabled and DualStack enabled", - "expect": { - "endpoint": { - "url": "https://transcribe-fips.eu-west-2.api.aws" - } - }, - "params": { - "UseFIPS": true, - "Region": "eu-west-2", - "UseDualStack": true - } - }, - { - "documentation": "For region eu-west-2 with FIPS enabled and DualStack disabled", - "expect": { - "endpoint": { - "url": "https://fips.transcribe.eu-west-2.amazonaws.com" - } - }, - "params": { - "UseFIPS": true, - "Region": "eu-west-2", - "UseDualStack": false - } - }, - { - "documentation": "For region eu-west-2 with FIPS disabled and DualStack enabled", - "expect": { - "endpoint": { - "url": "https://transcribe.eu-west-2.api.aws" - } - }, - "params": { - "UseFIPS": false, - "Region": "eu-west-2", - "UseDualStack": true - } - }, - { - "documentation": "For region eu-west-2 with FIPS disabled and DualStack disabled", - "expect": { - "endpoint": { - "url": "https://transcribe.eu-west-2.amazonaws.com" - } - }, - "params": { - "UseFIPS": false, - "Region": "eu-west-2", - "UseDualStack": false - } - }, - { - "documentation": "For region eu-west-1 with FIPS enabled and DualStack enabled", - "expect": { - "endpoint": { - "url": "https://transcribe-fips.eu-west-1.api.aws" - } - }, - "params": { - "UseFIPS": true, - "Region": "eu-west-1", - "UseDualStack": true - } - }, - { - "documentation": "For region eu-west-1 with FIPS enabled and DualStack disabled", - "expect": { - "endpoint": { - "url": "https://fips.transcribe.eu-west-1.amazonaws.com" - } - }, - "params": { - "UseFIPS": true, - "Region": "eu-west-1", - "UseDualStack": false - } - }, - { - "documentation": "For region eu-west-1 with FIPS disabled and DualStack enabled", - "expect": { - "endpoint": { - "url": "https://transcribe.eu-west-1.api.aws" - } - }, - "params": { - "UseFIPS": false, - "Region": "eu-west-1", - "UseDualStack": true - } - }, - { - "documentation": "For region eu-west-1 with FIPS disabled and DualStack disabled", - "expect": { - "endpoint": { - "url": "https://transcribe.eu-west-1.amazonaws.com" - } - }, - "params": { - "UseFIPS": false, - "Region": "eu-west-1", - "UseDualStack": false - } - }, - { - "documentation": "For region ap-northeast-2 with FIPS enabled and DualStack enabled", - "expect": { - "endpoint": { - "url": "https://transcribe-fips.ap-northeast-2.api.aws" - } - }, - "params": { - "UseFIPS": true, - "Region": "ap-northeast-2", - "UseDualStack": true - } - }, - { - "documentation": "For region ap-northeast-2 with FIPS enabled and DualStack disabled", - "expect": { - "endpoint": { - "url": "https://fips.transcribe.ap-northeast-2.amazonaws.com" - } - }, - "params": { - "UseFIPS": true, - "Region": "ap-northeast-2", - "UseDualStack": false - } - }, - { - "documentation": "For region ap-northeast-2 with FIPS disabled and DualStack enabled", - "expect": { - "endpoint": { - "url": "https://transcribe.ap-northeast-2.api.aws" - } - }, - "params": { - "UseFIPS": false, - "Region": "ap-northeast-2", - "UseDualStack": true - } - }, - { - "documentation": "For region ap-northeast-2 with FIPS disabled and DualStack disabled", + "documentation": "For region af-south-1 with FIPS disabled and DualStack disabled", "expect": { "endpoint": { - "url": "https://transcribe.ap-northeast-2.amazonaws.com" + "url": "https://transcribe.af-south-1.amazonaws.com" } }, "params": { "UseFIPS": false, - "Region": "ap-northeast-2", - "UseDualStack": false - } - }, - { - "documentation": "For region ap-northeast-1 with FIPS enabled and DualStack enabled", - "expect": { - "endpoint": { - "url": "https://transcribe-fips.ap-northeast-1.api.aws" - } - }, - "params": { - "UseFIPS": true, - "Region": "ap-northeast-1", - "UseDualStack": true - } - }, - { - "documentation": "For region ap-northeast-1 with FIPS enabled and DualStack disabled", - "expect": { - "endpoint": { - "url": "https://fips.transcribe.ap-northeast-1.amazonaws.com" - } - }, - "params": { - "UseFIPS": true, - "Region": "ap-northeast-1", - "UseDualStack": false + "UseDualStack": false, + "Region": "af-south-1" } }, { - "documentation": "For region ap-northeast-1 with FIPS disabled and DualStack enabled", + "documentation": "For region ap-east-1 with FIPS disabled and DualStack disabled", "expect": { "endpoint": { - "url": "https://transcribe.ap-northeast-1.api.aws" + "url": "https://transcribe.ap-east-1.amazonaws.com" } }, "params": { "UseFIPS": false, - "Region": "ap-northeast-1", - "UseDualStack": true + "UseDualStack": false, + "Region": "ap-east-1" } }, { @@ -672,580 +35,520 @@ }, "params": { "UseFIPS": false, - "Region": "ap-northeast-1", - "UseDualStack": false + "UseDualStack": false, + "Region": "ap-northeast-1" } }, { - "documentation": "For region me-south-1 with FIPS enabled and DualStack enabled", + "documentation": "For region ap-northeast-2 with FIPS disabled and DualStack disabled", "expect": { "endpoint": { - "url": "https://transcribe-fips.me-south-1.api.aws" + "url": "https://transcribe.ap-northeast-2.amazonaws.com" } }, "params": { - "UseFIPS": true, - "Region": "me-south-1", - "UseDualStack": true + "UseFIPS": false, + "UseDualStack": false, + "Region": "ap-northeast-2" } }, { - "documentation": "For region me-south-1 with FIPS enabled and DualStack disabled", + "documentation": "For region ap-south-1 with FIPS disabled and DualStack disabled", "expect": { "endpoint": { - "url": "https://fips.transcribe.me-south-1.amazonaws.com" + "url": "https://transcribe.ap-south-1.amazonaws.com" } }, "params": { - "UseFIPS": true, - "Region": "me-south-1", - "UseDualStack": false + "UseFIPS": false, + "UseDualStack": false, + "Region": "ap-south-1" } }, { - "documentation": "For region me-south-1 with FIPS disabled and DualStack enabled", + "documentation": "For region ap-southeast-1 with FIPS disabled and DualStack disabled", "expect": { "endpoint": { - "url": "https://transcribe.me-south-1.api.aws" + "url": "https://transcribe.ap-southeast-1.amazonaws.com" } }, "params": { "UseFIPS": false, - "Region": "me-south-1", - "UseDualStack": true + "UseDualStack": false, + "Region": "ap-southeast-1" } }, { - "documentation": "For region me-south-1 with FIPS disabled and DualStack disabled", + "documentation": "For region ap-southeast-2 with FIPS disabled and DualStack disabled", "expect": { "endpoint": { - "url": "https://transcribe.me-south-1.amazonaws.com" + "url": "https://transcribe.ap-southeast-2.amazonaws.com" } }, "params": { "UseFIPS": false, - "Region": "me-south-1", - "UseDualStack": false + "UseDualStack": false, + "Region": "ap-southeast-2" } }, { - "documentation": "For region sa-east-1 with FIPS enabled and DualStack enabled", + "documentation": "For region ca-central-1 with FIPS disabled and DualStack disabled", "expect": { "endpoint": { - "url": "https://transcribe-fips.sa-east-1.api.aws" + "url": "https://transcribe.ca-central-1.amazonaws.com" } }, "params": { - "UseFIPS": true, - "Region": "sa-east-1", - "UseDualStack": true + "UseFIPS": false, + "UseDualStack": false, + "Region": "ca-central-1" } }, { - "documentation": "For region sa-east-1 with FIPS enabled and DualStack disabled", + "documentation": "For region ca-central-1 with FIPS enabled and DualStack disabled", "expect": { "endpoint": { - "url": "https://fips.transcribe.sa-east-1.amazonaws.com" + "url": "https://fips.transcribe.ca-central-1.amazonaws.com" } }, "params": { "UseFIPS": true, - "Region": "sa-east-1", - "UseDualStack": false + "UseDualStack": false, + "Region": "ca-central-1" } }, { - "documentation": "For region sa-east-1 with FIPS disabled and DualStack enabled", + "documentation": "For region eu-central-1 with FIPS disabled and DualStack disabled", "expect": { "endpoint": { - "url": "https://transcribe.sa-east-1.api.aws" + "url": "https://transcribe.eu-central-1.amazonaws.com" } }, "params": { "UseFIPS": false, - "Region": "sa-east-1", - "UseDualStack": true + "UseDualStack": false, + "Region": "eu-central-1" } }, { - "documentation": "For region sa-east-1 with FIPS disabled and DualStack disabled", + "documentation": "For region eu-north-1 with FIPS disabled and DualStack disabled", "expect": { "endpoint": { - "url": "https://transcribe.sa-east-1.amazonaws.com" + "url": "https://transcribe.eu-north-1.amazonaws.com" } }, "params": { "UseFIPS": false, - "Region": "sa-east-1", - "UseDualStack": false + "UseDualStack": false, + "Region": "eu-north-1" } }, { - "documentation": "For region ap-east-1 with FIPS enabled and DualStack enabled", + "documentation": "For region eu-west-1 with FIPS disabled and DualStack disabled", "expect": { "endpoint": { - "url": "https://transcribe-fips.ap-east-1.api.aws" + "url": "https://transcribe.eu-west-1.amazonaws.com" } }, "params": { - "UseFIPS": true, - "Region": "ap-east-1", - "UseDualStack": true + "UseFIPS": false, + "UseDualStack": false, + "Region": "eu-west-1" } }, { - "documentation": "For region ap-east-1 with FIPS enabled and DualStack disabled", + "documentation": "For region eu-west-2 with FIPS disabled and DualStack disabled", "expect": { "endpoint": { - "url": "https://fips.transcribe.ap-east-1.amazonaws.com" + "url": "https://transcribe.eu-west-2.amazonaws.com" } }, "params": { - "UseFIPS": true, - "Region": "ap-east-1", - "UseDualStack": false + "UseFIPS": false, + "UseDualStack": false, + "Region": "eu-west-2" } }, { - "documentation": "For region ap-east-1 with FIPS disabled and DualStack enabled", + "documentation": "For region eu-west-3 with FIPS disabled and DualStack disabled", "expect": { "endpoint": { - "url": "https://transcribe.ap-east-1.api.aws" + "url": "https://transcribe.eu-west-3.amazonaws.com" } }, "params": { "UseFIPS": false, - "Region": "ap-east-1", - "UseDualStack": true + "UseDualStack": false, + "Region": "eu-west-3" } }, { - "documentation": "For region ap-east-1 with FIPS disabled and DualStack disabled", + "documentation": "For region me-south-1 with FIPS disabled and DualStack disabled", "expect": { "endpoint": { - "url": "https://transcribe.ap-east-1.amazonaws.com" + "url": "https://transcribe.me-south-1.amazonaws.com" } }, "params": { "UseFIPS": false, - "Region": "ap-east-1", - "UseDualStack": false + "UseDualStack": false, + "Region": "me-south-1" } }, { - "documentation": "For region cn-north-1 with FIPS enabled and DualStack enabled", + "documentation": "For region sa-east-1 with FIPS disabled and DualStack disabled", "expect": { "endpoint": { - "url": "https://transcribe-fips.cn-north-1.api.amazonwebservices.com.cn" + "url": "https://transcribe.sa-east-1.amazonaws.com" } }, "params": { - "UseFIPS": true, - "Region": "cn-north-1", - "UseDualStack": true + "UseFIPS": false, + "UseDualStack": false, + "Region": "sa-east-1" } }, { - "documentation": "For region cn-north-1 with FIPS enabled and DualStack disabled", + "documentation": "For region us-east-1 with FIPS disabled and DualStack disabled", "expect": { "endpoint": { - "url": "https://transcribe-fips.cn-north-1.amazonaws.com.cn" + "url": "https://transcribe.us-east-1.amazonaws.com" } }, "params": { - "UseFIPS": true, - "Region": "cn-north-1", - "UseDualStack": false + "UseFIPS": false, + "UseDualStack": false, + "Region": "us-east-1" } }, { - "documentation": "For region cn-north-1 with FIPS disabled and DualStack enabled", + "documentation": "For region us-east-1 with FIPS enabled and DualStack disabled", "expect": { "endpoint": { - "url": "https://transcribe.cn-north-1.api.amazonwebservices.com.cn" + "url": "https://fips.transcribe.us-east-1.amazonaws.com" } }, "params": { - "UseFIPS": false, - "Region": "cn-north-1", - "UseDualStack": true + "UseFIPS": true, + "UseDualStack": false, + "Region": "us-east-1" } }, { - "documentation": "For region cn-north-1 with FIPS disabled and DualStack disabled", + "documentation": "For region us-east-2 with FIPS disabled and DualStack disabled", "expect": { "endpoint": { - "url": "https://cn.transcribe.cn-north-1.amazonaws.com.cn" + "url": "https://transcribe.us-east-2.amazonaws.com" } }, "params": { "UseFIPS": false, - "Region": "cn-north-1", - "UseDualStack": false + "UseDualStack": false, + "Region": "us-east-2" } }, { - "documentation": "For region us-gov-west-1 with FIPS enabled and DualStack enabled", + "documentation": "For region us-east-2 with FIPS enabled and DualStack disabled", "expect": { "endpoint": { - "url": "https://transcribe-fips.us-gov-west-1.api.aws" + "url": "https://fips.transcribe.us-east-2.amazonaws.com" } }, "params": { "UseFIPS": true, - "Region": "us-gov-west-1", - "UseDualStack": true + "UseDualStack": false, + "Region": "us-east-2" } }, { - "documentation": "For region us-gov-west-1 with FIPS enabled and DualStack disabled", + "documentation": "For region us-west-1 with FIPS disabled and DualStack disabled", "expect": { "endpoint": { - "url": "https://fips.transcribe.us-gov-west-1.amazonaws.com" + "url": "https://transcribe.us-west-1.amazonaws.com" } }, "params": { - "UseFIPS": true, - "Region": "us-gov-west-1", - "UseDualStack": false + "UseFIPS": false, + "UseDualStack": false, + "Region": "us-west-1" } }, { - "documentation": "For region us-gov-west-1 with FIPS disabled and DualStack enabled", + "documentation": "For region us-west-1 with FIPS enabled and DualStack disabled", "expect": { "endpoint": { - "url": "https://transcribe.us-gov-west-1.api.aws" + "url": "https://fips.transcribe.us-west-1.amazonaws.com" } }, "params": { - "UseFIPS": false, - "Region": "us-gov-west-1", - "UseDualStack": true + "UseFIPS": true, + "UseDualStack": false, + "Region": "us-west-1" } }, { - "documentation": "For region us-gov-west-1 with FIPS disabled and DualStack disabled", + "documentation": "For region us-west-2 with FIPS disabled and DualStack disabled", "expect": { "endpoint": { - "url": "https://transcribe.us-gov-west-1.amazonaws.com" + "url": "https://transcribe.us-west-2.amazonaws.com" } }, "params": { "UseFIPS": false, - "Region": "us-gov-west-1", - "UseDualStack": false + "UseDualStack": false, + "Region": "us-west-2" } }, { - "documentation": "For region ap-southeast-1 with FIPS enabled and DualStack enabled", + "documentation": "For region us-west-2 with FIPS enabled and DualStack disabled", "expect": { "endpoint": { - "url": "https://transcribe-fips.ap-southeast-1.api.aws" + "url": "https://fips.transcribe.us-west-2.amazonaws.com" } }, "params": { "UseFIPS": true, - "Region": "ap-southeast-1", - "UseDualStack": true + "UseDualStack": false, + "Region": "us-west-2" } }, { - "documentation": "For region ap-southeast-1 with FIPS enabled and DualStack disabled", + "documentation": "For region us-east-1 with FIPS enabled and DualStack enabled", "expect": { "endpoint": { - "url": "https://fips.transcribe.ap-southeast-1.amazonaws.com" + "url": "https://transcribe-fips.us-east-1.api.aws" } }, "params": { "UseFIPS": true, - "Region": "ap-southeast-1", - "UseDualStack": false + "UseDualStack": true, + "Region": "us-east-1" } }, { - "documentation": "For region ap-southeast-1 with FIPS disabled and DualStack enabled", + "documentation": "For region us-east-1 with FIPS disabled and DualStack enabled", "expect": { "endpoint": { - "url": "https://transcribe.ap-southeast-1.api.aws" + "url": "https://transcribe.us-east-1.api.aws" } }, "params": { "UseFIPS": false, - "Region": "ap-southeast-1", - "UseDualStack": true + "UseDualStack": true, + "Region": "us-east-1" } }, { - "documentation": "For region ap-southeast-1 with FIPS disabled and DualStack disabled", + "documentation": "For region cn-north-1 with FIPS disabled and DualStack disabled", "expect": { "endpoint": { - "url": "https://transcribe.ap-southeast-1.amazonaws.com" + "url": "https://cn.transcribe.cn-north-1.amazonaws.com.cn" } }, "params": { "UseFIPS": false, - "Region": "ap-southeast-1", - "UseDualStack": false - } - }, - { - "documentation": "For region ap-southeast-2 with FIPS enabled and DualStack enabled", - "expect": { - "endpoint": { - "url": "https://transcribe-fips.ap-southeast-2.api.aws" - } - }, - "params": { - "UseFIPS": true, - "Region": "ap-southeast-2", - "UseDualStack": true - } - }, - { - "documentation": "For region ap-southeast-2 with FIPS enabled and DualStack disabled", - "expect": { - "endpoint": { - "url": "https://fips.transcribe.ap-southeast-2.amazonaws.com" - } - }, - "params": { - "UseFIPS": true, - "Region": "ap-southeast-2", - "UseDualStack": false + "UseDualStack": false, + "Region": "cn-north-1" } }, { - "documentation": "For region ap-southeast-2 with FIPS disabled and DualStack enabled", + "documentation": "For region cn-northwest-1 with FIPS disabled and DualStack disabled", "expect": { "endpoint": { - "url": "https://transcribe.ap-southeast-2.api.aws" + "url": "https://cn.transcribe.cn-northwest-1.amazonaws.com.cn" } }, "params": { "UseFIPS": false, - "Region": "ap-southeast-2", - "UseDualStack": true + "UseDualStack": false, + "Region": "cn-northwest-1" } }, { - "documentation": "For region ap-southeast-2 with FIPS disabled and DualStack disabled", + "documentation": "For region cn-north-1 with FIPS enabled and DualStack enabled", "expect": { "endpoint": { - "url": "https://transcribe.ap-southeast-2.amazonaws.com" + "url": "https://transcribe-fips.cn-north-1.api.amazonwebservices.com.cn" } }, - "params": { - "UseFIPS": false, - "Region": "ap-southeast-2", - "UseDualStack": false - } - }, - { - "documentation": "For region us-iso-east-1 with FIPS enabled and DualStack enabled", - "expect": { - "error": "FIPS and DualStack are enabled, but this partition does not support one or both" - }, "params": { "UseFIPS": true, - "Region": "us-iso-east-1", - "UseDualStack": true + "UseDualStack": true, + "Region": "cn-north-1" } }, { - "documentation": "For region us-iso-east-1 with FIPS enabled and DualStack disabled", + "documentation": "For region cn-north-1 with FIPS enabled and DualStack disabled", "expect": { "endpoint": { - "url": "https://transcribe-fips.us-iso-east-1.c2s.ic.gov" + "url": "https://transcribe-fips.cn-north-1.amazonaws.com.cn" } }, "params": { "UseFIPS": true, - "Region": "us-iso-east-1", - "UseDualStack": false - } - }, - { - "documentation": "For region us-iso-east-1 with FIPS disabled and DualStack enabled", - "expect": { - "error": "DualStack is enabled but this partition does not support DualStack" - }, - "params": { - "UseFIPS": false, - "Region": "us-iso-east-1", - "UseDualStack": true + "UseDualStack": false, + "Region": "cn-north-1" } }, { - "documentation": "For region us-iso-east-1 with FIPS disabled and DualStack disabled", + "documentation": "For region cn-north-1 with FIPS disabled and DualStack enabled", "expect": { "endpoint": { - "url": "https://transcribe.us-iso-east-1.c2s.ic.gov" + "url": "https://transcribe.cn-north-1.api.amazonwebservices.com.cn" } }, "params": { "UseFIPS": false, - "Region": "us-iso-east-1", - "UseDualStack": false + "UseDualStack": true, + "Region": "cn-north-1" } }, { - "documentation": "For region us-east-1 with FIPS enabled and DualStack enabled", + "documentation": "For region us-gov-east-1 with FIPS disabled and DualStack disabled", "expect": { "endpoint": { - "url": "https://transcribe-fips.us-east-1.api.aws" + "url": "https://transcribe.us-gov-east-1.amazonaws.com" } }, "params": { - "UseFIPS": true, - "Region": "us-east-1", - "UseDualStack": true + "UseFIPS": false, + "UseDualStack": false, + "Region": "us-gov-east-1" } }, { - "documentation": "For region us-east-1 with FIPS enabled and DualStack disabled", + "documentation": "For region us-gov-east-1 with FIPS enabled and DualStack disabled", "expect": { "endpoint": { - "url": "https://fips.transcribe.us-east-1.amazonaws.com" + "url": "https://fips.transcribe.us-gov-east-1.amazonaws.com" } }, "params": { "UseFIPS": true, - "Region": "us-east-1", - "UseDualStack": false - } - }, - { - "documentation": "For region us-east-1 with FIPS disabled and DualStack enabled", - "expect": { - "endpoint": { - "url": "https://transcribe.us-east-1.api.aws" - } - }, - "params": { - "UseFIPS": false, - "Region": "us-east-1", - "UseDualStack": true + "UseDualStack": false, + "Region": "us-gov-east-1" } }, { - "documentation": "For region us-east-1 with FIPS disabled and DualStack disabled", + "documentation": "For region us-gov-west-1 with FIPS disabled and DualStack disabled", "expect": { "endpoint": { - "url": "https://transcribe.us-east-1.amazonaws.com" + "url": "https://transcribe.us-gov-west-1.amazonaws.com" } }, "params": { "UseFIPS": false, - "Region": "us-east-1", - "UseDualStack": false + "UseDualStack": false, + "Region": "us-gov-west-1" } }, { - "documentation": "For region us-east-2 with FIPS enabled and DualStack enabled", + "documentation": "For region us-gov-west-1 with FIPS enabled and DualStack disabled", "expect": { "endpoint": { - "url": "https://transcribe-fips.us-east-2.api.aws" + "url": "https://fips.transcribe.us-gov-west-1.amazonaws.com" } }, "params": { "UseFIPS": true, - "Region": "us-east-2", - "UseDualStack": true + "UseDualStack": false, + "Region": "us-gov-west-1" } }, { - "documentation": "For region us-east-2 with FIPS enabled and DualStack disabled", + "documentation": "For region us-gov-east-1 with FIPS enabled and DualStack enabled", "expect": { "endpoint": { - "url": "https://fips.transcribe.us-east-2.amazonaws.com" + "url": "https://transcribe-fips.us-gov-east-1.api.aws" } }, "params": { "UseFIPS": true, - "Region": "us-east-2", - "UseDualStack": false + "UseDualStack": true, + "Region": "us-gov-east-1" } }, { - "documentation": "For region us-east-2 with FIPS disabled and DualStack enabled", + "documentation": "For region us-gov-east-1 with FIPS disabled and DualStack enabled", "expect": { "endpoint": { - "url": "https://transcribe.us-east-2.api.aws" + "url": "https://transcribe.us-gov-east-1.api.aws" } }, "params": { "UseFIPS": false, - "Region": "us-east-2", - "UseDualStack": true + "UseDualStack": true, + "Region": "us-gov-east-1" } }, { - "documentation": "For region us-east-2 with FIPS disabled and DualStack disabled", + "documentation": "For region us-iso-east-1 with FIPS disabled and DualStack disabled", "expect": { "endpoint": { - "url": "https://transcribe.us-east-2.amazonaws.com" + "url": "https://transcribe.us-iso-east-1.c2s.ic.gov" } }, "params": { "UseFIPS": false, - "Region": "us-east-2", - "UseDualStack": false + "UseDualStack": false, + "Region": "us-iso-east-1" } }, { - "documentation": "For region cn-northwest-1 with FIPS enabled and DualStack enabled", + "documentation": "For region us-iso-east-1 with FIPS enabled and DualStack disabled", "expect": { "endpoint": { - "url": "https://transcribe-fips.cn-northwest-1.api.amazonwebservices.com.cn" + "url": "https://transcribe-fips.us-iso-east-1.c2s.ic.gov" } }, "params": { "UseFIPS": true, - "Region": "cn-northwest-1", - "UseDualStack": true + "UseDualStack": false, + "Region": "us-iso-east-1" } }, { - "documentation": "For region cn-northwest-1 with FIPS enabled and DualStack disabled", + "documentation": "For region us-isob-east-1 with FIPS enabled and DualStack disabled", "expect": { "endpoint": { - "url": "https://transcribe-fips.cn-northwest-1.amazonaws.com.cn" + "url": "https://transcribe-fips.us-isob-east-1.sc2s.sgov.gov" } }, "params": { "UseFIPS": true, - "Region": "cn-northwest-1", - "UseDualStack": false + "UseDualStack": false, + "Region": "us-isob-east-1" } }, { - "documentation": "For region cn-northwest-1 with FIPS disabled and DualStack enabled", + "documentation": "For region us-isob-east-1 with FIPS disabled and DualStack disabled", "expect": { "endpoint": { - "url": "https://transcribe.cn-northwest-1.api.amazonwebservices.com.cn" + "url": "https://transcribe.us-isob-east-1.sc2s.sgov.gov" } }, "params": { "UseFIPS": false, - "Region": "cn-northwest-1", - "UseDualStack": true + "UseDualStack": false, + "Region": "us-isob-east-1" } }, { - "documentation": "For region cn-northwest-1 with FIPS disabled and DualStack disabled", + "documentation": "For custom endpoint with region set and fips disabled and dualstack disabled", "expect": { "endpoint": { - "url": "https://cn.transcribe.cn-northwest-1.amazonaws.com.cn" + "url": "https://example.com" } }, "params": { "UseFIPS": false, - "Region": "cn-northwest-1", - "UseDualStack": false + "UseDualStack": false, + "Region": "us-east-1", + "Endpoint": "https://example.com" } }, { - "documentation": "For custom endpoint with fips disabled and dualstack disabled", + "documentation": "For custom endpoint with region not set and fips disabled and dualstack disabled", "expect": { "endpoint": { "url": "https://example.com" @@ -1253,7 +556,6 @@ }, "params": { "UseFIPS": false, - "Region": "us-east-1", "UseDualStack": false, "Endpoint": "https://example.com" } @@ -1265,8 +567,8 @@ }, "params": { "UseFIPS": true, - "Region": "us-east-1", "UseDualStack": false, + "Region": "us-east-1", "Endpoint": "https://example.com" } }, @@ -1277,8 +579,8 @@ }, "params": { "UseFIPS": false, - "Region": "us-east-1", "UseDualStack": true, + "Region": "us-east-1", "Endpoint": "https://example.com" } } diff --git a/models/endpoints/endpoints.json b/models/endpoints/endpoints.json index 9ee83f5e807..9e48747dc1f 100644 --- a/models/endpoints/endpoints.json +++ b/models/endpoints/endpoints.json @@ -10590,6 +10590,7 @@ "deprecated" : true, "hostname" : "network-firewall-fips.us-west-2.amazonaws.com" }, + "me-central-1" : { }, "me-south-1" : { }, "sa-east-1" : { }, "us-east-1" : { @@ -11461,7 +11462,6 @@ "ap-south-1" : { }, "ap-southeast-1" : { }, "ap-southeast-2" : { }, - "api" : { }, "ca-central-1" : { }, "eu-central-1" : { }, "eu-north-1" : { }, @@ -20416,11 +20416,18 @@ }, "mediaconvert" : { "endpoints" : { - "us-gov-west-1" : { + "fips-us-gov-west-1" : { "credentialScope" : { "region" : "us-gov-west-1" }, + "deprecated" : true, "hostname" : "mediaconvert.us-gov-west-1.amazonaws.com" + }, + "us-gov-west-1" : { + "variants" : [ { + "hostname" : "mediaconvert.us-gov-west-1.amazonaws.com", + "tags" : [ "fips" ] + } ] } } }, diff --git a/service/dynamodb/api.go b/service/dynamodb/api.go index c7727107eb8..a20146ce5e1 100644 --- a/service/dynamodb/api.go +++ b/service/dynamodb/api.go @@ -773,8 +773,13 @@ func (c *DynamoDB) CreateGlobalTableRequest(input *CreateGlobalTableInput) (req // relationship between two or more DynamoDB tables with the same table name // in the provided Regions. // -// This operation only applies to Version 2017.11.29 (https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/globaltables.V1.html) -// of global tables. +// This operation only applies to Version 2017.11.29 (Legacy) (https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/globaltables.V1.html) +// of global tables. We recommend using Version 2019.11.21 (Current) (https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/globaltables.V2.html) +// when creating new global tables, as it provides greater flexibility, higher +// efficiency and consumes less write capacity than 2017.11.29 (Legacy). To +// determine which version you are using, see Determining the version (https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/globaltables.DetermineVersion.html). +// To update existing global tables from version 2017.11.29 (Legacy) to version +// 2019.11.21 (Current), see Updating global tables (https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/V2globaltables_upgrade.html). // // If you want to add a new replica table to a global table, each of the following // conditions must be true: @@ -1363,6 +1368,9 @@ func (c *DynamoDB) DeleteTableRequest(input *DeleteTableInput) (req *request.Req // If the specified table does not exist, DynamoDB returns a ResourceNotFoundException. // If table is already in the DELETING state, no error is returned. // +// This operation only applies to Version 2019.11.21 (Current) (https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/globaltables.V2.html) +// of global tables. +// // DynamoDB might continue to accept data read and write operations, such as // GetItem and PutItem, on a table in the DELETING state until the table deletion // is complete. @@ -1708,7 +1716,7 @@ func (c *DynamoDB) DescribeContributorInsightsRequest(input *DescribeContributor // DescribeContributorInsights API operation for Amazon DynamoDB. // -// Returns information about contributor insights, for a given table or global +// Returns information about contributor insights for a given table or global // secondary index. // // Returns awserr.Error for service API and SDK errors. Use runtime type assertions @@ -1792,7 +1800,10 @@ func (c *DynamoDB) DescribeEndpointsRequest(input *DescribeEndpointsInput) (req // DescribeEndpoints API operation for Amazon DynamoDB. // -// Returns the regional endpoint information. +// Returns the regional endpoint information. This action must be included in +// your VPC endpoint policies, or access to the DescribeEndpoints API will be +// denied. For more information on policy permissions, please see Internetwork +// traffic privacy (https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/inter-network-traffic-privacy.html#inter-network-traffic-DescribeEndpoints). // // Returns awserr.Error for service API and SDK errors. Use runtime type assertions // with awserr.Error's Code and Message methods to get detailed information about @@ -2063,10 +2074,13 @@ func (c *DynamoDB) DescribeGlobalTableRequest(input *DescribeGlobalTableInput) ( // // Returns information about the specified global table. // -// This operation only applies to Version 2017.11.29 (https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/globaltables.V1.html) -// of global tables. If you are using global tables Version 2019.11.21 (https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/globaltables.V2.html) -// you can use DescribeTable (https://docs.aws.amazon.com/amazondynamodb/latest/APIReference/API_DescribeTable.html) -// instead. +// This operation only applies to Version 2017.11.29 (Legacy) (https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/globaltables.V1.html) +// of global tables. We recommend using Version 2019.11.21 (Current) (https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/globaltables.V2.html) +// when creating new global tables, as it provides greater flexibility, higher +// efficiency and consumes less write capacity than 2017.11.29 (Legacy). To +// determine which version you are using, see Determining the version (https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/globaltables.DetermineVersion.html). +// To update existing global tables from version 2017.11.29 (Legacy) to version +// 2019.11.21 (Current), see Updating global tables (https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/V2globaltables_upgrade.html). // // Returns awserr.Error for service API and SDK errors. Use runtime type assertions // with awserr.Error's Code and Message methods to get detailed information about @@ -2175,8 +2189,13 @@ func (c *DynamoDB) DescribeGlobalTableSettingsRequest(input *DescribeGlobalTable // // Describes Region-specific settings for a global table. // -// This operation only applies to Version 2017.11.29 (https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/globaltables.V1.html) -// of global tables. +// This operation only applies to Version 2017.11.29 (Legacy) (https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/globaltables.V1.html) +// of global tables. We recommend using Version 2019.11.21 (Current) (https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/globaltables.V2.html) +// when creating new global tables, as it provides greater flexibility, higher +// efficiency and consumes less write capacity than 2017.11.29 (Legacy). To +// determine which version you are using, see Determining the version (https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/globaltables.DetermineVersion.html). +// To update existing global tables from version 2017.11.29 (Legacy) to version +// 2019.11.21 (Current), see Updating global tables (https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/V2globaltables_upgrade.html). // // Returns awserr.Error for service API and SDK errors. Use runtime type assertions // with awserr.Error's Code and Message methods to get detailed information about @@ -2635,6 +2654,9 @@ func (c *DynamoDB) DescribeTableRequest(input *DescribeTableInput) (req *request // table, when it was created, the primary key schema, and any indexes on the // table. // +// This operation only applies to Version 2019.11.21 (Current) (https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/globaltables.V2.html) +// of global tables. +// // If you issue a DescribeTable request immediately after a CreateTable request, // DynamoDB might return a ResourceNotFoundException. This is because DescribeTable // uses an eventually consistent query, and the metadata for your table might @@ -2724,7 +2746,7 @@ func (c *DynamoDB) DescribeTableReplicaAutoScalingRequest(input *DescribeTableRe // // Describes auto scaling settings across replicas of the global table at once. // -// This operation only applies to Version 2019.11.21 (https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/globaltables.V2.html) +// This operation only applies to Version 2019.11.21 (Current) (https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/globaltables.V2.html) // of global tables. // // Returns awserr.Error for service API and SDK errors. Use runtime type assertions @@ -3192,7 +3214,8 @@ func (c *DynamoDB) ExecuteStatementRequest(input *ExecuteStatementInput) (req *r // A single SELECT statement response can return up to the maximum number of // items (if using the Limit parameter) or a maximum of 1 MB of data (and then // apply any filtering to the results using WHERE clause). If LastEvaluatedKey -// is present in the response, you need to paginate the result set. +// is present in the response, you need to paginate the result set. If NextToken +// is present, you need to paginate the result set and include NextToken. // // Returns awserr.Error for service API and SDK errors. Use runtime type assertions // with awserr.Error's Code and Message methods to get detailed information about @@ -3396,7 +3419,7 @@ func (c *DynamoDB) ExecuteTransactionRequest(input *ExecuteTransactionInput) (re // as DynamoDB is automatically scaling the table. Throughput exceeds the // current capacity for one or more global secondary indexes. DynamoDB is // automatically scaling your index so please try again shortly. This message -// is returned when when writes get throttled on an On-Demand GSI as DynamoDB +// is returned when writes get throttled on an On-Demand GSI as DynamoDB // is automatically scaling the GSI. // // - Validation Error: Code: ValidationError Messages: One or more parameter @@ -3414,6 +3437,47 @@ func (c *DynamoDB) ExecuteTransactionRequest(input *ExecuteTransactionInput) (re // - TransactionInProgressException // The transaction with the given request token is already in progress. // +// Recommended Settings +// +// This is a general recommendation for handling the TransactionInProgressException. +// These settings help ensure that the client retries will trigger completion +// of the ongoing TransactWriteItems request. +// +// - Set clientExecutionTimeout to a value that allows at least one retry +// to be processed after 5 seconds have elapsed since the first attempt for +// the TransactWriteItems operation. +// +// - Set socketTimeout to a value a little lower than the requestTimeout +// setting. +// +// - requestTimeout should be set based on the time taken for the individual +// retries of a single HTTP request for your use case, but setting it to +// 1 second or higher should work well to reduce chances of retries and TransactionInProgressException +// errors. +// +// - Use exponential backoff when retrying and tune backoff if needed. +// +// Assuming default retry policy (https://github.com/aws/aws-sdk-java/blob/fd409dee8ae23fb8953e0bb4dbde65536a7e0514/aws-java-sdk-core/src/main/java/com/amazonaws/retry/PredefinedRetryPolicies.java#L97), +// example timeout settings based on the guidelines above are as follows: +// +// Example timeline: +// +// - 0-1000 first attempt +// +// - 1000-1500 first sleep/delay (default retry policy uses 500 ms as base +// delay for 4xx errors) +// +// - 1500-2500 second attempt +// +// - 2500-3500 second sleep/delay (500 * 2, exponential backoff) +// +// - 3500-4500 third attempt +// +// - 4500-6500 third sleep/delay (500 * 2^2) +// +// - 6500-7500 fourth attempt (this can trigger inline recovery since 5 seconds +// have elapsed since the first attempt reached TC) +// // - IdempotentParameterMismatchException // DynamoDB rejected the request because you retried a request with a different // payload but with an idempotent token that was already used. @@ -4274,8 +4338,13 @@ func (c *DynamoDB) ListGlobalTablesRequest(input *ListGlobalTablesInput) (req *r // // Lists all global tables that have a replica in the specified Region. // -// This operation only applies to Version 2017.11.29 (https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/globaltables.V1.html) -// of global tables. +// This operation only applies to Version 2017.11.29 (Legacy) (https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/globaltables.V1.html) +// of global tables. We recommend using Version 2019.11.21 (Current) (https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/globaltables.V2.html) +// when creating new global tables, as it provides greater flexibility, higher +// efficiency and consumes less write capacity than 2017.11.29 (Legacy). To +// determine which version you are using, see Determining the version (https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/globaltables.DetermineVersion.html). +// To update existing global tables from version 2017.11.29 (Legacy) to version +// 2019.11.21 (Current), see Updating global tables (https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/V2globaltables_upgrade.html). // // Returns awserr.Error for service API and SDK errors. Use runtime type assertions // with awserr.Error's Code and Message methods to get detailed information about @@ -5871,7 +5940,7 @@ func (c *DynamoDB) TransactGetItemsRequest(input *TransactGetItemsInput) (req *r // // - There is a user error, such as an invalid data format. // -// - The aggregate size of the items in the transaction cannot exceed 4 MB. +// - The aggregate size of the items in the transaction exceeded 4 MB. // // Returns awserr.Error for service API and SDK errors. Use runtime type assertions // with awserr.Error's Code and Message methods to get detailed information about @@ -5958,7 +6027,7 @@ func (c *DynamoDB) TransactGetItemsRequest(input *TransactGetItemsInput) (req *r // as DynamoDB is automatically scaling the table. Throughput exceeds the // current capacity for one or more global secondary indexes. DynamoDB is // automatically scaling your index so please try again shortly. This message -// is returned when when writes get throttled on an On-Demand GSI as DynamoDB +// is returned when writes get throttled on an On-Demand GSI as DynamoDB // is automatically scaling the GSI. // // - Validation Error: Code: ValidationError Messages: One or more parameter @@ -6219,7 +6288,7 @@ func (c *DynamoDB) TransactWriteItemsRequest(input *TransactWriteItemsInput) (re // as DynamoDB is automatically scaling the table. Throughput exceeds the // current capacity for one or more global secondary indexes. DynamoDB is // automatically scaling your index so please try again shortly. This message -// is returned when when writes get throttled on an On-Demand GSI as DynamoDB +// is returned when writes get throttled on an On-Demand GSI as DynamoDB // is automatically scaling the GSI. // // - Validation Error: Code: ValidationError Messages: One or more parameter @@ -6237,6 +6306,47 @@ func (c *DynamoDB) TransactWriteItemsRequest(input *TransactWriteItemsInput) (re // - TransactionInProgressException // The transaction with the given request token is already in progress. // +// Recommended Settings +// +// This is a general recommendation for handling the TransactionInProgressException. +// These settings help ensure that the client retries will trigger completion +// of the ongoing TransactWriteItems request. +// +// - Set clientExecutionTimeout to a value that allows at least one retry +// to be processed after 5 seconds have elapsed since the first attempt for +// the TransactWriteItems operation. +// +// - Set socketTimeout to a value a little lower than the requestTimeout +// setting. +// +// - requestTimeout should be set based on the time taken for the individual +// retries of a single HTTP request for your use case, but setting it to +// 1 second or higher should work well to reduce chances of retries and TransactionInProgressException +// errors. +// +// - Use exponential backoff when retrying and tune backoff if needed. +// +// Assuming default retry policy (https://github.com/aws/aws-sdk-java/blob/fd409dee8ae23fb8953e0bb4dbde65536a7e0514/aws-java-sdk-core/src/main/java/com/amazonaws/retry/PredefinedRetryPolicies.java#L97), +// example timeout settings based on the guidelines above are as follows: +// +// Example timeline: +// +// - 0-1000 first attempt +// +// - 1000-1500 first sleep/delay (default retry policy uses 500 ms as base +// delay for 4xx errors) +// +// - 1500-2500 second attempt +// +// - 2500-3500 second sleep/delay (500 * 2, exponential backoff) +// +// - 3500-4500 third attempt +// +// - 4500-6500 third sleep/delay (500 * 2^2) +// +// - 6500-7500 fourth attempt (this can trigger inline recovery since 5 seconds +// have elapsed since the first attempt reached TC) +// // - IdempotentParameterMismatchException // DynamoDB rejected the request because you retried a request with a different // payload but with an idempotent token that was already used. @@ -6700,6 +6810,19 @@ func (c *DynamoDB) UpdateGlobalTableRequest(input *UpdateGlobalTableInput) (req // schema, have DynamoDB Streams enabled, and have the same provisioned and // maximum write capacity units. // +// This operation only applies to Version 2017.11.29 (Legacy) (https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/globaltables.V1.html) +// of global tables. We recommend using Version 2019.11.21 (Current) (https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/globaltables.V2.html) +// when creating new global tables, as it provides greater flexibility, higher +// efficiency and consumes less write capacity than 2017.11.29 (Legacy). To +// determine which version you are using, see Determining the version (https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/globaltables.DetermineVersion.html). +// To update existing global tables from version 2017.11.29 (Legacy) to version +// 2019.11.21 (Current), see Updating global tables (https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/V2globaltables_upgrade.html). +// +// This operation only applies to Version 2017.11.29 (https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/globaltables.V1.html) +// of global tables. If you are using global tables Version 2019.11.21 (https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/globaltables.V2.html) +// you can use DescribeTable (https://docs.aws.amazon.com/amazondynamodb/latest/APIReference/API_DescribeTable.html) +// instead. +// // Although you can use UpdateGlobalTable to add replicas and remove replicas // in a single request, for simplicity we recommend that you issue separate // requests for adding or removing replicas. @@ -6833,6 +6956,14 @@ func (c *DynamoDB) UpdateGlobalTableSettingsRequest(input *UpdateGlobalTableSett // // Updates settings for a global table. // +// This operation only applies to Version 2017.11.29 (Legacy) (https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/globaltables.V1.html) +// of global tables. We recommend using Version 2019.11.21 (Current) (https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/globaltables.V2.html) +// when creating new global tables, as it provides greater flexibility, higher +// efficiency and consumes less write capacity than 2017.11.29 (Legacy). To +// determine which version you are using, see Determining the version (https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/globaltables.DetermineVersion.html). +// To update existing global tables from version 2017.11.29 (Legacy) to version +// 2019.11.21 (Current), see Updating global tables (https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/V2globaltables_upgrade.html). +// // Returns awserr.Error for service API and SDK errors. Use runtime type assertions // with awserr.Error's Code and Message methods to get detailed information about // the error. @@ -7107,6 +7238,9 @@ func (c *DynamoDB) UpdateTableRequest(input *UpdateTableInput) (req *request.Req // Modifies the provisioned throughput settings, global secondary indexes, or // DynamoDB Streams settings for a given table. // +// This operation only applies to Version 2019.11.21 (Current) (https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/globaltables.V2.html) +// of global tables. +// // You can only perform one of the following operations at once: // // - Modify the provisioned throughput settings of the table. @@ -7226,7 +7360,7 @@ func (c *DynamoDB) UpdateTableReplicaAutoScalingRequest(input *UpdateTableReplic // // Updates auto scaling settings on your global tables at once. // -// This operation only applies to Version 2019.11.21 (https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/globaltables.V2.html) +// This operation only applies to Version 2019.11.21 (Current) (https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/globaltables.V2.html) // of global tables. // // Returns awserr.Error for service API and SDK errors. Use runtime type assertions @@ -9717,14 +9851,20 @@ type ConditionCheck struct { _ struct{} `type:"structure"` // A condition that must be satisfied in order for a conditional update to succeed. + // For more information, see Condition expressions (https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/Expressions.ConditionExpressions.html) + // in the Amazon DynamoDB Developer Guide. // // ConditionExpression is a required field ConditionExpression *string `type:"string" required:"true"` - // One or more substitution tokens for attribute names in an expression. + // One or more substitution tokens for attribute names in an expression. For + // more information, see Expression attribute names (https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/Expressions.ExpressionAttributeNames.html) + // in the Amazon DynamoDB Developer Guide. ExpressionAttributeNames map[string]*string `type:"map"` - // One or more values that can be substituted in an expression. + // One or more values that can be substituted in an expression. For more information, + // see Condition expressions (https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/Expressions.ConditionExpressions.html) + // in the Amazon DynamoDB Developer Guide. ExpressionAttributeValues map[string]*AttributeValue `type:"map"` // The primary key of the item to be checked. Each element consists of an attribute @@ -11332,7 +11472,7 @@ type DeleteItemInput struct { // A map of attribute names to AttributeValue objects, representing the primary // key of the item to delete. // - // For the primary key, you must provide all of the attributes. For example, + // For the primary key, you must provide all of the key attributes. For example, // with a simple primary key, you only need to provide a value for the partition // key. For a composite primary key, you must provide values for both the partition // key and the sort key. @@ -11493,7 +11633,7 @@ type DeleteItemOutput struct { // includes the total provisioned throughput consumed, along with statistics // for the table and any indexes involved in the operation. ConsumedCapacity // is only returned if the ReturnConsumedCapacity parameter was specified. For - // more information, see Provisioned Mode (https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/ProvisionedThroughputIntro.html) + // more information, see Provisioned Throughput (https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/ProvisionedThroughputIntro.html) // in the Amazon DynamoDB Developer Guide. ConsumedCapacity *ConsumedCapacity `type:"structure"` @@ -14647,7 +14787,7 @@ type GetItemOutput struct { // the total provisioned throughput consumed, along with statistics for the // table and any indexes involved in the operation. ConsumedCapacity is only // returned if the ReturnConsumedCapacity parameter was specified. For more - // information, see Read/Write Capacity Mode (https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/ProvisionedThroughputIntro.html) + // information, see Provisioned Throughput (https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/ProvisionedThroughput.html#ItemSizeCalculations.Reads) // in the Amazon DynamoDB Developer Guide. ConsumedCapacity *ConsumedCapacity `type:"structure"` @@ -19024,7 +19164,7 @@ type PutItemOutput struct { // the total provisioned throughput consumed, along with statistics for the // table and any indexes involved in the operation. ConsumedCapacity is only // returned if the ReturnConsumedCapacity parameter was specified. For more - // information, see Read/Write Capacity Mode (https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/ProvisionedThroughputIntro.html) + // information, see Provisioned Throughput (https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/ProvisionedThroughputIntro.html) // in the Amazon DynamoDB Developer Guide. ConsumedCapacity *ConsumedCapacity `type:"structure"` @@ -19382,7 +19522,8 @@ type QueryInput struct { // to specifying ALL_ATTRIBUTES. // // * COUNT - Returns the number of matching items, rather than the matching - // items themselves. + // items themselves. Note that this uses the same quantity of read capacity + // units as getting the items, and is subject to the same item size calculations. // // * SPECIFIC_ATTRIBUTES - Returns only the attributes listed in ProjectionExpression. // This return value is equivalent to specifying ProjectionExpression without @@ -21938,7 +22079,8 @@ type ScanInput struct { // to specifying ALL_ATTRIBUTES. // // * COUNT - Returns the number of matching items, rather than the matching - // items themselves. + // items themselves. Note that this uses the same quantity of read capacity + // units as getting the items, and is subject to the same item size calculations. // // * SPECIFIC_ATTRIBUTES - Returns only the attributes listed in ProjectionExpression. // This return value is equivalent to specifying ProjectionExpression without @@ -22144,7 +22286,7 @@ type ScanOutput struct { // the total provisioned throughput consumed, along with statistics for the // table and any indexes involved in the operation. ConsumedCapacity is only // returned if the ReturnConsumedCapacity parameter was specified. For more - // information, see Provisioned Throughput (https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/ProvisionedThroughputIntro.html) + // information, see Provisioned Throughput (https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/ProvisionedThroughput.html#ItemSizeCalculations.Reads) // in the Amazon DynamoDB Developer Guide. ConsumedCapacity *ConsumedCapacity `type:"structure"` @@ -23870,7 +24012,7 @@ type TransactWriteItemsInput struct { // // Although multiple identical calls using the same client request token produce // the same result on the server (no side effects), the responses to the calls - // might not be the same. If the ReturnConsumedCapacity> parameter is set, then + // might not be the same. If the ReturnConsumedCapacity parameter is set, then // the initial TransactWriteItems call returns the amount of write capacity // units consumed in making the changes. Subsequent TransactWriteItems calls // with the same client token return the number of read capacity units consumed @@ -24103,7 +24245,7 @@ func (s *TransactWriteItemsOutput) SetItemCollectionMetrics(v map[string][]*Item // as DynamoDB is automatically scaling the table. Throughput exceeds the // current capacity for one or more global secondary indexes. DynamoDB is // automatically scaling your index so please try again shortly. This message -// is returned when when writes get throttled on an On-Demand GSI as DynamoDB +// is returned when writes get throttled on an On-Demand GSI as DynamoDB // is automatically scaling the GSI. // // - Validation Error: Code: ValidationError Messages: One or more parameter @@ -24248,6 +24390,47 @@ func (s *TransactionConflictException) RequestID() string { } // The transaction with the given request token is already in progress. +// +// # Recommended Settings +// +// This is a general recommendation for handling the TransactionInProgressException. +// These settings help ensure that the client retries will trigger completion +// of the ongoing TransactWriteItems request. +// +// - Set clientExecutionTimeout to a value that allows at least one retry +// to be processed after 5 seconds have elapsed since the first attempt for +// the TransactWriteItems operation. +// +// - Set socketTimeout to a value a little lower than the requestTimeout +// setting. +// +// - requestTimeout should be set based on the time taken for the individual +// retries of a single HTTP request for your use case, but setting it to +// 1 second or higher should work well to reduce chances of retries and TransactionInProgressException +// errors. +// +// - Use exponential backoff when retrying and tune backoff if needed. +// +// Assuming default retry policy (https://github.com/aws/aws-sdk-java/blob/fd409dee8ae23fb8953e0bb4dbde65536a7e0514/aws-java-sdk-core/src/main/java/com/amazonaws/retry/PredefinedRetryPolicies.java#L97), +// example timeout settings based on the guidelines above are as follows: +// +// Example timeline: +// +// - 0-1000 first attempt +// +// - 1000-1500 first sleep/delay (default retry policy uses 500 ms as base +// delay for 4xx errors) +// +// - 1500-2500 second attempt +// +// - 2500-3500 second sleep/delay (500 * 2, exponential backoff) +// +// - 3500-4500 third attempt +// +// - 4500-6500 third sleep/delay (500 * 2^2) +// +// - 6500-7500 fourth attempt (this can trigger inline recovery since 5 seconds +// have elapsed since the first attempt reached TC) type TransactionInProgressException struct { _ struct{} `type:"structure"` RespMetadata protocol.ResponseMetadata `json:"-" xml:"-"` @@ -25224,7 +25407,8 @@ type UpdateItemInput struct { ReturnItemCollectionMetrics *string `type:"string" enum:"ReturnItemCollectionMetrics"` // Use ReturnValues if you want to get the item attributes as they appear before - // or after they are updated. For UpdateItem, the valid values are: + // or after they are successfully updated. For UpdateItem, the valid values + // are: // // * NONE - If ReturnValues is not specified, or if its value is NONE, then // nothing is returned. (This setting is the default for ReturnValues.) @@ -25429,15 +25613,16 @@ type UpdateItemOutput struct { // A map of attribute values as they appear before or after the UpdateItem operation, // as determined by the ReturnValues parameter. // - // The Attributes map is only present if ReturnValues was specified as something - // other than NONE in the request. Each element represents one attribute. + // The Attributes map is only present if the update was successful and ReturnValues + // was specified as something other than NONE in the request. Each element represents + // one attribute. Attributes map[string]*AttributeValue `type:"map"` // The capacity units consumed by the UpdateItem operation. The data returned // includes the total provisioned throughput consumed, along with statistics // for the table and any indexes involved in the operation. ConsumedCapacity // is only returned if the ReturnConsumedCapacity parameter was specified. For - // more information, see Provisioned Throughput (https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/ProvisionedThroughputIntro.html) + // more information, see Provisioned Throughput (https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/ProvisionedThroughput.html#ItemSizeCalculations.Reads) // in the Amazon DynamoDB Developer Guide. ConsumedCapacity *ConsumedCapacity `type:"structure"` @@ -25648,7 +25833,7 @@ type UpdateTableInput struct { // A list of replica update actions (create, delete, or update) for the table. // - // This property only applies to Version 2019.11.21 (https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/globaltables.V2.html) + // This property only applies to Version 2019.11.21 (Current) (https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/globaltables.V2.html) // of global tables. ReplicaUpdates []*ReplicationGroupUpdate `min:"1" type:"list"` diff --git a/service/dynamodb/errors.go b/service/dynamodb/errors.go index 9f7baf88f6b..d3af3262eb6 100644 --- a/service/dynamodb/errors.go +++ b/service/dynamodb/errors.go @@ -283,7 +283,7 @@ const ( // as DynamoDB is automatically scaling the table. Throughput exceeds the // current capacity for one or more global secondary indexes. DynamoDB is // automatically scaling your index so please try again shortly. This message - // is returned when when writes get throttled on an On-Demand GSI as DynamoDB + // is returned when writes get throttled on an On-Demand GSI as DynamoDB // is automatically scaling the GSI. // // * Validation Error: Code: ValidationError Messages: One or more parameter @@ -309,6 +309,47 @@ const ( // "TransactionInProgressException". // // The transaction with the given request token is already in progress. + // + // Recommended Settings + // + // This is a general recommendation for handling the TransactionInProgressException. + // These settings help ensure that the client retries will trigger completion + // of the ongoing TransactWriteItems request. + // + // * Set clientExecutionTimeout to a value that allows at least one retry + // to be processed after 5 seconds have elapsed since the first attempt for + // the TransactWriteItems operation. + // + // * Set socketTimeout to a value a little lower than the requestTimeout + // setting. + // + // * requestTimeout should be set based on the time taken for the individual + // retries of a single HTTP request for your use case, but setting it to + // 1 second or higher should work well to reduce chances of retries and TransactionInProgressException + // errors. + // + // * Use exponential backoff when retrying and tune backoff if needed. + // + // Assuming default retry policy (https://github.com/aws/aws-sdk-java/blob/fd409dee8ae23fb8953e0bb4dbde65536a7e0514/aws-java-sdk-core/src/main/java/com/amazonaws/retry/PredefinedRetryPolicies.java#L97), + // example timeout settings based on the guidelines above are as follows: + // + // Example timeline: + // + // * 0-1000 first attempt + // + // * 1000-1500 first sleep/delay (default retry policy uses 500 ms as base + // delay for 4xx errors) + // + // * 1500-2500 second attempt + // + // * 2500-3500 second sleep/delay (500 * 2, exponential backoff) + // + // * 3500-4500 third attempt + // + // * 4500-6500 third sleep/delay (500 * 2^2) + // + // * 6500-7500 fourth attempt (this can trigger inline recovery since 5 seconds + // have elapsed since the first attempt reached TC) ErrCodeTransactionInProgressException = "TransactionInProgressException" ) diff --git a/service/ec2/api.go b/service/ec2/api.go index bbab3bc424b..ae8e2b716ec 100644 --- a/service/ec2/api.go +++ b/service/ec2/api.go @@ -67965,9 +67965,7 @@ type CreateCustomerGatewayInput struct { // For devices that support BGP, the customer gateway's BGP ASN. // // Default: 65000 - // - // BgpAsn is a required field - BgpAsn *int64 `type:"integer" required:"true"` + BgpAsn *int64 `type:"integer"` // The Amazon Resource Name (ARN) for the customer gateway certificate. CertificateArn *string `type:"string"` @@ -68021,9 +68019,6 @@ func (s CreateCustomerGatewayInput) GoString() string { // Validate inspects the fields of the type to determine if they are valid. func (s *CreateCustomerGatewayInput) Validate() error { invalidParams := request.ErrInvalidParams{Context: "CreateCustomerGatewayInput"} - if s.BgpAsn == nil { - invalidParams.Add(request.NewErrParamRequired("BgpAsn")) - } if s.Type == nil { invalidParams.Add(request.NewErrParamRequired("Type")) } @@ -127124,7 +127119,14 @@ type Instance struct { // Any block device mapping entries for the instance. BlockDeviceMappings []*InstanceBlockDeviceMapping `locationName:"blockDeviceMapping" locationNameList:"item" type:"list"` - // The boot mode of the instance. For more information, see Boot modes (https://docs.aws.amazon.com/AWSEC2/latest/UserGuide/ami-boot.html) + // The boot mode that was specified by the AMI. If the value is uefi-preferred, + // the AMI supports both UEFI and Legacy BIOS. The currentInstanceBootMode parameter + // is the boot mode that is used to boot the instance at launch or start. + // + // The operating system contained in the AMI must be configured to support the + // specified boot mode. + // + // For more information, see Boot modes (https://docs.aws.amazon.com/AWSEC2/latest/UserGuide/ami-boot.html) // in the Amazon EC2 User Guide. BootMode *string `locationName:"bootMode" type:"string" enum:"BootModeValues"` @@ -127140,6 +127142,11 @@ type Instance struct { // The CPU options for the instance. CpuOptions *CpuOptions `locationName:"cpuOptions" type:"structure"` + // The boot mode that is used to boot the instance at launch or start. For more + // information, see Boot modes (https://docs.aws.amazon.com/AWSEC2/latest/UserGuide/ami-boot.html) + // in the Amazon EC2 User Guide. + CurrentInstanceBootMode *string `locationName:"currentInstanceBootMode" type:"string" enum:"InstanceBootModeValues"` + // Indicates whether the instance is optimized for Amazon EBS I/O. This optimization // provides dedicated throughput to Amazon EBS and an optimized configuration // stack to provide optimal I/O performance. This optimization isn't available @@ -127378,6 +127385,12 @@ func (s *Instance) SetCpuOptions(v *CpuOptions) *Instance { return s } +// SetCurrentInstanceBootMode sets the CurrentInstanceBootMode field's value. +func (s *Instance) SetCurrentInstanceBootMode(v string) *Instance { + s.CurrentInstanceBootMode = &v + return s +} + // SetEbsOptimized sets the EbsOptimized field's value. func (s *Instance) SetEbsOptimized(v bool) *Instance { s.EbsOptimized = &v @@ -154421,7 +154434,13 @@ type RegisterImageInput struct { // in the Amazon EC2 User Guide. BlockDeviceMappings []*BlockDeviceMapping `locationName:"BlockDeviceMapping" locationNameList:"BlockDeviceMapping" type:"list"` - // The boot mode of the AMI. For more information, see Boot modes (https://docs.aws.amazon.com/AWSEC2/latest/UserGuide/ami-boot.html) + // The boot mode of the AMI. A value of uefi-preferred indicates that the AMI + // supports both UEFI and Legacy BIOS. + // + // The operating system contained in the AMI must be configured to support the + // specified boot mode. + // + // For more information, see Boot modes (https://docs.aws.amazon.com/AWSEC2/latest/UserGuide/ami-boot.html) // in the Amazon EC2 User Guide. BootMode *string `type:"string" enum:"BootModeValues"` @@ -180256,6 +180275,9 @@ const ( // BootModeValuesUefi is a BootModeValues enum value BootModeValuesUefi = "uefi" + + // BootModeValuesUefiPreferred is a BootModeValues enum value + BootModeValuesUefiPreferred = "uefi-preferred" ) // BootModeValues_Values returns all elements of the BootModeValues enum @@ -180263,6 +180285,7 @@ func BootModeValues_Values() []string { return []string{ BootModeValuesLegacyBios, BootModeValuesUefi, + BootModeValuesUefiPreferred, } } @@ -182162,6 +182185,22 @@ func InstanceAutoRecoveryState_Values() []string { } } +const ( + // InstanceBootModeValuesLegacyBios is a InstanceBootModeValues enum value + InstanceBootModeValuesLegacyBios = "legacy-bios" + + // InstanceBootModeValuesUefi is a InstanceBootModeValues enum value + InstanceBootModeValuesUefi = "uefi" +) + +// InstanceBootModeValues_Values returns all elements of the InstanceBootModeValues enum +func InstanceBootModeValues_Values() []string { + return []string{ + InstanceBootModeValuesLegacyBios, + InstanceBootModeValuesUefi, + } +} + const ( // InstanceEventWindowStateCreating is a InstanceEventWindowState enum value InstanceEventWindowStateCreating = "creating" diff --git a/service/macie2/api.go b/service/macie2/api.go index f67d77a062e..f956cf9766c 100644 --- a/service/macie2/api.go +++ b/service/macie2/api.go @@ -9859,10 +9859,10 @@ func (s *BucketCountByEffectivePermission) SetUnknown(v int64) *BucketCountByEff return s } -// Provides information about the number of S3 buckets that use certain types -// of server-side encryption by default or don't encrypt new objects by default. -// For detailed information about these settings, see Setting default server-side -// encryption behavior for Amazon S3 buckets (https://docs.aws.amazon.com/AmazonS3/latest/userguide/bucket-encryption.html) +// Provides information about the number of S3 buckets whose settings do or +// don't specify default server-side encryption behavior for objects that are +// added to the buckets. For detailed information about these settings, see +// Setting default server-side encryption behavior for Amazon S3 buckets (https://docs.aws.amazon.com/AmazonS3/latest/userguide/bucket-encryption.html) // in the Amazon Simple Storage Service User Guide. type BucketCountByEncryptionType struct { _ struct{} `type:"structure"` @@ -9979,7 +9979,7 @@ func (s *BucketCountBySharedAccessType) SetUnknown(v int64) *BucketCountByShared } // Provides information about the number of S3 buckets whose bucket policies -// do or don't require server-side encryption of objects when objects are uploaded +// do or don't require server-side encryption of objects when objects are added // to the buckets. type BucketCountPolicyAllowsUnencryptedObjectUploads struct { _ struct{} `type:"structure"` @@ -14716,15 +14716,15 @@ type GetBucketStatisticsOutput struct { // due to a combination of permissions settings for each bucket. BucketCountByEffectivePermission *BucketCountByEffectivePermission `locationName:"bucketCountByEffectivePermission" type:"structure"` - // Provides information about the number of S3 buckets that use certain types - // of server-side encryption by default or don't encrypt new objects by default. - // For detailed information about these settings, see Setting default server-side - // encryption behavior for Amazon S3 buckets (https://docs.aws.amazon.com/AmazonS3/latest/userguide/bucket-encryption.html) + // Provides information about the number of S3 buckets whose settings do or + // don't specify default server-side encryption behavior for objects that are + // added to the buckets. For detailed information about these settings, see + // Setting default server-side encryption behavior for Amazon S3 buckets (https://docs.aws.amazon.com/AmazonS3/latest/userguide/bucket-encryption.html) // in the Amazon Simple Storage Service User Guide. BucketCountByEncryptionType *BucketCountByEncryptionType `locationName:"bucketCountByEncryptionType" type:"structure"` // Provides information about the number of S3 buckets whose bucket policies - // do or don't require server-side encryption of objects when objects are uploaded + // do or don't require server-side encryption of objects when objects are added // to the buckets. BucketCountByObjectEncryptionRequirement *BucketCountPolicyAllowsUnencryptedObjectUploads `locationName:"bucketCountByObjectEncryptionRequirement" type:"structure"` @@ -20168,8 +20168,8 @@ type S3Bucket struct { CreatedAt *time.Time `locationName:"createdAt" type:"timestamp" timestampFormat:"iso8601"` - // Provides information about the server-side encryption settings for an S3 - // bucket or S3 object. + // Provides information about the default server-side encryption settings for + // an S3 bucket or the encryption settings for an S3 object. DefaultServerSideEncryption *ServerSideEncryption `locationName:"defaultServerSideEncryption" type:"structure"` Name *string `locationName:"name" type:"string"` @@ -20743,8 +20743,8 @@ type S3Object struct { PublicAccess *bool `locationName:"publicAccess" type:"boolean"` - // Provides information about the server-side encryption settings for an S3 - // bucket or S3 object. + // Provides information about the default server-side encryption settings for + // an S3 bucket or the encryption settings for an S3 object. ServerSideEncryption *ServerSideEncryption `locationName:"serverSideEncryption" type:"structure"` Size *int64 `locationName:"size" type:"long"` @@ -21663,13 +21663,14 @@ func (s *SensitivityInspectionTemplatesEntry) SetName(v string) *SensitivityInsp return s } -// Provides information about the server-side encryption settings for an S3 -// bucket or S3 object. +// Provides information about the default server-side encryption settings for +// an S3 bucket or the encryption settings for an S3 object. type ServerSideEncryption struct { _ struct{} `type:"structure"` - // The type of server-side encryption that's used to encrypt an S3 object or - // objects in an S3 bucket. Possible values are: + // The server-side encryption algorithm that was used to encrypt an S3 object + // or is used by default to encrypt objects that are added to an S3 bucket. + // Possible values are: EncryptionType *string `locationName:"encryptionType" type:"string" enum:"EncryptionType"` KmsMasterKeyId *string `locationName:"kmsMasterKeyId" type:"string"` @@ -24867,8 +24868,9 @@ func EffectivePermission_Values() []string { } } -// The type of server-side encryption that's used to encrypt an S3 object or -// objects in an S3 bucket. Possible values are: +// The server-side encryption algorithm that was used to encrypt an S3 object +// or is used by default to encrypt objects that are added to an S3 bucket. +// Possible values are: const ( // EncryptionTypeNone is a EncryptionType enum value EncryptionTypeNone = "NONE" diff --git a/service/mediaconvert/api.go b/service/mediaconvert/api.go index 674272be66b..ec6342000af 100644 --- a/service/mediaconvert/api.go +++ b/service/mediaconvert/api.go @@ -5206,9 +5206,9 @@ func (s *BadRequestException) RequestID() string { // QVBR output, with little or no perceptual decrease in quality. Or, use to // increase the video quality of outputs with other rate control modes relative // to the bitrate that you specify. Bandwidth reduction increases further when -// your input is low quality or noisy.Outputs that use this feature incur pro-tier -// pricing.When you include Bandwidth reduction filter, you cannot include the -// Noise reducer preprocessor. +// your input is low quality or noisy. Outputs that use this feature incur pro-tier +// pricing. When you include Bandwidth reduction filter, you cannot include +// the Noise reducer preprocessor. type BandwidthReductionFilter struct { _ struct{} `type:"structure"` @@ -5220,13 +5220,11 @@ type BandwidthReductionFilter struct { Sharpening *string `locationName:"sharpening" type:"string" enum:"BandwidthReductionFilterSharpening"` // Specify the strength of the Bandwidth reduction filter. For most workflows, - // we recommend that you choose Auto. Your output bandwidth will be reduced - // by at least 8 percent with no perceptual decrease in video quality. If your - // output bandwidth isn't constrained, set Filter strength to Low or Medium. - // Low results in minimal to no impact in perceptual quality. For more bandwidth - // reduction, choose High. The filter helps equalize quality between all scenes - // and increases video softness. We recommend that you choose High for low bitrate - // outputs. + // we recommend that you choose Auto to reduce the bandwidth of your output + // with little to no perceptual decrease in video quality. For high quality + // and high bitrate outputs, choose Low. For the most bandwidth reduction, choose + // High. We recommend that you choose High for low bitrate outputs. Note that + // High may incur a slight increase in the softness of your output. Strength *string `locationName:"strength" type:"string" enum:"BandwidthReductionFilterStrength"` } @@ -7415,11 +7413,13 @@ type ColorCorrector struct { // between HDR formats, between SDR formats, from SDR to HDR, and from HDR to // SDR. SDR to HDR conversion doesn't upgrade the dynamic range. The converted // video has an HDR format, but visually appears the same as an unconverted - // output. HDR to SDR conversion uses Elemental tone mapping technology to approximate - // the outcome of manually regrading from HDR to SDR. Select Force P3D65 (SDR) - // to set the output color space metadata to the following: * Color primaries: - // Display P3 * Transfer characteristics: SMPTE 428M * Matrix coefficients: - // BT.709 + // output. HDR to SDR conversion uses tone mapping to approximate the outcome + // of manually regrading from HDR to SDR. When you specify an output color space, + // MediaConvert uses the following color space metadata, which includes color + // primaries, transfer characteristics, and matrix coefficients: * HDR 10: BT.2020, + // PQ, BT.2020 non-constant * HLG 2020: BT.2020, HLG, BT.2020 non-constant * + // P3DCI (Theater): DCIP3, SMPTE 428M, BT.709 * P3D65 (SDR): Display P3, sRGB, + // BT.709 * P3D65 (HDR): Display P3, PQ, BT.709 ColorSpaceConversion *string `locationName:"colorSpaceConversion" type:"string" enum:"ColorSpaceConversion"` // Contrast level. @@ -12273,9 +12273,9 @@ type H264Settings struct { // QVBR output, with little or no perceptual decrease in quality. Or, use to // increase the video quality of outputs with other rate control modes relative // to the bitrate that you specify. Bandwidth reduction increases further when - // your input is low quality or noisy.Outputs that use this feature incur pro-tier - // pricing.When you include Bandwidth reduction filter, you cannot include the - // Noise reducer preprocessor. + // your input is low quality or noisy. Outputs that use this feature incur pro-tier + // pricing. When you include Bandwidth reduction filter, you cannot include + // the Noise reducer preprocessor. BandwidthReductionFilter *BandwidthReductionFilter `locationName:"bandwidthReductionFilter" type:"structure"` // Specify the average bitrate in bits per second. Required for VBR and CBR. @@ -25396,17 +25396,20 @@ type VideoSelector struct { AlphaBehavior *string `locationName:"alphaBehavior" type:"string" enum:"AlphaBehavior"` // If your input video has accurate color space metadata, or if you don't know - // about color space, leave this set to the default value Follow. The service - // will automatically detect your input color space. If your input video has - // metadata indicating the wrong color space, specify the accurate color space - // here. If your input video is HDR 10 and the SMPTE ST 2086 Mastering Display - // Color Volume static metadata isn't present in your video stream, or if that - // metadata is present but not accurate, choose Force HDR 10 here and specify + // about color space: Keep the default value, Follow. MediaConvert will automatically + // detect your input color space. If your input video has metadata indicating + // the wrong color space, or has missing metadata: Specify the accurate color + // space here. If your input video is HDR 10 and the SMPTE ST 2086 Mastering + // Display Color Volume static metadata isn't present in your video stream, + // or if that metadata is present but not accurate: Choose Force HDR 10. Specify // correct values in the input HDR 10 metadata settings. For more information - // about MediaConvert HDR jobs, see https://docs.aws.amazon.com/console/mediaconvert/hdr. - // Select P3D65 (SDR) to set the input color space metadata to the following: - // * Color primaries: Display P3 * Transfer characteristics: SMPTE 428M * Matrix - // coefficients: BT.709 + // about HDR jobs, see https://docs.aws.amazon.com/console/mediaconvert/hdr. + // When you specify an input color space, MediaConvert uses the following color + // space metadata, which includes color primaries, transfer characteristics, + // and matrix coefficients: * HDR 10: BT.2020, PQ, BT.2020 non-constant * HLG + // 2020: BT.2020, HLG, BT.2020 non-constant * P3DCI (Theater): DCIP3, SMPTE + // 428M, BT.709 * P3D65 (SDR): Display P3, sRGB, BT.709 * P3D65 (HDR): Display + // P3, PQ, BT.709 ColorSpace *string `locationName:"colorSpace" type:"string" enum:"ColorSpace"` // There are two sources for color metadata, the input file and the job input @@ -28274,13 +28277,11 @@ func BandwidthReductionFilterSharpening_Values() []string { } // Specify the strength of the Bandwidth reduction filter. For most workflows, -// we recommend that you choose Auto. Your output bandwidth will be reduced -// by at least 8 percent with no perceptual decrease in video quality. If your -// output bandwidth isn't constrained, set Filter strength to Low or Medium. -// Low results in minimal to no impact in perceptual quality. For more bandwidth -// reduction, choose High. The filter helps equalize quality between all scenes -// and increases video softness. We recommend that you choose High for low bitrate -// outputs. +// we recommend that you choose Auto to reduce the bandwidth of your output +// with little to no perceptual decrease in video quality. For high quality +// and high bitrate outputs, choose Low. For the most bandwidth reduction, choose +// High. We recommend that you choose High for low bitrate outputs. Note that +// High may incur a slight increase in the softness of your output. const ( // BandwidthReductionFilterStrengthLow is a BandwidthReductionFilterStrength enum value BandwidthReductionFilterStrengthLow = "LOW" @@ -29407,17 +29408,20 @@ func ColorMetadata_Values() []string { } // If your input video has accurate color space metadata, or if you don't know -// about color space, leave this set to the default value Follow. The service -// will automatically detect your input color space. If your input video has -// metadata indicating the wrong color space, specify the accurate color space -// here. If your input video is HDR 10 and the SMPTE ST 2086 Mastering Display -// Color Volume static metadata isn't present in your video stream, or if that -// metadata is present but not accurate, choose Force HDR 10 here and specify +// about color space: Keep the default value, Follow. MediaConvert will automatically +// detect your input color space. If your input video has metadata indicating +// the wrong color space, or has missing metadata: Specify the accurate color +// space here. If your input video is HDR 10 and the SMPTE ST 2086 Mastering +// Display Color Volume static metadata isn't present in your video stream, +// or if that metadata is present but not accurate: Choose Force HDR 10. Specify // correct values in the input HDR 10 metadata settings. For more information -// about MediaConvert HDR jobs, see https://docs.aws.amazon.com/console/mediaconvert/hdr. -// Select P3D65 (SDR) to set the input color space metadata to the following: -// * Color primaries: Display P3 * Transfer characteristics: SMPTE 428M * Matrix -// coefficients: BT.709 +// about HDR jobs, see https://docs.aws.amazon.com/console/mediaconvert/hdr. +// When you specify an input color space, MediaConvert uses the following color +// space metadata, which includes color primaries, transfer characteristics, +// and matrix coefficients: * HDR 10: BT.2020, PQ, BT.2020 non-constant * HLG +// 2020: BT.2020, HLG, BT.2020 non-constant * P3DCI (Theater): DCIP3, SMPTE +// 428M, BT.709 * P3D65 (SDR): Display P3, sRGB, BT.709 * P3D65 (HDR): Display +// P3, PQ, BT.709 const ( // ColorSpaceFollow is a ColorSpace enum value ColorSpaceFollow = "FOLLOW" @@ -29439,6 +29443,9 @@ const ( // ColorSpaceP3d65Sdr is a ColorSpace enum value ColorSpaceP3d65Sdr = "P3D65_SDR" + + // ColorSpaceP3d65Hdr is a ColorSpace enum value + ColorSpaceP3d65Hdr = "P3D65_HDR" ) // ColorSpace_Values returns all elements of the ColorSpace enum @@ -29451,6 +29458,7 @@ func ColorSpace_Values() []string { ColorSpaceHlg2020, ColorSpaceP3dci, ColorSpaceP3d65Sdr, + ColorSpaceP3d65Hdr, } } @@ -29458,11 +29466,13 @@ func ColorSpace_Values() []string { // between HDR formats, between SDR formats, from SDR to HDR, and from HDR to // SDR. SDR to HDR conversion doesn't upgrade the dynamic range. The converted // video has an HDR format, but visually appears the same as an unconverted -// output. HDR to SDR conversion uses Elemental tone mapping technology to approximate -// the outcome of manually regrading from HDR to SDR. Select Force P3D65 (SDR) -// to set the output color space metadata to the following: * Color primaries: -// Display P3 * Transfer characteristics: SMPTE 428M * Matrix coefficients: -// BT.709 +// output. HDR to SDR conversion uses tone mapping to approximate the outcome +// of manually regrading from HDR to SDR. When you specify an output color space, +// MediaConvert uses the following color space metadata, which includes color +// primaries, transfer characteristics, and matrix coefficients: * HDR 10: BT.2020, +// PQ, BT.2020 non-constant * HLG 2020: BT.2020, HLG, BT.2020 non-constant * +// P3DCI (Theater): DCIP3, SMPTE 428M, BT.709 * P3D65 (SDR): Display P3, sRGB, +// BT.709 * P3D65 (HDR): Display P3, PQ, BT.709 const ( // ColorSpaceConversionNone is a ColorSpaceConversion enum value ColorSpaceConversionNone = "NONE" @@ -29484,6 +29494,9 @@ const ( // ColorSpaceConversionForceP3d65Sdr is a ColorSpaceConversion enum value ColorSpaceConversionForceP3d65Sdr = "FORCE_P3D65_SDR" + + // ColorSpaceConversionForceP3d65Hdr is a ColorSpaceConversion enum value + ColorSpaceConversionForceP3d65Hdr = "FORCE_P3D65_HDR" ) // ColorSpaceConversion_Values returns all elements of the ColorSpaceConversion enum @@ -29496,6 +29509,7 @@ func ColorSpaceConversion_Values() []string { ColorSpaceConversionForceHlg2020, ColorSpaceConversionForceP3dci, ColorSpaceConversionForceP3d65Sdr, + ColorSpaceConversionForceP3d65Hdr, } } diff --git a/service/medialive/api.go b/service/medialive/api.go index f4c68261cb2..ff3997c5f98 100644 --- a/service/medialive/api.go +++ b/service/medialive/api.go @@ -7868,7 +7868,10 @@ func (s *AvailBlanking) SetState(v string) *AvailBlanking { type AvailConfiguration struct { _ struct{} `type:"structure"` - // Ad avail settings. + // Controls how SCTE-35 messages create cues. Splice Insert mode treats all + // segmentation signals traditionally. With Time Signal APOS mode only Time + // Signal Placement Opportunity and Break messages create segment breaks. With + // ESAM mode, signals are forwarded to an ESAM server for possible update. AvailSettings *AvailSettings `locationName:"availSettings" type:"structure"` } @@ -7918,10 +7921,12 @@ type AvailSettings struct { // Esam Esam *Esam `locationName:"esam" type:"structure"` - // Scte35 Splice Insert + // Typical configuration that applies breaks on splice inserts in addition to + // time signal placement opportunities, breaks, and advertisements. Scte35SpliceInsert *Scte35SpliceInsert `locationName:"scte35SpliceInsert" type:"structure"` - // Scte35 Time Signal Apos + // Atypical configuration that applies segment breaks only on SCTE-35 time signal + // placement opportunities and breaks. Scte35TimeSignalApos *Scte35TimeSignalApos `locationName:"scte35TimeSignalApos" type:"structure"` } @@ -9626,8 +9631,7 @@ func (s *CaptionRectangle) SetWidth(v float64) *CaptionRectangle { return s } -// Output groups for this Live Event. Output groups contain information about -// where streams should be distributed. +// Caption Selector type CaptionSelector struct { _ struct{} `type:"structure"` @@ -17317,7 +17321,8 @@ type HlsAkamaiSettings struct { HttpTransferMode *string `locationName:"httpTransferMode" type:"string" enum:"HlsAkamaiHttpTransferMode"` // Number of retry attempts that will be made before the Live Event is put into - // an error state. + // an error state. Applies only if the CDN destination URI begins with "s3" + // or "mediastore". For other URIs, the value is always 3. NumRetries *int64 `locationName:"numRetries" type:"integer"` // If a streaming output fails, number of seconds to wait until a restart is @@ -17403,7 +17408,8 @@ type HlsBasicPutSettings struct { FilecacheDuration *int64 `locationName:"filecacheDuration" type:"integer"` // Number of retry attempts that will be made before the Live Event is put into - // an error state. + // an error state. Applies only if the CDN destination URI begins with "s3" + // or "mediastore". For other URIs, the value is always 3. NumRetries *int64 `locationName:"numRetries" type:"integer"` // If a streaming output fails, number of seconds to wait until a restart is @@ -18219,7 +18225,8 @@ type HlsMediaStoreSettings struct { MediaStoreStorageClass *string `locationName:"mediaStoreStorageClass" type:"string" enum:"HlsMediaStoreStorageClass"` // Number of retry attempts that will be made before the Live Event is put into - // an error state. + // an error state. Applies only if the CDN destination URI begins with "s3" + // or "mediastore". For other URIs, the value is always 3. NumRetries *int64 `locationName:"numRetries" type:"integer"` // If a streaming output fails, number of seconds to wait until a restart is @@ -18533,7 +18540,8 @@ type HlsWebdavSettings struct { HttpTransferMode *string `locationName:"httpTransferMode" type:"string" enum:"HlsWebdavHttpTransferMode"` // Number of retry attempts that will be made before the Live Event is put into - // an error state. + // an error state. Applies only if the CDN destination URI begins with "s3" + // or "mediastore". For other URIs, the value is always 3. NumRetries *int64 `locationName:"numRetries" type:"integer"` // If a streaming output fails, number of seconds to wait until a restart is @@ -24890,6 +24898,10 @@ type NielsenNaesIiNw struct { // // Sid is a required field Sid *float64 `locationName:"sid" type:"double" required:"true"` + + // Choose the timezone for the time stamps in the watermark. If not provided,the + // timestamps will be in Coordinated Universal Time (UTC) + Timezone *string `locationName:"timezone" type:"string" enum:"NielsenWatermarkTimezones"` } // String returns the string representation. @@ -24941,6 +24953,12 @@ func (s *NielsenNaesIiNw) SetSid(v float64) *NielsenNaesIiNw { return s } +// SetTimezone sets the Timezone field's value. +func (s *NielsenNaesIiNw) SetTimezone(v string) *NielsenNaesIiNw { + s.Timezone = &v + return s +} + // Nielsen Watermarks Settings type NielsenWatermarksSettings struct { _ struct{} `type:"structure"` @@ -27996,7 +28014,8 @@ func (s *Scte35SegmentationDescriptor) SetSubSegmentsExpected(v int64) *Scte35Se return s } -// Scte35 Splice Insert +// Typical configuration that applies breaks on splice inserts in addition to +// time signal placement opportunities, breaks, and advertisements. type Scte35SpliceInsert struct { _ struct{} `type:"structure"` @@ -28124,7 +28143,8 @@ func (s *Scte35SpliceInsertScheduleActionSettings) SetSpliceEventId(v int64) *Sc return s } -// Scte35 Time Signal Apos +// Atypical configuration that applies segment breaks only on SCTE-35 time signal +// placement opportunities and breaks. type Scte35TimeSignalApos struct { _ struct{} `type:"structure"` @@ -36311,6 +36331,55 @@ func NielsenPcmToId3TaggingState_Values() []string { } } +// Nielsen Watermark Timezones +const ( + // NielsenWatermarkTimezonesAmericaPuertoRico is a NielsenWatermarkTimezones enum value + NielsenWatermarkTimezonesAmericaPuertoRico = "AMERICA_PUERTO_RICO" + + // NielsenWatermarkTimezonesUsAlaska is a NielsenWatermarkTimezones enum value + NielsenWatermarkTimezonesUsAlaska = "US_ALASKA" + + // NielsenWatermarkTimezonesUsArizona is a NielsenWatermarkTimezones enum value + NielsenWatermarkTimezonesUsArizona = "US_ARIZONA" + + // NielsenWatermarkTimezonesUsCentral is a NielsenWatermarkTimezones enum value + NielsenWatermarkTimezonesUsCentral = "US_CENTRAL" + + // NielsenWatermarkTimezonesUsEastern is a NielsenWatermarkTimezones enum value + NielsenWatermarkTimezonesUsEastern = "US_EASTERN" + + // NielsenWatermarkTimezonesUsHawaii is a NielsenWatermarkTimezones enum value + NielsenWatermarkTimezonesUsHawaii = "US_HAWAII" + + // NielsenWatermarkTimezonesUsMountain is a NielsenWatermarkTimezones enum value + NielsenWatermarkTimezonesUsMountain = "US_MOUNTAIN" + + // NielsenWatermarkTimezonesUsPacific is a NielsenWatermarkTimezones enum value + NielsenWatermarkTimezonesUsPacific = "US_PACIFIC" + + // NielsenWatermarkTimezonesUsSamoa is a NielsenWatermarkTimezones enum value + NielsenWatermarkTimezonesUsSamoa = "US_SAMOA" + + // NielsenWatermarkTimezonesUtc is a NielsenWatermarkTimezones enum value + NielsenWatermarkTimezonesUtc = "UTC" +) + +// NielsenWatermarkTimezones_Values returns all elements of the NielsenWatermarkTimezones enum +func NielsenWatermarkTimezones_Values() []string { + return []string{ + NielsenWatermarkTimezonesAmericaPuertoRico, + NielsenWatermarkTimezonesUsAlaska, + NielsenWatermarkTimezonesUsArizona, + NielsenWatermarkTimezonesUsCentral, + NielsenWatermarkTimezonesUsEastern, + NielsenWatermarkTimezonesUsHawaii, + NielsenWatermarkTimezonesUsMountain, + NielsenWatermarkTimezonesUsPacific, + NielsenWatermarkTimezonesUsSamoa, + NielsenWatermarkTimezonesUtc, + } +} + // Nielsen Watermarks Cbet Stepaside const ( // NielsenWatermarksCbetStepasideDisabled is a NielsenWatermarksCbetStepaside enum value diff --git a/service/transcribeservice/api.go b/service/transcribeservice/api.go index 3b1b67092ac..79c78a78081 100644 --- a/service/transcribeservice/api.go +++ b/service/transcribeservice/api.go @@ -64,10 +64,10 @@ func (c *TranscribeService) CreateCallAnalyticsCategoryRequest(input *CreateCall // applied retroactively. // // When creating a new category, you can use the InputType parameter to label -// the category as a batch category (POST_CALL) or a streaming category (REAL_TIME). -// Batch categories can only be applied to batch transcriptions and streaming -// categories can only be applied to streaming transcriptions. If you do not -// include InputType, your category is created as a batch category by default. +// the category as a POST_CALL or a REAL_TIME category. POST_CALL categories +// can only be applied to post-call transcriptions and REAL_TIME categories +// can only be applied to real-time transcriptions. If you do not include InputType, +// your category is created as a POST_CALL category by default. // // Call Analytics categories are composed of rules. For each category, you must // create between 1 and 20 rules. Rules can include these parameters: , , , @@ -76,8 +76,8 @@ func (c *TranscribeService) CreateCallAnalyticsCategoryRequest(input *CreateCall // To update an existing category, see . // // To learn more about Call Analytics categories, see Creating categories for -// batch transcriptions (https://docs.aws.amazon.com/transcribe/latest/dg/tca-categories-batch.html) -// and Creating categories for streaming transcriptions (https://docs.aws.amazon.com/transcribe/latest/dg/tca-categories-stream.html). +// post-call transcriptions (https://docs.aws.amazon.com/transcribe/latest/dg/tca-categories-batch.html) +// and Creating categories for real-time transcriptions (https://docs.aws.amazon.com/transcribe/latest/dg/tca-categories-stream.html). // // Returns awserr.Error for service API and SDK errors. Use runtime type assertions // with awserr.Error's Code and Message methods to get detailed information about @@ -280,10 +280,10 @@ func (c *TranscribeService) CreateMedicalVocabularyRequest(input *CreateMedicalV // Creates a new custom medical vocabulary. // // Before creating a new custom medical vocabulary, you must first upload a -// text file that contains your new entries, phrases, and terms into an Amazon -// S3 bucket. Note that this differs from , where you can include a list of -// terms within your request using the Phrases flag; CreateMedicalVocabulary -// does not support the Phrases flag. +// text file that contains your vocabulary table into an Amazon S3 bucket. Note +// that this differs from , where you can include a list of terms within your +// request using the Phrases flag; CreateMedicalVocabulary does not support +// the Phrases flag and only accepts vocabularies in table format. // // Each language has a character set that contains all allowed characters for // that specific language. If you use unsupported characters, your custom vocabulary @@ -3481,14 +3481,16 @@ func (c *TranscribeService) StartCallAnalyticsJobRequest(input *StartCallAnalyti // If you want to apply categories to your Call Analytics job, you must create // them before submitting your job request. Categories cannot be retroactively // applied to a job. To create a new category, use the operation. To learn more -// about Call Analytics categories, see Creating categories for batch transcriptions +// about Call Analytics categories, see Creating categories for post-call transcriptions // (https://docs.aws.amazon.com/transcribe/latest/dg/tca-categories-batch.html) -// and Creating categories for streaming transcriptions (https://docs.aws.amazon.com/transcribe/latest/dg/tca-categories-stream.html). +// and Creating categories for real-time transcriptions (https://docs.aws.amazon.com/transcribe/latest/dg/tca-categories-stream.html). // // To make a StartCallAnalyticsJob request, you must first upload your media // file into an Amazon S3 bucket; you can then specify the Amazon S3 location // of the file using the Media parameter. // +// Note that job queuing is enabled by default for Call Analytics jobs. +// // You must include the following parameters in your StartCallAnalyticsJob request: // // - region: The Amazon Web Services Region where you are making your request. @@ -5362,17 +5364,17 @@ type CreateCallAnalyticsCategoryInput struct { // CategoryName is a required field CategoryName *string `min:"1" type:"string" required:"true"` - // Choose whether you want to create a streaming or a batch category for your - // Call Analytics transcription. + // Choose whether you want to create a real-time or a post-call category for + // your Call Analytics transcription. // - // Specifying POST_CALL assigns your category to batch transcriptions; categories + // Specifying POST_CALL assigns your category to post-call transcriptions; categories // with this input type cannot be applied to streaming (real-time) transcriptions. // // Specifying REAL_TIME assigns your category to streaming transcriptions; categories - // with this input type cannot be applied to batch (post-call) transcriptions. + // with this input type cannot be applied to post-call transcriptions. // - // If you do not include InputType, your category is created as a batch category - // by default. + // If you do not include InputType, your category is created as a post-call + // category by default. InputType *string `type:"string" enum:"InputType"` // Rules define a Call Analytics category. When creating a new category, you @@ -5904,6 +5906,17 @@ func (s *CreateMedicalVocabularyOutput) SetVocabularyState(v string) *CreateMedi type CreateVocabularyFilterInput struct { _ struct{} `type:"structure"` + // The Amazon Resource Name (ARN) of an IAM role that has permissions to access + // the Amazon S3 bucket that contains your input files (in this case, your custom + // vocabulary filter). If the role that you specify doesn’t have the appropriate + // permissions to access the specified Amazon S3 location, your request fails. + // + // IAM role ARNs have the format arn:partition:iam::account:role/role-name-with-path. + // For example: arn:aws:iam::111122223333:role/Admin. + // + // For more information, see IAM ARNs (https://docs.aws.amazon.com/IAM/latest/UserGuide/reference_identifiers.html#identifiers-arns). + DataAccessRoleArn *string `min:"20" type:"string"` + // The language code that represents the language of the entries in your vocabulary // filter. Each custom vocabulary filter must contain terms in only one language. // @@ -5983,6 +5996,9 @@ func (s CreateVocabularyFilterInput) GoString() string { // Validate inspects the fields of the type to determine if they are valid. func (s *CreateVocabularyFilterInput) Validate() error { invalidParams := request.ErrInvalidParams{Context: "CreateVocabularyFilterInput"} + if s.DataAccessRoleArn != nil && len(*s.DataAccessRoleArn) < 20 { + invalidParams.Add(request.NewErrParamMinLen("DataAccessRoleArn", 20)) + } if s.LanguageCode == nil { invalidParams.Add(request.NewErrParamRequired("LanguageCode")) } @@ -6018,6 +6034,12 @@ func (s *CreateVocabularyFilterInput) Validate() error { return nil } +// SetDataAccessRoleArn sets the DataAccessRoleArn field's value. +func (s *CreateVocabularyFilterInput) SetDataAccessRoleArn(v string) *CreateVocabularyFilterInput { + s.DataAccessRoleArn = &v + return s +} + // SetLanguageCode sets the LanguageCode field's value. func (s *CreateVocabularyFilterInput) SetLanguageCode(v string) *CreateVocabularyFilterInput { s.LanguageCode = &v @@ -6103,6 +6125,17 @@ func (s *CreateVocabularyFilterOutput) SetVocabularyFilterName(v string) *Create type CreateVocabularyInput struct { _ struct{} `type:"structure"` + // The Amazon Resource Name (ARN) of an IAM role that has permissions to access + // the Amazon S3 bucket that contains your input files (in this case, your custom + // vocabulary). If the role that you specify doesn’t have the appropriate + // permissions to access the specified Amazon S3 location, your request fails. + // + // IAM role ARNs have the format arn:partition:iam::account:role/role-name-with-path. + // For example: arn:aws:iam::111122223333:role/Admin. + // + // For more information, see IAM ARNs (https://docs.aws.amazon.com/IAM/latest/UserGuide/reference_identifiers.html#identifiers-arns). + DataAccessRoleArn *string `min:"20" type:"string"` + // The language code that represents the language of the entries in your custom // vocabulary. Each custom vocabulary must contain terms in only one language. // @@ -6182,6 +6215,9 @@ func (s CreateVocabularyInput) GoString() string { // Validate inspects the fields of the type to determine if they are valid. func (s *CreateVocabularyInput) Validate() error { invalidParams := request.ErrInvalidParams{Context: "CreateVocabularyInput"} + if s.DataAccessRoleArn != nil && len(*s.DataAccessRoleArn) < 20 { + invalidParams.Add(request.NewErrParamMinLen("DataAccessRoleArn", 20)) + } if s.LanguageCode == nil { invalidParams.Add(request.NewErrParamRequired("LanguageCode")) } @@ -6214,6 +6250,12 @@ func (s *CreateVocabularyInput) Validate() error { return nil } +// SetDataAccessRoleArn sets the DataAccessRoleArn field's value. +func (s *CreateVocabularyInput) SetDataAccessRoleArn(v string) *CreateVocabularyInput { + s.DataAccessRoleArn = &v + return s +} + // SetLanguageCode sets the LanguageCode field's value. func (s *CreateVocabularyInput) SetLanguageCode(v string) *CreateVocabularyInput { s.LanguageCode = &v @@ -7862,7 +7904,7 @@ func (s *InternalFailureException) RequestID() string { // // - A lack of interruptions // -// See Rule criteria for batch categories (https://docs.aws.amazon.com/transcribe/latest/dg/tca-categories-batch.html#tca-rules-batch) +// See Rule criteria for post-call categories (https://docs.aws.amazon.com/transcribe/latest/dg/tca-categories-batch.html#tca-rules-batch) // for usage examples. type InterruptionFilter struct { _ struct{} `type:"structure"` @@ -7952,8 +7994,6 @@ type JobExecutionSettings struct { // transcription job requests exceed the concurrent request limit, you get a // LimitExceededException error. // - // Note that job queuing is enabled by default for Call Analytics jobs. - // // If you include AllowDeferredExecution in your request, you must also include // DataAccessRoleArn. AllowDeferredExecution *bool `type:"boolean"` @@ -8175,11 +8215,23 @@ func (s *LanguageIdSettings) SetVocabularyName(v string) *LanguageIdSettings { return s } -// Provides information about a custom language model, including the base model -// name, when the model was created, the location of the files used to train -// the model, when the model was last modified, the name you chose for the model, -// its language, its processing state, and if there is an upgrade available -// for the base model. +// Provides information about a custom language model, including: +// +// - The base model name +// +// - When the model was created +// +// - The location of the files used to train the model +// +// - When the model was last modified +// +// - The name you chose for the model +// +// - The model's language +// +// - The model's processing state +// +// - Any available upgrades for the base model type LanguageModel struct { _ struct{} `type:"structure"` @@ -10232,7 +10284,7 @@ func (s *ModelSettings) SetLanguageModelName(v string) *ModelSettings { // // - The presence of speech at specified periods throughout the call // -// See Rule criteria for batch categories (https://docs.aws.amazon.com/transcribe/latest/dg/tca-categories-batch.html#tca-rules-batch) +// See Rule criteria for post-call categories (https://docs.aws.amazon.com/transcribe/latest/dg/tca-categories-batch.html#tca-rules-batch) // for usage examples. type NonTalkTimeFilter struct { _ struct{} `type:"structure"` @@ -10447,8 +10499,8 @@ func (s *RelativeTimeRange) SetStartPercentage(v int64) *RelativeTimeRange { // Rules can include these parameters: , , , and . // // To learn more about Call Analytics rules and categories, see Creating categories -// for batch transcriptions (https://docs.aws.amazon.com/transcribe/latest/dg/tca-categories-batch.html) -// and Creating categories for streaming transcriptions (https://docs.aws.amazon.com/transcribe/latest/dg/tca-categories-stream.html). +// for post-call transcriptions (https://docs.aws.amazon.com/transcribe/latest/dg/tca-categories-batch.html) +// and Creating categories for real-time transcriptions (https://docs.aws.amazon.com/transcribe/latest/dg/tca-categories-stream.html). // // To learn more about Call Analytics, see Analyzing call center audio with // Call Analytics (https://docs.aws.amazon.com/transcribe/latest/dg/call-analytics.html). @@ -10551,7 +10603,7 @@ func (s *Rule) SetTranscriptFilter(v *TranscriptFilter) *Rule { // - The presence or absence of a mixed sentiment felt by the customer, the // agent, or both at specified points in the call // -// See Rule criteria for batch categories (https://docs.aws.amazon.com/transcribe/latest/dg/tca-categories-batch.html#tca-rules-batch) +// See Rule criteria for post-call categories (https://docs.aws.amazon.com/transcribe/latest/dg/tca-categories-batch.html#tca-rules-batch) // for usage examples. type SentimentFilter struct { _ struct{} `type:"structure"` @@ -10882,7 +10934,7 @@ type StartCallAnalyticsJobInput struct { // If you specify a KMS key to encrypt your output, you must also specify an // output location using the OutputLocation parameter. // - // Note that the user making the request must have permission to use the specified + // Note that the role making the request must have permission to use the specified // KMS key. OutputEncryptionKMSKeyId *string `min:"1" type:"string"` @@ -11150,7 +11202,7 @@ type StartMedicalTranscriptionJobInput struct { // If you specify a KMS key to encrypt your output, you must also specify an // output location using the OutputLocation parameter. // - // Note that the user making the request must have permission to use the specified + // Note that the role making the request must have permission to use the specified // KMS key. OutputEncryptionKMSKeyId *string `min:"1" type:"string"` @@ -11603,7 +11655,7 @@ type StartTranscriptionJobInput struct { // If you specify a KMS key to encrypt your output, you must also specify an // output location using the OutputLocation parameter. // - // Note that the user making the request must have permission to use the specified + // Note that the role making the request must have permission to use the specified // KMS key. OutputEncryptionKMSKeyId *string `min:"1" type:"string"` @@ -12302,7 +12354,7 @@ func (s *Transcript) SetTranscriptFileUri(v string) *Transcript { // // - Custom words or phrases that occur at a specific time frame // -// See Rule criteria for batch categories (https://docs.aws.amazon.com/transcribe/latest/dg/tca-categories-batch.html#tca-rules-batch) +// See Rule criteria for post-call categories (https://docs.aws.amazon.com/transcribe/latest/dg/tca-categories-batch.html#tca-rules-batch) // and Rule criteria for streaming categories (https://docs.aws.amazon.com/transcribe/latest/dg/tca-categories-stream.html#tca-rules-stream) // for usage examples. type TranscriptFilter struct { @@ -13012,8 +13064,8 @@ type UpdateCallAnalyticsCategoryInput struct { // CategoryName is a required field CategoryName *string `min:"1" type:"string" required:"true"` - // Choose whether you want to update a streaming or a batch Call Analytics category. - // The input type you specify must match the input type specified when the category + // Choose whether you want to update a real-time or a post-call category. The + // input type you specify must match the input type specified when the category // was created. For example, if you created a category with the POST_CALL input // type, you must use POST_CALL as the input type when updating this category. InputType *string `type:"string" enum:"InputType"` @@ -13279,6 +13331,17 @@ func (s *UpdateMedicalVocabularyOutput) SetVocabularyState(v string) *UpdateMedi type UpdateVocabularyFilterInput struct { _ struct{} `type:"structure"` + // The Amazon Resource Name (ARN) of an IAM role that has permissions to access + // the Amazon S3 bucket that contains your input files (in this case, your custom + // vocabulary filter). If the role that you specify doesn’t have the appropriate + // permissions to access the specified Amazon S3 location, your request fails. + // + // IAM role ARNs have the format arn:partition:iam::account:role/role-name-with-path. + // For example: arn:aws:iam::111122223333:role/Admin. + // + // For more information, see IAM ARNs (https://docs.aws.amazon.com/IAM/latest/UserGuide/reference_identifiers.html#identifiers-arns). + DataAccessRoleArn *string `min:"20" type:"string"` + // The Amazon S3 location of the text file that contains your custom vocabulary // filter terms. The URI must be located in the same Amazon Web Services Region // as the resource you're calling. @@ -13332,6 +13395,9 @@ func (s UpdateVocabularyFilterInput) GoString() string { // Validate inspects the fields of the type to determine if they are valid. func (s *UpdateVocabularyFilterInput) Validate() error { invalidParams := request.ErrInvalidParams{Context: "UpdateVocabularyFilterInput"} + if s.DataAccessRoleArn != nil && len(*s.DataAccessRoleArn) < 20 { + invalidParams.Add(request.NewErrParamMinLen("DataAccessRoleArn", 20)) + } if s.VocabularyFilterFileUri != nil && len(*s.VocabularyFilterFileUri) < 1 { invalidParams.Add(request.NewErrParamMinLen("VocabularyFilterFileUri", 1)) } @@ -13351,6 +13417,12 @@ func (s *UpdateVocabularyFilterInput) Validate() error { return nil } +// SetDataAccessRoleArn sets the DataAccessRoleArn field's value. +func (s *UpdateVocabularyFilterInput) SetDataAccessRoleArn(v string) *UpdateVocabularyFilterInput { + s.DataAccessRoleArn = &v + return s +} + // SetVocabularyFilterFileUri sets the VocabularyFilterFileUri field's value. func (s *UpdateVocabularyFilterInput) SetVocabularyFilterFileUri(v string) *UpdateVocabularyFilterInput { s.VocabularyFilterFileUri = &v @@ -13424,6 +13496,17 @@ func (s *UpdateVocabularyFilterOutput) SetVocabularyFilterName(v string) *Update type UpdateVocabularyInput struct { _ struct{} `type:"structure"` + // The Amazon Resource Name (ARN) of an IAM role that has permissions to access + // the Amazon S3 bucket that contains your input files (in this case, your custom + // vocabulary). If the role that you specify doesn’t have the appropriate + // permissions to access the specified Amazon S3 location, your request fails. + // + // IAM role ARNs have the format arn:partition:iam::account:role/role-name-with-path. + // For example: arn:aws:iam::111122223333:role/Admin. + // + // For more information, see IAM ARNs (https://docs.aws.amazon.com/IAM/latest/UserGuide/reference_identifiers.html#identifiers-arns). + DataAccessRoleArn *string `min:"20" type:"string"` + // The language code that represents the language of the entries in the custom // vocabulary you want to update. Each custom vocabulary must contain terms // in only one language. @@ -13493,6 +13576,9 @@ func (s UpdateVocabularyInput) GoString() string { // Validate inspects the fields of the type to determine if they are valid. func (s *UpdateVocabularyInput) Validate() error { invalidParams := request.ErrInvalidParams{Context: "UpdateVocabularyInput"} + if s.DataAccessRoleArn != nil && len(*s.DataAccessRoleArn) < 20 { + invalidParams.Add(request.NewErrParamMinLen("DataAccessRoleArn", 20)) + } if s.LanguageCode == nil { invalidParams.Add(request.NewErrParamRequired("LanguageCode")) } @@ -13512,6 +13598,12 @@ func (s *UpdateVocabularyInput) Validate() error { return nil } +// SetDataAccessRoleArn sets the DataAccessRoleArn field's value. +func (s *UpdateVocabularyInput) SetDataAccessRoleArn(v string) *UpdateVocabularyInput { + s.DataAccessRoleArn = &v + return s +} + // SetLanguageCode sets the LanguageCode field's value. func (s *UpdateVocabularyInput) SetLanguageCode(v string) *UpdateVocabularyInput { s.LanguageCode = &v