From a284d376a6c0f86a0db61810dff7635d41ced960 Mon Sep 17 00:00:00 2001 From: aws-sdk-go-automation <43143561+aws-sdk-go-automation@users.noreply.github.com> Date: Fri, 24 Mar 2023 11:24:15 -0700 Subject: [PATCH] Release v1.44.229 (2023-03-24) (#4773) Release v1.44.229 (2023-03-24) === ### Service Client Updates * `service/comprehend`: Updates service API and documentation * `service/monitoring`: Updates service documentation * Doc-only update to correct alarm actions list * `service/rds`: Updates service API, documentation, waiters, paginators, and examples * Added error code CreateCustomDBEngineVersionFault for when the create custom engine version for Custom engines fails. --- CHANGELOG.md | 10 + aws/endpoints/defaults.go | 25 + aws/version.go | 2 +- models/apis/comprehend/2017-11-27/api-2.json | 6 +- models/apis/comprehend/2017-11-27/docs-2.json | 205 +-- .../2017-11-27/endpoint-tests-1.json | 180 +- models/apis/monitoring/2010-08-01/docs-2.json | 24 +- .../2010-08-01/endpoint-rule-set-1.json | 412 +++-- .../2010-08-01/endpoint-tests-1.json | 1573 +++-------------- models/apis/rds/2014-10-31/api-2.json | 14 +- models/apis/rds/2014-10-31/docs-2.json | 39 +- .../apis/rds/2014-10-31/endpoint-tests-1.json | 146 +- models/endpoints/endpoints.json | 11 + service/cloudwatch/api.go | 145 +- service/comprehend/api.go | 518 +++--- service/comprehend/doc.go | 8 +- service/rds/api.go | 144 +- service/rds/errors.go | 6 + 18 files changed, 1300 insertions(+), 2168 deletions(-) diff --git a/CHANGELOG.md b/CHANGELOG.md index 6e350d64b9e..25b95ba09f9 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -1,3 +1,13 @@ +Release v1.44.229 (2023-03-24) +=== + +### Service Client Updates +* `service/comprehend`: Updates service API and documentation +* `service/monitoring`: Updates service documentation + * Doc-only update to correct alarm actions list +* `service/rds`: Updates service API, documentation, waiters, paginators, and examples + * Added error code CreateCustomDBEngineVersionFault for when the create custom engine version for Custom engines fails. + Release v1.44.228 (2023-03-23) === diff --git a/aws/endpoints/defaults.go b/aws/endpoints/defaults.go index cbd98b571bd..d3789cce827 100644 --- a/aws/endpoints/defaults.go +++ b/aws/endpoints/defaults.go @@ -14341,6 +14341,31 @@ var awsPartition = partition{ }: endpoint{}, }, }, + "ivsrealtime": service{ + Endpoints: serviceEndpoints{ + endpointKey{ + Region: "ap-northeast-1", + }: endpoint{}, + endpointKey{ + Region: "ap-northeast-2", + }: endpoint{}, + endpointKey{ + Region: "ap-south-1", + }: endpoint{}, + endpointKey{ + Region: "eu-central-1", + }: endpoint{}, + endpointKey{ + Region: "eu-west-1", + }: endpoint{}, + endpointKey{ + Region: "us-east-1", + }: endpoint{}, + endpointKey{ + Region: "us-west-2", + }: endpoint{}, + }, + }, "kafka": service{ Endpoints: serviceEndpoints{ endpointKey{ diff --git a/aws/version.go b/aws/version.go index f61d223cc0a..55e65f65519 100644 --- a/aws/version.go +++ b/aws/version.go @@ -5,4 +5,4 @@ package aws const SDKName = "aws-sdk-go" // SDKVersion is the version of this SDK -const SDKVersion = "1.44.228" +const SDKVersion = "1.44.229" diff --git a/models/apis/comprehend/2017-11-27/api-2.json b/models/apis/comprehend/2017-11-27/api-2.json index fede2943c8c..6857e61da83 100644 --- a/models/apis/comprehend/2017-11-27/api-2.json +++ b/models/apis/comprehend/2017-11-27/api-2.json @@ -2702,7 +2702,8 @@ "LanguageCode":{"shape":"LanguageCode"}, "DataAccessRoleArn":{"shape":"IamRoleArn"}, "VolumeKmsKeyId":{"shape":"KmsKeyId"}, - "VpcConfig":{"shape":"VpcConfig"} + "VpcConfig":{"shape":"VpcConfig"}, + "FlywheelArn":{"shape":"ComprehendFlywheelArn"} } }, "EntitiesDetectionJobPropertiesList":{ @@ -3690,7 +3691,8 @@ "STOP_REQUESTED", "STOPPED", "IN_ERROR", - "TRAINED" + "TRAINED", + "TRAINED_WITH_WARNING" ] }, "ModelType":{ diff --git a/models/apis/comprehend/2017-11-27/docs-2.json b/models/apis/comprehend/2017-11-27/docs-2.json index 3cdb9d2b93e..2ddba539845 100644 --- a/models/apis/comprehend/2017-11-27/docs-2.json +++ b/models/apis/comprehend/2017-11-27/docs-2.json @@ -1,6 +1,6 @@ { "version": "2.0", - "service": "
Amazon Comprehend is an AWS service for gaining insight into the content of documents. Use these actions to determine the topics contained in your documents, the topics they discuss, the predominant sentiment expressed in them, the predominant language used, and more.
", + "service": "Amazon Comprehend is an Amazon Web Services service for gaining insight into the content of documents. Use these actions to determine the topics contained in your documents, the topics they discuss, the predominant sentiment expressed in them, the predominant language used, and more.
", "operations": { "BatchDetectDominantLanguage": "Determines the dominant language of the input text for a batch of documents. For a list of languages that Amazon Comprehend can detect, see Amazon Comprehend Supported Languages.
", "BatchDetectEntities": "Inspects the text of a batch of documents for named entities and returns information about them. For more information about named entities, see Entities in the Comprehend Developer Guide.
", @@ -14,7 +14,7 @@ "CreateDocumentClassifier": "Creates a new document classifier that you can use to categorize documents. To create a classifier, you provide a set of training documents that labeled with the categories that you want to use. After the classifier is trained you can use it to categorize a set of labeled documents into the categories. For more information, see Document Classification in the Comprehend Developer Guide.
", "CreateEndpoint": "Creates a model-specific endpoint for synchronous inference for a previously trained custom model For information about endpoints, see Managing endpoints.
", "CreateEntityRecognizer": "Creates an entity recognizer using submitted files. After your CreateEntityRecognizer
request is submitted, you can check job status using the DescribeEntityRecognizer
API.
A flywheel is an AWS resource that orchestrates the ongoing training of a model for custom classification or custom entity recognition. You can create a flywheel to start with an existing trained model, or Comprehend can create and train a new model.
When you create the flywheel, Comprehend creates a data lake in your account. The data lake holds the training data and test data for all versions of the model.
To use a flywheel with an existing trained model, you specify the active model version. Comprehend copies the model's training data and test data into the flywheel's data lake.
To use the flywheel with a new model, you need to provide a dataset for training data (and optional test data) when you create the flywheel.
For more information about flywheels, see Flywheel overview in the Amazon Comprehend Developer Guide.
", + "CreateFlywheel": "A flywheel is an Amazon Web Services resource that orchestrates the ongoing training of a model for custom classification or custom entity recognition. You can create a flywheel to start with an existing trained model, or Comprehend can create and train a new model.
When you create the flywheel, Comprehend creates a data lake in your account. The data lake holds the training data and test data for all versions of the model.
To use a flywheel with an existing trained model, you specify the active model version. Comprehend copies the model's training data and test data into the flywheel's data lake.
To use the flywheel with a new model, you need to provide a dataset for training data (and optional test data) when you create the flywheel.
For more information about flywheels, see Flywheel overview in the Amazon Comprehend Developer Guide.
", "DeleteDocumentClassifier": "Deletes a previously created document classifier
Only those classifiers that are in terminated states (IN_ERROR, TRAINED) will be deleted. If an active inference job is using the model, a ResourceInUseException
will be returned.
This is an asynchronous action that puts the classifier into a DELETING state, and it is then removed by a background job. Once removed, the classifier disappears from your account and is no longer available for use.
", "DeleteEndpoint": "Deletes a model-specific endpoint for a previously-trained custom model. All endpoints must be deleted in order for the model to be deleted. For information about endpoints, see Managing endpoints.
", "DeleteEntityRecognizer": "Deletes an entity recognizer.
Only those recognizers that are in terminated states (IN_ERROR, TRAINED) will be deleted. If an active inference job is using the model, a ResourceInUseException
will be returned.
This is an asynchronous action that puts the recognizer into a DELETING state, and it is then removed by a background job. Once removed, the recognizer disappears from your account and is no longer available for use.
", @@ -43,8 +43,8 @@ "DetectSentiment": "Inspects text and returns an inference of the prevailing sentiment (POSITIVE
, NEUTRAL
, MIXED
, or NEGATIVE
).
Inspects text for syntax and the part of speech of words in the document. For more information, see Syntax in the Comprehend Developer Guide.
", "DetectTargetedSentiment": "Inspects the input text and returns a sentiment analysis for each entity identified in the text.
For more information about targeted sentiment, see Targeted sentiment.
", - "ImportModel": "Creates a new custom model that replicates a source custom model that you import. The source model can be in your AWS account or another one.
If the source model is in another AWS account, then it must have a resource-based policy that authorizes you to import it.
The source model must be in the same AWS region that you're using when you import. You can't import a model that's in a different region.
", - "ListDatasets": "List the datasets that you have configured in this region. For more information about datasets, see Flywheel overview in the Amazon Comprehend Developer Guide.
", + "ImportModel": "Creates a new custom model that replicates a source custom model that you import. The source model can be in your Amazon Web Services account or another one.
If the source model is in another Amazon Web Services account, then it must have a resource-based policy that authorizes you to import it.
The source model must be in the same Amazon Web Services Region that you're using when you import. You can't import a model that's in a different Region.
", + "ListDatasets": "List the datasets that you have configured in this Region. For more information about datasets, see Flywheel overview in the Amazon Comprehend Developer Guide.
", "ListDocumentClassificationJobs": "Gets a list of the documentation classification jobs that you have submitted.
", "ListDocumentClassifierSummaries": "Gets a list of summaries of the document classifiers that you have created
", "ListDocumentClassifiers": "Gets a list of the document classifiers that you have created.
", @@ -62,7 +62,7 @@ "ListTagsForResource": "Lists all tags associated with a given Amazon Comprehend resource.
", "ListTargetedSentimentDetectionJobs": "Gets a list of targeted sentiment detection jobs that you have submitted.
", "ListTopicsDetectionJobs": "Gets a list of the topic detection jobs that you have submitted.
", - "PutResourcePolicy": "Attaches a resource-based policy to a custom model. You can use this policy to authorize an entity in another AWS account to import the custom model, which replicates it in Amazon Comprehend in their account.
", + "PutResourcePolicy": "Attaches a resource-based policy to a custom model. You can use this policy to authorize an entity in another Amazon Web Services account to import the custom model, which replicates it in Amazon Comprehend in their account.
", "StartDocumentClassificationJob": "Starts an asynchronous document classification job. Use the DescribeDocumentClassificationJob
operation to track the progress of the job.
Starts an asynchronous dominant language detection job for a collection of documents. Use the operation to track the status of a job.
", "StartEntitiesDetectionJob": "Starts an asynchronous entity detection job for a collection of documents. Use the operation to track the status of a job.
This API can be used for either standard entity detection or custom entity recognition. In order to be used for custom entity recognition, the optional EntityRecognizerArn
must be used in order to provide access to the recognizer being used to detect the custom entity.
The Amazon Resource Name (ARN) of the document classification job. It is a unique, fully qualified identifier for the job. It includes the AWS account, Region, and the job ID. The format of the ARN is as follows:
arn:<partition>:comprehend:<region>:<account-id>:document-classification-job/<job-id>
The following is an example job ARN:
arn:aws:comprehend:us-west-2:111122223333:document-classification-job/1234abcd12ab34cd56ef1234567890ab
The Amazon Resource Name (ARN) of the dominant language detection job. It is a unique, fully qualified identifier for the job. It includes the AWS account, Region, and the job ID. The format of the ARN is as follows:
arn:<partition>:comprehend:<region>:<account-id>:dominant-language-detection-job/<job-id>
The following is an example job ARN:
arn:aws:comprehend:us-west-2:111122223333:dominant-language-detection-job/1234abcd12ab34cd56ef1234567890ab
The Amazon Resource Name (ARN) of the entities detection job. It is a unique, fully qualified identifier for the job. It includes the AWS account, Region, and the job ID. The format of the ARN is as follows:
arn:<partition>:comprehend:<region>:<account-id>:entities-detection-job/<job-id>
The following is an example job ARN:
arn:aws:comprehend:us-west-2:111122223333:entities-detection-job/1234abcd12ab34cd56ef1234567890ab
The Amazon Resource Name (ARN) of the events detection job. It is a unique, fully qualified identifier for the job. It includes the AWS account, Region, and the job ID. The format of the ARN is as follows:
arn:<partition>:comprehend:<region>:<account-id>:events-detection-job/<job-id>
The following is an example job ARN:
arn:aws:comprehend:us-west-2:111122223333:events-detection-job/1234abcd12ab34cd56ef1234567890ab
The Amazon Resource Name (ARN) of the key phrases detection job. It is a unique, fully qualified identifier for the job. It includes the AWS account, Region, and the job ID. The format of the ARN is as follows:
arn:<partition>:comprehend:<region>:<account-id>:key-phrases-detection-job/<job-id>
The following is an example job ARN:
arn:aws:comprehend:us-west-2:111122223333:key-phrases-detection-job/1234abcd12ab34cd56ef1234567890ab
The Amazon Resource Name (ARN) of the document classification job. It is a unique, fully qualified identifier for the job. It includes the Amazon Web Services account, Amazon Web Services Region, and the job ID. The format of the ARN is as follows:
arn:<partition>:comprehend:<region>:<account-id>:document-classification-job/<job-id>
The following is an example job ARN:
arn:aws:comprehend:us-west-2:111122223333:document-classification-job/1234abcd12ab34cd56ef1234567890ab
The Amazon Resource Name (ARN) of the dominant language detection job. It is a unique, fully qualified identifier for the job. It includes the Amazon Web Services account, Amazon Web Services Region, and the job ID. The format of the ARN is as follows:
arn:<partition>:comprehend:<region>:<account-id>:dominant-language-detection-job/<job-id>
The following is an example job ARN:
arn:aws:comprehend:us-west-2:111122223333:dominant-language-detection-job/1234abcd12ab34cd56ef1234567890ab
The Amazon Resource Name (ARN) of the entities detection job. It is a unique, fully qualified identifier for the job. It includes the Amazon Web Services account, Amazon Web Services Region, and the job ID. The format of the ARN is as follows:
arn:<partition>:comprehend:<region>:<account-id>:entities-detection-job/<job-id>
The following is an example job ARN:
arn:aws:comprehend:us-west-2:111122223333:entities-detection-job/1234abcd12ab34cd56ef1234567890ab
The Amazon Resource Name (ARN) of the events detection job. It is a unique, fully qualified identifier for the job. It includes the Amazon Web Services account, Amazon Web Services Region, and the job ID. The format of the ARN is as follows:
arn:<partition>:comprehend:<region>:<account-id>:events-detection-job/<job-id>
The following is an example job ARN:
arn:aws:comprehend:us-west-2:111122223333:events-detection-job/1234abcd12ab34cd56ef1234567890ab
The Amazon Resource Name (ARN) of the key phrases detection job. It is a unique, fully qualified identifier for the job. It includes the Amazon Web Services account, Amazon Web Services Region, and the job ID. The format of the ARN is as follows:
arn:<partition>:comprehend:<region>:<account-id>:key-phrases-detection-job/<job-id>
The following is an example job ARN:
arn:aws:comprehend:us-west-2:111122223333:key-phrases-detection-job/1234abcd12ab34cd56ef1234567890ab
The Amazon Resource Name (ARN) of the given Amazon Comprehend resource you are querying.
", "ListTagsForResourceResponse$ResourceArn": "The Amazon Resource Name (ARN) of the given Amazon Comprehend resource you are querying.
", - "PiiEntitiesDetectionJobProperties$JobArn": "The Amazon Resource Name (ARN) of the PII entities detection job. It is a unique, fully qualified identifier for the job. It includes the AWS account, Region, and the job ID. The format of the ARN is as follows:
arn:<partition>:comprehend:<region>:<account-id>:pii-entities-detection-job/<job-id>
The following is an example job ARN:
arn:aws:comprehend:us-west-2:111122223333:pii-entities-detection-job/1234abcd12ab34cd56ef1234567890ab
The Amazon Resource Name (ARN) of the sentiment detection job. It is a unique, fully qualified identifier for the job. It includes the AWS account, Region, and the job ID. The format of the ARN is as follows:
arn:<partition>:comprehend:<region>:<account-id>:sentiment-detection-job/<job-id>
The following is an example job ARN:
arn:aws:comprehend:us-west-2:111122223333:sentiment-detection-job/1234abcd12ab34cd56ef1234567890ab
The Amazon Resource Name (ARN) of the document classification job. It is a unique, fully qualified identifier for the job. It includes the AWS account, Region, and the job ID. The format of the ARN is as follows:
arn:<partition>:comprehend:<region>:<account-id>:document-classification-job/<job-id>
The following is an example job ARN:
arn:aws:comprehend:us-west-2:111122223333:document-classification-job/1234abcd12ab34cd56ef1234567890ab
The Amazon Resource Name (ARN) of the dominant language detection job. It is a unique, fully qualified identifier for the job. It includes the AWS account, Region, and the job ID. The format of the ARN is as follows:
arn:<partition>:comprehend:<region>:<account-id>:dominant-language-detection-job/<job-id>
The following is an example job ARN:
arn:aws:comprehend:us-west-2:111122223333:dominant-language-detection-job/1234abcd12ab34cd56ef1234567890ab
The Amazon Resource Name (ARN) of the entities detection job. It is a unique, fully qualified identifier for the job. It includes the AWS account, Region, and the job ID. The format of the ARN is as follows:
arn:<partition>:comprehend:<region>:<account-id>:entities-detection-job/<job-id>
The following is an example job ARN:
arn:aws:comprehend:us-west-2:111122223333:entities-detection-job/1234abcd12ab34cd56ef1234567890ab
The Amazon Resource Name (ARN) of the events detection job. It is a unique, fully qualified identifier for the job. It includes the AWS account, Region, and the job ID. The format of the ARN is as follows:
arn:<partition>:comprehend:<region>:<account-id>:events-detection-job/<job-id>
The following is an example job ARN:
arn:aws:comprehend:us-west-2:111122223333:events-detection-job/1234abcd12ab34cd56ef1234567890ab
The Amazon Resource Name (ARN) of the key phrase detection job. It is a unique, fully qualified identifier for the job. It includes the AWS account, Region, and the job ID. The format of the ARN is as follows:
arn:<partition>:comprehend:<region>:<account-id>:key-phrases-detection-job/<job-id>
The following is an example job ARN:
arn:aws:comprehend:us-west-2:111122223333:key-phrases-detection-job/1234abcd12ab34cd56ef1234567890ab
The Amazon Resource Name (ARN) of the PII entity detection job. It is a unique, fully qualified identifier for the job. It includes the AWS account, Region, and the job ID. The format of the ARN is as follows:
arn:<partition>:comprehend:<region>:<account-id>:pii-entities-detection-job/<job-id>
The following is an example job ARN:
arn:aws:comprehend:us-west-2:111122223333:pii-entities-detection-job/1234abcd12ab34cd56ef1234567890ab
The Amazon Resource Name (ARN) of the sentiment detection job. It is a unique, fully qualified identifier for the job. It includes the AWS account, Region, and the job ID. The format of the ARN is as follows:
arn:<partition>:comprehend:<region>:<account-id>:sentiment-detection-job/<job-id>
The following is an example job ARN:
arn:aws:comprehend:us-west-2:111122223333:sentiment-detection-job/1234abcd12ab34cd56ef1234567890ab
The Amazon Resource Name (ARN) of the targeted sentiment detection job. It is a unique, fully qualified identifier for the job. It includes the AWS account, Region, and the job ID. The format of the ARN is as follows:
arn:<partition>:comprehend:<region>:<account-id>:targeted-sentiment-detection-job/<job-id>
The following is an example job ARN:
arn:aws:comprehend:us-west-2:111122223333:targeted-sentiment-detection-job/1234abcd12ab34cd56ef1234567890ab
The Amazon Resource Name (ARN) of the topics detection job. It is a unique, fully qualified identifier for the job. It includes the AWS account, Region, and the job ID. The format of the ARN is as follows:
arn:<partition>:comprehend:<region>:<account-id>:topics-detection-job/<job-id>
The following is an example job ARN:
arn:aws:comprehend:us-west-2:111122223333:document-classification-job/1234abcd12ab34cd56ef1234567890ab
The Amazon Resource Name (ARN) of the PII entities detection job. It is a unique, fully qualified identifier for the job. It includes the Amazon Web Services account, Amazon Web Services Region, and the job ID. The format of the ARN is as follows:
arn:<partition>:comprehend:<region>:<account-id>:pii-entities-detection-job/<job-id>
The following is an example job ARN:
arn:aws:comprehend:us-west-2:111122223333:pii-entities-detection-job/1234abcd12ab34cd56ef1234567890ab
The Amazon Resource Name (ARN) of the sentiment detection job. It is a unique, fully qualified identifier for the job. It includes the Amazon Web Services account, Amazon Web Services Region, and the job ID. The format of the ARN is as follows:
arn:<partition>:comprehend:<region>:<account-id>:sentiment-detection-job/<job-id>
The following is an example job ARN:
arn:aws:comprehend:us-west-2:111122223333:sentiment-detection-job/1234abcd12ab34cd56ef1234567890ab
The Amazon Resource Name (ARN) of the document classification job. It is a unique, fully qualified identifier for the job. It includes the Amazon Web Services account, Amazon Web Services Region, and the job ID. The format of the ARN is as follows:
arn:<partition>:comprehend:<region>:<account-id>:document-classification-job/<job-id>
The following is an example job ARN:
arn:aws:comprehend:us-west-2:111122223333:document-classification-job/1234abcd12ab34cd56ef1234567890ab
The Amazon Resource Name (ARN) of the dominant language detection job. It is a unique, fully qualified identifier for the job. It includes the Amazon Web Services account, Amazon Web Services Region, and the job ID. The format of the ARN is as follows:
arn:<partition>:comprehend:<region>:<account-id>:dominant-language-detection-job/<job-id>
The following is an example job ARN:
arn:aws:comprehend:us-west-2:111122223333:dominant-language-detection-job/1234abcd12ab34cd56ef1234567890ab
The Amazon Resource Name (ARN) of the entities detection job. It is a unique, fully qualified identifier for the job. It includes the Amazon Web Services account, Amazon Web Services Region, and the job ID. The format of the ARN is as follows:
arn:<partition>:comprehend:<region>:<account-id>:entities-detection-job/<job-id>
The following is an example job ARN:
arn:aws:comprehend:us-west-2:111122223333:entities-detection-job/1234abcd12ab34cd56ef1234567890ab
The Amazon Resource Name (ARN) of the events detection job. It is a unique, fully qualified identifier for the job. It includes the Amazon Web Services account, Amazon Web Services Region, and the job ID. The format of the ARN is as follows:
arn:<partition>:comprehend:<region>:<account-id>:events-detection-job/<job-id>
The following is an example job ARN:
arn:aws:comprehend:us-west-2:111122223333:events-detection-job/1234abcd12ab34cd56ef1234567890ab
The Amazon Resource Name (ARN) of the key phrase detection job. It is a unique, fully qualified identifier for the job. It includes the Amazon Web Services account, Amazon Web Services Region, and the job ID. The format of the ARN is as follows:
arn:<partition>:comprehend:<region>:<account-id>:key-phrases-detection-job/<job-id>
The following is an example job ARN:
arn:aws:comprehend:us-west-2:111122223333:key-phrases-detection-job/1234abcd12ab34cd56ef1234567890ab
The Amazon Resource Name (ARN) of the PII entity detection job. It is a unique, fully qualified identifier for the job. It includes the Amazon Web Services account, Amazon Web Services Region, and the job ID. The format of the ARN is as follows:
arn:<partition>:comprehend:<region>:<account-id>:pii-entities-detection-job/<job-id>
The following is an example job ARN:
arn:aws:comprehend:us-west-2:111122223333:pii-entities-detection-job/1234abcd12ab34cd56ef1234567890ab
The Amazon Resource Name (ARN) of the sentiment detection job. It is a unique, fully qualified identifier for the job. It includes the Amazon Web Services account, Amazon Web Services Region, and the job ID. The format of the ARN is as follows:
arn:<partition>:comprehend:<region>:<account-id>:sentiment-detection-job/<job-id>
The following is an example job ARN:
arn:aws:comprehend:us-west-2:111122223333:sentiment-detection-job/1234abcd12ab34cd56ef1234567890ab
The Amazon Resource Name (ARN) of the targeted sentiment detection job. It is a unique, fully qualified identifier for the job. It includes the Amazon Web Services account, Amazon Web Services Region, and the job ID. The format of the ARN is as follows:
arn:<partition>:comprehend:<region>:<account-id>:targeted-sentiment-detection-job/<job-id>
The following is an example job ARN:
arn:aws:comprehend:us-west-2:111122223333:targeted-sentiment-detection-job/1234abcd12ab34cd56ef1234567890ab
The Amazon Resource Name (ARN) of the topics detection job. It is a unique, fully qualified identifier for the job. It includes the Amazon Web Services account, Amazon Web Services Region, and the job ID. The format of the ARN is as follows:
arn:<partition>:comprehend:<region>:<account-id>:topics-detection-job/<job-id>
The following is an example job ARN:
arn:aws:comprehend:us-west-2:111122223333:document-classification-job/1234abcd12ab34cd56ef1234567890ab
The Amazon Resource Name (ARN) of the given Amazon Comprehend resource to which you want to associate the tags.
", - "TargetedSentimentDetectionJobProperties$JobArn": "The Amazon Resource Name (ARN) of the targeted sentiment detection job. It is a unique, fully qualified identifier for the job. It includes the AWS account, Region, and the job ID. The format of the ARN is as follows:
arn:<partition>:comprehend:<region>:<account-id>:targeted-sentiment-detection-job/<job-id>
The following is an example job ARN:
arn:aws:comprehend:us-west-2:111122223333:targeted-sentiment-detection-job/1234abcd12ab34cd56ef1234567890ab
The Amazon Resource Name (ARN) of the topics detection job. It is a unique, fully qualified identifier for the job. It includes the AWS account, Region, and the job ID. The format of the ARN is as follows:
arn:<partition>:comprehend:<region>:<account-id>:topics-detection-job/<job-id>
The following is an example job ARN:
arn:aws:comprehend:us-west-2:111122223333:topics-detection-job/1234abcd12ab34cd56ef1234567890ab
The Amazon Resource Name (ARN) of the targeted sentiment detection job. It is a unique, fully qualified identifier for the job. It includes the Amazon Web Services account, Amazon Web Services Region, and the job ID. The format of the ARN is as follows:
arn:<partition>:comprehend:<region>:<account-id>:targeted-sentiment-detection-job/<job-id>
The following is an example job ARN:
arn:aws:comprehend:us-west-2:111122223333:targeted-sentiment-detection-job/1234abcd12ab34cd56ef1234567890ab
The Amazon Resource Name (ARN) of the topics detection job. It is a unique, fully qualified identifier for the job. It includes the Amazon Web Services account, Amazon Web Services Region, and the job ID. The format of the ARN is as follows:
arn:<partition>:comprehend:<region>:<account-id>:topics-detection-job/<job-id>
The following is an example job ARN:
arn:aws:comprehend:us-west-2:111122223333:topics-detection-job/1234abcd12ab34cd56ef1234567890ab
The Amazon Resource Name (ARN) of the given Amazon Comprehend resource from which you want to remove the tags.
" } }, @@ -359,7 +359,7 @@ "refs": { "CreateDatasetRequest$DatasetName": "Name of the dataset.
", "CreateDocumentClassifierRequest$DocumentClassifierName": "The name of the document classifier.
", - "CreateEntityRecognizerRequest$RecognizerName": "The name given to the newly created recognizer. Recognizer names can be a maximum of 256 characters. Alphanumeric characters, hyphens (-) and underscores (_) are allowed. The name must be unique in the account/region.
", + "CreateEntityRecognizerRequest$RecognizerName": "The name given to the newly created recognizer. Recognizer names can be a maximum of 256 characters. Alphanumeric characters, hyphens (-) and underscores (_) are allowed. The name must be unique in the account/Region.
", "CreateFlywheelRequest$FlywheelName": "Name for the flywheel.
", "DatasetProperties$DatasetName": "The name of the dataset.
", "DocumentClassifierFilter$DocumentClassifierName": "The name that you assigned to the document classifier
", @@ -405,6 +405,7 @@ "DocumentClassificationJobProperties$FlywheelArn": "The Amazon Resource Number (ARN) of the flywheel
", "DocumentClassifierProperties$FlywheelArn": "The Amazon Resource Number (ARN) of the flywheel
", "EndpointProperties$FlywheelArn": "The Amazon Resource Number (ARN) of the flywheel
", + "EntitiesDetectionJobProperties$FlywheelArn": "The Amazon Resource Name (ARN) of the flywheel associated with this job.
", "EntityRecognizerProperties$FlywheelArn": "The Amazon Resource Number (ARN) of the flywheel
", "FlywheelIterationProperties$FlywheelArn": "", "FlywheelProperties$FlywheelArn": "The Amazon Resource Number (ARN) of the flywheel.
", @@ -516,7 +517,7 @@ "DetectDominantLanguageRequest$Text": "A UTF-8 text string. The string must contain at least 20 characters. The maximum string size is 100 KB.
", "DetectEntitiesRequest$Text": "A UTF-8 text string. The maximum string size is 100 KB. If you enter text using this parameter, do not use the Bytes
parameter.
A UTF-8 text string. The string must contain less than 100 KB of UTF-8 encoded characters.
", - "DetectSentimentRequest$Text": "A UTF-8 text string. The maximum string size is 5 KB.
Amazon Comprehend performs real-time sentiment analysis on the first 500 characters of the input text and ignores any additional text in the input.
A UTF-8 text string. The maximum string size is 5 KB.
", "DetectSyntaxRequest$Text": "A UTF-8 string. The maximum string size is 5 KB.
", "DetectTargetedSentimentRequest$Text": "A UTF-8 text string. The maximum string length is 5 KB.
" } @@ -527,7 +528,7 @@ "BatchDetectDominantLanguageRequest$TextList": "A list containing the UTF-8 encoded text of the input documents. The list can contain a maximum of 25 documents. Each document should contain at least 20 characters. The maximum size of each document is 5 KB.
", "BatchDetectEntitiesRequest$TextList": "A list containing the UTF-8 encoded text of the input documents. The list can contain a maximum of 25 documents. The maximum size of each document is 5 KB.
", "BatchDetectKeyPhrasesRequest$TextList": "A list containing the UTF-8 encoded text of the input documents. The list can contain a maximum of 25 documents. The maximum size of each document is 5 KB.
", - "BatchDetectSentimentRequest$TextList": "A list containing the UTF-8 encoded text of the input documents. The list can contain a maximum of 25 documents. The maximum size of each document is 5 KB.
Amazon Comprehend performs real-time sentiment analysis on the first 500 characters of the input text and ignores any additional text in the input.
A list containing the UTF-8 encoded text of the input documents. The list can contain a maximum of 25 documents. The maximum size of each document is 5 KB.
", "BatchDetectSyntaxRequest$TextList": "A list containing the UTF-8 encoded text of the input documents. The list can contain a maximum of 25 documents. The maximum size for each document is 5 KB.
", "BatchDetectTargetedSentimentRequest$TextList": "A list containing the UTF-8 encoded text of the input documents. The list can contain a maximum of 25 documents. The maximum size of each document is 5 KB.
" } @@ -953,7 +954,7 @@ "DescribeDocumentClassifierRequest$DocumentClassifierArn": "The Amazon Resource Name (ARN) that identifies the document classifier. The CreateDocumentClassifier
operation returns this identifier in its response.
The Amazon Resource Name (ARN) that identifies the document classifier.
", "DocumentClassifierProperties$DocumentClassifierArn": "The Amazon Resource Name (ARN) that identifies the document classifier.
", - "DocumentClassifierProperties$SourceModelArn": "The Amazon Resource Name (ARN) of the source model. This model was imported from a different AWS account to create the document classifier model in your AWS account.
", + "DocumentClassifierProperties$SourceModelArn": "The Amazon Resource Name (ARN) of the source model. This model was imported from a different Amazon Web Services account to create the document classifier model in your Amazon Web Services account.
", "StartDocumentClassificationJobRequest$DocumentClassifierArn": "The Amazon Resource Name (ARN) of the document classifier to use to process the job.
", "StartDocumentClassificationJobResponse$DocumentClassifierArn": "The ARN of the custom classification model.
", "StopTrainingDocumentClassifierRequest$DocumentClassifierArn": "The Amazon Resource Name (ARN) that identifies the document classifier currently being trained.
" @@ -1206,7 +1207,7 @@ "DescribeEntityRecognizerRequest$EntityRecognizerArn": "The Amazon Resource Name (ARN) that identifies the entity recognizer.
", "EntitiesDetectionJobProperties$EntityRecognizerArn": "The Amazon Resource Name (ARN) that identifies the entity recognizer.
", "EntityRecognizerProperties$EntityRecognizerArn": "The Amazon Resource Name (ARN) that identifies the entity recognizer.
", - "EntityRecognizerProperties$SourceModelArn": "The Amazon Resource Name (ARN) of the source model. This model was imported from a different AWS account to create the entity recognizer model in your AWS account.
", + "EntityRecognizerProperties$SourceModelArn": "The Amazon Resource Name (ARN) of the source model. This model was imported from a different Amazon Web Services account to create the entity recognizer model in your Amazon Web Services account.
", "StartEntitiesDetectionJobRequest$EntityRecognizerArn": "The Amazon Resource Name (ARN) that identifies the specific entity recognizer to be used by the StartEntitiesDetectionJob
. This ARN is optional and is only used for a custom entity recognition job.
The ARN of the custom entity recognition model.
", "StopTrainingEntityRecognizerRequest$EntityRecognizerArn": "The Amazon Resource Name (ARN) that identifies the entity recognizer currently being trained.
" @@ -1257,7 +1258,7 @@ "EntityRecognizerInputDataConfig": { "base": "Specifies the format and location of the input data.
", "refs": { - "CreateEntityRecognizerRequest$InputDataConfig": "Specifies the format and location of the input data. The S3 bucket containing the input data must be located in the same region as the entity recognizer being created.
", + "CreateEntityRecognizerRequest$InputDataConfig": "Specifies the format and location of the input data. The S3 bucket containing the input data must be located in the same Region as the entity recognizer being created.
", "EntityRecognizerProperties$InputDataConfig": "The input data properties of an entity recognizer.
" } }, @@ -1493,36 +1494,36 @@ "IamRoleArn": { "base": null, "refs": { - "CreateDocumentClassifierRequest$DataAccessRoleArn": "The Amazon Resource Name (ARN) of the AWS Identity and Access Management (IAM) role that grants Amazon Comprehend read access to your input data.
", - "CreateEndpointRequest$DataAccessRoleArn": "The Amazon Resource Name (ARN) of the AWS Identity and Access Management (IAM) role that grants Amazon Comprehend read access to trained custom models encrypted with a customer managed key (ModelKmsKeyId).
", - "CreateEntityRecognizerRequest$DataAccessRoleArn": "The Amazon Resource Name (ARN) of the AWS Identity and Access Management (IAM) role that grants Amazon Comprehend read access to your input data.
", - "CreateFlywheelRequest$DataAccessRoleArn": "The Amazon Resource Name (ARN) of the AWS Identity and Access Management (IAM) role that grants Amazon Comprehend the permissions required to access the flywheel data in the data lake.
", - "DocumentClassificationJobProperties$DataAccessRoleArn": "The Amazon Resource Name (ARN) of the AWS Identity and Access Management (IAM) role that grants Amazon Comprehend read access to your input data.
", - "DocumentClassifierProperties$DataAccessRoleArn": "The Amazon Resource Name (ARN) of the AWS Identity and Access Management (IAM) role that grants Amazon Comprehend read access to your input data.
", - "DominantLanguageDetectionJobProperties$DataAccessRoleArn": "The Amazon Resource Name (ARN) of the AWS Identity and Access Management (IAM) role that grants Amazon Comprehend read access to your input data.
", - "EndpointProperties$DataAccessRoleArn": "The Amazon Resource Name (ARN) of the AWS Identity and Access Management (IAM) role that grants Amazon Comprehend read access to trained custom models encrypted with a customer managed key (ModelKmsKeyId).
", + "CreateDocumentClassifierRequest$DataAccessRoleArn": "The Amazon Resource Name (ARN) of the IAM role that grants Amazon Comprehend read access to your input data.
", + "CreateEndpointRequest$DataAccessRoleArn": "The Amazon Resource Name (ARN) of the IAM role that grants Amazon Comprehend read access to trained custom models encrypted with a customer managed key (ModelKmsKeyId).
", + "CreateEntityRecognizerRequest$DataAccessRoleArn": "The Amazon Resource Name (ARN) of the IAM role that grants Amazon Comprehend read access to your input data.
", + "CreateFlywheelRequest$DataAccessRoleArn": "The Amazon Resource Name (ARN) of the IAM role that grants Amazon Comprehend the permissions required to access the flywheel data in the data lake.
", + "DocumentClassificationJobProperties$DataAccessRoleArn": "The Amazon Resource Name (ARN) of the IAM role that grants Amazon Comprehend read access to your input data.
", + "DocumentClassifierProperties$DataAccessRoleArn": "The Amazon Resource Name (ARN) of the IAM role that grants Amazon Comprehend read access to your input data.
", + "DominantLanguageDetectionJobProperties$DataAccessRoleArn": "The Amazon Resource Name (ARN) of the IAM role that grants Amazon Comprehend read access to your input data.
", + "EndpointProperties$DataAccessRoleArn": "The Amazon Resource Name (ARN) of the IAM role that grants Amazon Comprehend read access to trained custom models encrypted with a customer managed key (ModelKmsKeyId).
", "EndpointProperties$DesiredDataAccessRoleArn": "Data access role ARN to use in case the new model is encrypted with a customer KMS key.
", - "EntitiesDetectionJobProperties$DataAccessRoleArn": "The Amazon Resource Name (ARN) of the AWS Identity and Access Management (IAM) role that grants Amazon Comprehend read access to your input data.
", - "EntityRecognizerProperties$DataAccessRoleArn": "The Amazon Resource Name (ARN) of the AWS Identity and Access Management (IAM) role that grants Amazon Comprehend read access to your input data.
", - "EventsDetectionJobProperties$DataAccessRoleArn": "The Amazon Resource Name (ARN) of the AWS Identity and Access Management (IAM) role that grants Amazon Comprehend read access to your input data.
", - "FlywheelProperties$DataAccessRoleArn": "The Amazon Resource Name (ARN) of the AWS Identity and Access Management (IAM) role that grants Amazon Comprehend permission to access the flywheel data.
", - "ImportModelRequest$DataAccessRoleArn": "The Amazon Resource Name (ARN) of the AWS Identity and Access Management (IAM) role that grants Amazon Comprehend permission to use Amazon Key Management Service (KMS) to encrypt or decrypt the custom model.
", - "KeyPhrasesDetectionJobProperties$DataAccessRoleArn": "The Amazon Resource Name (ARN) of the AWS Identity and Access Management (IAM) role that grants Amazon Comprehend read access to your input data.
", - "PiiEntitiesDetectionJobProperties$DataAccessRoleArn": "The Amazon Resource Name (ARN) of the AWS Identity and Access Management (IAM) role that grants Amazon Comprehend read access to your input data.
", - "SentimentDetectionJobProperties$DataAccessRoleArn": "The Amazon Resource Name (ARN) of the AWS Identity and Access Management (IAM) role that grants Amazon Comprehend read access to your input data.
", - "StartDocumentClassificationJobRequest$DataAccessRoleArn": "The Amazon Resource Name (ARN) of the AWS Identity and Access Management (IAM) role that grants Amazon Comprehend read access to your input data.
", - "StartDominantLanguageDetectionJobRequest$DataAccessRoleArn": "The Amazon Resource Name (ARN) of the AWS Identity and Access Management (IAM) role that grants Amazon Comprehend read access to your input data. For more information, see https://docs.aws.amazon.com/comprehend/latest/dg/access-control-managing-permissions.html#auth-role-permissions.
", - "StartEntitiesDetectionJobRequest$DataAccessRoleArn": "The Amazon Resource Name (ARN) of the AWS Identity and Access Management (IAM) role that grants Amazon Comprehend read access to your input data. For more information, see https://docs.aws.amazon.com/comprehend/latest/dg/access-control-managing-permissions.html#auth-role-permissions.
", - "StartEventsDetectionJobRequest$DataAccessRoleArn": "The Amazon Resource Name (ARN) of the AWS Identity and Access Management (IAM) role that grants Amazon Comprehend read access to your input data.
", - "StartKeyPhrasesDetectionJobRequest$DataAccessRoleArn": "The Amazon Resource Name (ARN) of the AWS Identity and Access Management (IAM) role that grants Amazon Comprehend read access to your input data. For more information, see https://docs.aws.amazon.com/comprehend/latest/dg/access-control-managing-permissions.html#auth-role-permissions.
", - "StartPiiEntitiesDetectionJobRequest$DataAccessRoleArn": "The Amazon Resource Name (ARN) of the AWS Identity and Access Management (IAM) role that grants Amazon Comprehend read access to your input data.
", - "StartSentimentDetectionJobRequest$DataAccessRoleArn": "The Amazon Resource Name (ARN) of the AWS Identity and Access Management (IAM) role that grants Amazon Comprehend read access to your input data. For more information, see https://docs.aws.amazon.com/comprehend/latest/dg/access-control-managing-permissions.html#auth-role-permissions.
", - "StartTargetedSentimentDetectionJobRequest$DataAccessRoleArn": "The Amazon Resource Name (ARN) of the AWS Identity and Access Management (IAM) role that grants Amazon Comprehend read access to your input data. For more information, see Role-based permissions.
", - "StartTopicsDetectionJobRequest$DataAccessRoleArn": "The Amazon Resource Name (ARN) of the AWS Identity and Access Management (IAM) role that grants Amazon Comprehend read access to your input data. For more information, see https://docs.aws.amazon.com/comprehend/latest/dg/access-control-managing-permissions.html#auth-role-permissions.
", - "TargetedSentimentDetectionJobProperties$DataAccessRoleArn": "The Amazon Resource Name (ARN) of the AWS Identity and Access Management (IAM) role that grants Amazon Comprehend read access to your input data.
", - "TopicsDetectionJobProperties$DataAccessRoleArn": "The Amazon Resource Name (ARN) of the AWS Identity and Access Management (IAM) role that grants Amazon Comprehend read access to your job data.
", + "EntitiesDetectionJobProperties$DataAccessRoleArn": "The Amazon Resource Name (ARN) of the IAM role that grants Amazon Comprehend read access to your input data.
", + "EntityRecognizerProperties$DataAccessRoleArn": "The Amazon Resource Name (ARN) of the IAM role that grants Amazon Comprehend read access to your input data.
", + "EventsDetectionJobProperties$DataAccessRoleArn": "The Amazon Resource Name (ARN) of the IAM role that grants Amazon Comprehend read access to your input data.
", + "FlywheelProperties$DataAccessRoleArn": "The Amazon Resource Name (ARN) of the IAM role that grants Amazon Comprehend permission to access the flywheel data.
", + "ImportModelRequest$DataAccessRoleArn": "The Amazon Resource Name (ARN) of the IAM role that grants Amazon Comprehend permission to use Amazon Key Management Service (KMS) to encrypt or decrypt the custom model.
", + "KeyPhrasesDetectionJobProperties$DataAccessRoleArn": "The Amazon Resource Name (ARN) of the IAM role that grants Amazon Comprehend read access to your input data.
", + "PiiEntitiesDetectionJobProperties$DataAccessRoleArn": "The Amazon Resource Name (ARN) of the IAM role that grants Amazon Comprehend read access to your input data.
", + "SentimentDetectionJobProperties$DataAccessRoleArn": "The Amazon Resource Name (ARN) of the IAM role that grants Amazon Comprehend read access to your input data.
", + "StartDocumentClassificationJobRequest$DataAccessRoleArn": "The Amazon Resource Name (ARN) of the IAM role that grants Amazon Comprehend read access to your input data.
", + "StartDominantLanguageDetectionJobRequest$DataAccessRoleArn": "The Amazon Resource Name (ARN) of the IAM role that grants Amazon Comprehend read access to your input data. For more information, see Role-based permissions.
", + "StartEntitiesDetectionJobRequest$DataAccessRoleArn": "The Amazon Resource Name (ARN) of the IAM role that grants Amazon Comprehend read access to your input data. For more information, see Role-based permissions.
", + "StartEventsDetectionJobRequest$DataAccessRoleArn": "The Amazon Resource Name (ARN) of the IAM role that grants Amazon Comprehend read access to your input data.
", + "StartKeyPhrasesDetectionJobRequest$DataAccessRoleArn": "The Amazon Resource Name (ARN) of the IAM role that grants Amazon Comprehend read access to your input data. For more information, see Role-based permissions.
", + "StartPiiEntitiesDetectionJobRequest$DataAccessRoleArn": "The Amazon Resource Name (ARN) of the IAM role that grants Amazon Comprehend read access to your input data.
", + "StartSentimentDetectionJobRequest$DataAccessRoleArn": "The Amazon Resource Name (ARN) of the IAM role that grants Amazon Comprehend read access to your input data. For more information, see Role-based permissions.
", + "StartTargetedSentimentDetectionJobRequest$DataAccessRoleArn": "The Amazon Resource Name (ARN) of the IAM role that grants Amazon Comprehend read access to your input data. For more information, see Role-based permissions.
", + "StartTopicsDetectionJobRequest$DataAccessRoleArn": "The Amazon Resource Name (ARN) of the IAM role that grants Amazon Comprehend read access to your input data. For more information, see Role-based permissions.
", + "TargetedSentimentDetectionJobProperties$DataAccessRoleArn": "The Amazon Resource Name (ARN) of the IAM role that grants Amazon Comprehend read access to your input data.
", + "TopicsDetectionJobProperties$DataAccessRoleArn": "The Amazon Resource Name (ARN) of the IAM role that grants Amazon Comprehend read access to your job data.
", "UpdateEndpointRequest$DesiredDataAccessRoleArn": "Data access role ARN to use in case the new model is encrypted with a customer CMK.
", - "UpdateFlywheelRequest$DataAccessRoleArn": "The Amazon Resource Name (ARN) of the AWS Identity and Access Management (IAM) role that grants Amazon Comprehend permission to access the flywheel data.
" + "UpdateFlywheelRequest$DataAccessRoleArn": "The Amazon Resource Name (ARN) of the IAM role that grants Amazon Comprehend permission to access the flywheel data.
" } }, "ImportModelRequest": { @@ -1803,36 +1804,36 @@ "KmsKeyId": { "base": null, "refs": { - "CreateDocumentClassifierRequest$VolumeKmsKeyId": "ID for the AWS Key Management Service (KMS) key that Amazon Comprehend uses to encrypt data on the storage volume attached to the ML compute instance(s) that process the analysis job. The VolumeKmsKeyId can be either of the following formats:
KMS Key ID: \"1234abcd-12ab-34cd-56ef-1234567890ab\"
Amazon Resource Name (ARN) of a KMS Key: \"arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab\"
ID for the AWS Key Management Service (KMS) key that Amazon Comprehend uses to encrypt trained custom models. The ModelKmsKeyId can be either of the following formats:
KMS Key ID: \"1234abcd-12ab-34cd-56ef-1234567890ab\"
Amazon Resource Name (ARN) of a KMS Key: \"arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab\"
ID for the AWS Key Management Service (KMS) key that Amazon Comprehend uses to encrypt data on the storage volume attached to the ML compute instance(s) that process the analysis job. The VolumeKmsKeyId can be either of the following formats:
KMS Key ID: \"1234abcd-12ab-34cd-56ef-1234567890ab\"
Amazon Resource Name (ARN) of a KMS Key: \"arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab\"
ID for the AWS Key Management Service (KMS) key that Amazon Comprehend uses to encrypt trained custom models. The ModelKmsKeyId can be either of the following formats:
KMS Key ID: \"1234abcd-12ab-34cd-56ef-1234567890ab\"
Amazon Resource Name (ARN) of a KMS Key: \"arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab\"
ID for the AWS Key Management Service (KMS) key that Amazon Comprehend uses to encrypt trained custom models. The ModelKmsKeyId can be either of the following formats:
KMS Key ID: \"1234abcd-12ab-34cd-56ef-1234567890ab\"
Amazon Resource Name (ARN) of a KMS Key: \"arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab\"
ID for the Amazon Web Services Key Management Service (KMS) key that Amazon Comprehend uses to encrypt data on the storage volume attached to the ML compute instance(s) that process the analysis job. The VolumeKmsKeyId can be either of the following formats:
KMS Key ID: \"1234abcd-12ab-34cd-56ef-1234567890ab\"
Amazon Resource Name (ARN) of a KMS Key: \"arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab\"
ID for the KMS key that Amazon Comprehend uses to encrypt trained custom models. The ModelKmsKeyId can be either of the following formats:
KMS Key ID: \"1234abcd-12ab-34cd-56ef-1234567890ab\"
Amazon Resource Name (ARN) of a KMS Key: \"arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab\"
ID for the Amazon Web Services Key Management Service (KMS) key that Amazon Comprehend uses to encrypt data on the storage volume attached to the ML compute instance(s) that process the analysis job. The VolumeKmsKeyId can be either of the following formats:
KMS Key ID: \"1234abcd-12ab-34cd-56ef-1234567890ab\"
Amazon Resource Name (ARN) of a KMS Key: \"arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab\"
ID for the KMS key that Amazon Comprehend uses to encrypt trained custom models. The ModelKmsKeyId can be either of the following formats:
KMS Key ID: \"1234abcd-12ab-34cd-56ef-1234567890ab\"
Amazon Resource Name (ARN) of a KMS Key: \"arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab\"
ID for the KMS key that Amazon Comprehend uses to encrypt trained custom models. The ModelKmsKeyId can be either of the following formats:
KMS Key ID: \"1234abcd-12ab-34cd-56ef-1234567890ab\"
Amazon Resource Name (ARN) of a KMS Key: \"arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab\"
ID for the KMS key that Amazon Comprehend uses to encrypt the volume.
", "DataSecurityConfig$DataLakeKmsKeyId": "ID for the KMS key that Amazon Comprehend uses to encrypt the data in the data lake.
", - "DocumentClassificationJobProperties$VolumeKmsKeyId": "ID for the AWS Key Management Service (KMS) key that Amazon Comprehend uses to encrypt data on the storage volume attached to the ML compute instance(s) that process the analysis job. The VolumeKmsKeyId can be either of the following formats:
KMS Key ID: \"1234abcd-12ab-34cd-56ef-1234567890ab\"
Amazon Resource Name (ARN) of a KMS Key: \"arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab\"
ID for the AWS Key Management Service (KMS) key that Amazon Comprehend uses to encrypt the output results from an analysis job. The KmsKeyId can be one of the following formats:
KMS Key ID: \"1234abcd-12ab-34cd-56ef-1234567890ab\"
Amazon Resource Name (ARN) of a KMS Key: \"arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab\"
KMS Key Alias: \"alias/ExampleAlias\"
ARN of a KMS Key Alias: \"arn:aws:kms:us-west-2:111122223333:alias/ExampleAlias\"
ID for the AWS Key Management Service (KMS) key that Amazon Comprehend uses to encrypt data on the storage volume attached to the ML compute instance(s) that process the analysis job. The VolumeKmsKeyId can be either of the following formats:
KMS Key ID: \"1234abcd-12ab-34cd-56ef-1234567890ab\"
Amazon Resource Name (ARN) of a KMS Key: \"arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab\"
ID for the AWS Key Management Service (KMS) key that Amazon Comprehend uses to encrypt trained custom models. The ModelKmsKeyId can be either of the following formats:
KMS Key ID: \"1234abcd-12ab-34cd-56ef-1234567890ab\"
Amazon Resource Name (ARN) of a KMS Key: \"arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab\"
ID for the AWS Key Management Service (KMS) key that Amazon Comprehend uses to encrypt data on the storage volume attached to the ML compute instance(s) that process the analysis job. The VolumeKmsKeyId can be either of the following formats:
KMS Key ID: \"1234abcd-12ab-34cd-56ef-1234567890ab\"
Amazon Resource Name (ARN) of a KMS Key: \"arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab\"
ID for the AWS Key Management Service (KMS) key that Amazon Comprehend uses to encrypt data on the storage volume attached to the ML compute instance(s) that process the analysis job. The VolumeKmsKeyId can be either of the following formats:
KMS Key ID: \"1234abcd-12ab-34cd-56ef-1234567890ab\"
Amazon Resource Name (ARN) of a KMS Key: \"arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab\"
ID for the AWS Key Management Service (KMS) key that Amazon Comprehend uses to encrypt data on the storage volume attached to the ML compute instance(s) that process the analysis job. The VolumeKmsKeyId can be either of the following formats:
KMS Key ID: \"1234abcd-12ab-34cd-56ef-1234567890ab\"
Amazon Resource Name (ARN) of a KMS Key: \"arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab\"
ID for the AWS Key Management Service (KMS) key that Amazon Comprehend uses to encrypt trained custom models. The ModelKmsKeyId can be either of the following formats:
KMS Key ID: \"1234abcd-12ab-34cd-56ef-1234567890ab\"
Amazon Resource Name (ARN) of a KMS Key: \"arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab\"
ID for the AWS Key Management Service (KMS) key that Amazon Comprehend uses to encrypt trained custom models. The ModelKmsKeyId can be either of the following formats:
KMS Key ID: \"1234abcd-12ab-34cd-56ef-1234567890ab\"
Amazon Resource Name (ARN) of a KMS Key: \"arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab\"
ID for the AWS Key Management Service (KMS) key that Amazon Comprehend uses to encrypt data on the storage volume attached to the ML compute instance(s) that process the analysis job. The VolumeKmsKeyId can be either of the following formats:
KMS Key ID: \"1234abcd-12ab-34cd-56ef-1234567890ab\"
Amazon Resource Name (ARN) of a KMS Key: \"arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab\"
ID for the AWS Key Management Service (KMS) key that Amazon Comprehend uses to encrypt the output results from an analysis job. The KmsKeyId can be one of the following formats:
KMS Key ID: \"1234abcd-12ab-34cd-56ef-1234567890ab\"
Amazon Resource Name (ARN) of a KMS Key: \"arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab\"
KMS Key Alias: \"alias/ExampleAlias\"
ARN of a KMS Key Alias: \"arn:aws:kms:us-west-2:111122223333:alias/ExampleAlias\"
ID for the AWS Key Management Service (KMS) key that Amazon Comprehend uses to encrypt the output results from an analysis job.
", - "SentimentDetectionJobProperties$VolumeKmsKeyId": "ID for the AWS Key Management Service (KMS) key that Amazon Comprehend uses to encrypt data on the storage volume attached to the ML compute instance(s) that process the analysis job. The VolumeKmsKeyId can be either of the following formats:
KMS Key ID: \"1234abcd-12ab-34cd-56ef-1234567890ab\"
Amazon Resource Name (ARN) of a KMS Key: \"arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab\"
ID for the AWS Key Management Service (KMS) key that Amazon Comprehend uses to encrypt data on the storage volume attached to the ML compute instance(s) that process the analysis job. The VolumeKmsKeyId can be either of the following formats:
KMS Key ID: \"1234abcd-12ab-34cd-56ef-1234567890ab\"
Amazon Resource Name (ARN) of a KMS Key: \"arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab\"
ID for the AWS Key Management Service (KMS) key that Amazon Comprehend uses to encrypt data on the storage volume attached to the ML compute instance(s) that process the analysis job. The VolumeKmsKeyId can be either of the following formats:
KMS Key ID: \"1234abcd-12ab-34cd-56ef-1234567890ab\"
Amazon Resource Name (ARN) of a KMS Key: \"arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab\"
ID for the AWS Key Management Service (KMS) key that Amazon Comprehend uses to encrypt data on the storage volume attached to the ML compute instance(s) that process the analysis job. The VolumeKmsKeyId can be either of the following formats:
KMS Key ID: \"1234abcd-12ab-34cd-56ef-1234567890ab\"
Amazon Resource Name (ARN) of a KMS Key: \"arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab\"
ID for the AWS Key Management Service (KMS) key that Amazon Comprehend uses to encrypt data on the storage volume attached to the ML compute instance(s) that process the analysis job. The VolumeKmsKeyId can be either of the following formats:
KMS Key ID: \"1234abcd-12ab-34cd-56ef-1234567890ab\"
Amazon Resource Name (ARN) of a KMS Key: \"arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab\"
ID for the AWS Key Management Service (KMS) key that Amazon Comprehend uses to encrypt data on the storage volume attached to the ML compute instance(s) that process the analysis job. The VolumeKmsKeyId can be either of the following formats:
KMS Key ID: \"1234abcd-12ab-34cd-56ef-1234567890ab\"
Amazon Resource Name (ARN) of a KMS Key: \"arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab\"
ID for the Amazon Web Services Key Management Service (KMS) key that Amazon Comprehend uses to encrypt data on the storage volume attached to the ML compute instance(s) that process the analysis job. The VolumeKmsKeyId can be either of the following formats:
KMS Key ID: \"1234abcd-12ab-34cd-56ef-1234567890ab\"
Amazon Resource Name (ARN) of a KMS Key: \"arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab\"
ID for the Amazon Web Services Key Management Service (KMS) key that Amazon Comprehend uses to encrypt the output results from an analysis job. The KmsKeyId can be one of the following formats:
KMS Key ID: \"1234abcd-12ab-34cd-56ef-1234567890ab\"
Amazon Resource Name (ARN) of a KMS Key: \"arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab\"
KMS Key Alias: \"alias/ExampleAlias\"
ARN of a KMS Key Alias: \"arn:aws:kms:us-west-2:111122223333:alias/ExampleAlias\"
ID for the Amazon Web Services Key Management Service (KMS) key that Amazon Comprehend uses to encrypt data on the storage volume attached to the ML compute instance(s) that process the analysis job. The VolumeKmsKeyId can be either of the following formats:
KMS Key ID: \"1234abcd-12ab-34cd-56ef-1234567890ab\"
Amazon Resource Name (ARN) of a KMS Key: \"arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab\"
ID for the KMS key that Amazon Comprehend uses to encrypt trained custom models. The ModelKmsKeyId can be either of the following formats:
KMS Key ID: \"1234abcd-12ab-34cd-56ef-1234567890ab\"
Amazon Resource Name (ARN) of a KMS Key: \"arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab\"
ID for the Amazon Web Services Key Management Service (KMS) key that Amazon Comprehend uses to encrypt data on the storage volume attached to the ML compute instance(s) that process the analysis job. The VolumeKmsKeyId can be either of the following formats:
KMS Key ID: \"1234abcd-12ab-34cd-56ef-1234567890ab\"
Amazon Resource Name (ARN) of a KMS Key: \"arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab\"
ID for the Amazon Web Services Key Management Service (KMS) key that Amazon Comprehend uses to encrypt data on the storage volume attached to the ML compute instance(s) that process the analysis job. The VolumeKmsKeyId can be either of the following formats:
KMS Key ID: \"1234abcd-12ab-34cd-56ef-1234567890ab\"
Amazon Resource Name (ARN) of a KMS Key: \"arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab\"
ID for the Amazon Web Services Key Management Service (KMS) key that Amazon Comprehend uses to encrypt data on the storage volume attached to the ML compute instance(s) that process the analysis job. The VolumeKmsKeyId can be either of the following formats:
KMS Key ID: \"1234abcd-12ab-34cd-56ef-1234567890ab\"
Amazon Resource Name (ARN) of a KMS Key: \"arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab\"
ID for the KMS key that Amazon Comprehend uses to encrypt trained custom models. The ModelKmsKeyId can be either of the following formats:
KMS Key ID: \"1234abcd-12ab-34cd-56ef-1234567890ab\"
Amazon Resource Name (ARN) of a KMS Key: \"arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab\"
ID for the KMS key that Amazon Comprehend uses to encrypt trained custom models. The ModelKmsKeyId can be either of the following formats:
KMS Key ID: \"1234abcd-12ab-34cd-56ef-1234567890ab\"
Amazon Resource Name (ARN) of a KMS Key: \"arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab\"
ID for the KMS key that Amazon Comprehend uses to encrypt data on the storage volume attached to the ML compute instance(s) that process the analysis job. The VolumeKmsKeyId can be either of the following formats:
KMS Key ID: \"1234abcd-12ab-34cd-56ef-1234567890ab\"
Amazon Resource Name (ARN) of a KMS Key: \"arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab\"
ID for the Amazon Web Services Key Management Service (KMS) key that Amazon Comprehend uses to encrypt the output results from an analysis job. The KmsKeyId can be one of the following formats:
KMS Key ID: \"1234abcd-12ab-34cd-56ef-1234567890ab\"
Amazon Resource Name (ARN) of a KMS Key: \"arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab\"
KMS Key Alias: \"alias/ExampleAlias\"
ARN of a KMS Key Alias: \"arn:aws:kms:us-west-2:111122223333:alias/ExampleAlias\"
ID for the Amazon Web Services Key Management Service (KMS) key that Amazon Comprehend uses to encrypt the output results from an analysis job.
", + "SentimentDetectionJobProperties$VolumeKmsKeyId": "ID for the Amazon Web Services Key Management Service (KMS) key that Amazon Comprehend uses to encrypt data on the storage volume attached to the ML compute instance(s) that process the analysis job. The VolumeKmsKeyId can be either of the following formats:
KMS Key ID: \"1234abcd-12ab-34cd-56ef-1234567890ab\"
Amazon Resource Name (ARN) of a KMS Key: \"arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab\"
ID for the Amazon Web Services Key Management Service (KMS) key that Amazon Comprehend uses to encrypt data on the storage volume attached to the ML compute instance(s) that process the analysis job. The VolumeKmsKeyId can be either of the following formats:
KMS Key ID: \"1234abcd-12ab-34cd-56ef-1234567890ab\"
Amazon Resource Name (ARN) of a KMS Key: \"arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab\"
ID for the Amazon Web Services Key Management Service (KMS) key that Amazon Comprehend uses to encrypt data on the storage volume attached to the ML compute instance(s) that process the analysis job. The VolumeKmsKeyId can be either of the following formats:
KMS Key ID: \"1234abcd-12ab-34cd-56ef-1234567890ab\"
Amazon Resource Name (ARN) of a KMS Key: \"arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab\"
ID for the Amazon Web Services Key Management Service (KMS) key that Amazon Comprehend uses to encrypt data on the storage volume attached to the ML compute instance(s) that process the analysis job. The VolumeKmsKeyId can be either of the following formats:
KMS Key ID: \"1234abcd-12ab-34cd-56ef-1234567890ab\"
Amazon Resource Name (ARN) of a KMS Key: \"arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab\"
ID for the Amazon Web Services Key Management Service (KMS) key that Amazon Comprehend uses to encrypt data on the storage volume attached to the ML compute instance(s) that process the analysis job. The VolumeKmsKeyId can be either of the following formats:
KMS Key ID: \"1234abcd-12ab-34cd-56ef-1234567890ab\"
Amazon Resource Name (ARN) of a KMS Key: \"arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab\"
ID for the Amazon Web Services Key Management Service (KMS) key that Amazon Comprehend uses to encrypt data on the storage volume attached to the ML compute instance(s) that process the analysis job. The VolumeKmsKeyId can be either of the following formats:
KMS Key ID: \"1234abcd-12ab-34cd-56ef-1234567890ab\"
Amazon Resource Name (ARN) of a KMS Key: \"arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab\"
ID for the KMS key that Amazon Comprehend uses to encrypt data on the storage volume attached to the ML compute instance(s) that process the analysis job. The VolumeKmsKeyId can be either of the following formats:
KMS Key ID: \"1234abcd-12ab-34cd-56ef-1234567890ab\"
Amazon Resource Name (ARN) of a KMS Key: \"arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab\"
ID for the AWS Key Management Service (KMS) key that Amazon Comprehend uses to encrypt data on the storage volume attached to the ML compute instance(s) that process the analysis job. The VolumeKmsKeyId can be either of the following formats:
KMS Key ID: \"1234abcd-12ab-34cd-56ef-1234567890ab\"
Amazon Resource Name (ARN) of a KMS Key: \"arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab\"
ID for the Amazon Web Services Key Management Service (KMS) key that Amazon Comprehend uses to encrypt data on the storage volume attached to the ML compute instance(s) that process the analysis job. The VolumeKmsKeyId can be either of the following formats:
KMS Key ID: \"1234abcd-12ab-34cd-56ef-1234567890ab\"
Amazon Resource Name (ARN) of a KMS Key: \"arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab\"
ID for the KMS key that Amazon Comprehend uses to encrypt the data on the storage volume attached to the ML compute instance(s) that process the targeted sentiment detection job. The VolumeKmsKeyId can be either of the following formats:
KMS Key ID: \"1234abcd-12ab-34cd-56ef-1234567890ab\"
Amazon Resource Name (ARN) of a KMS Key: \"arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab\"
ID for the AWS Key Management Service (KMS) key that Amazon Comprehend uses to encrypt data on the storage volume attached to the ML compute instance(s) that process the analysis job. The VolumeKmsKeyId can be either of the following formats:
KMS Key ID: \"1234abcd-12ab-34cd-56ef-1234567890ab\"
Amazon Resource Name (ARN) of a KMS Key: \"arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab\"
ID for the AWS Key Management Service (KMS) key that Amazon Comprehend uses to encrypt trained custom models. The ModelKmsKeyId can be either of the following formats:
KMS Key ID: \"1234abcd-12ab-34cd-56ef-1234567890ab\"
Amazon Resource Name (ARN) of a KMS Key: \"arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab\"
ID for the Amazon Web Services Key Management Service (KMS) key that Amazon Comprehend uses to encrypt data on the storage volume attached to the ML compute instance(s) that process the analysis job. The VolumeKmsKeyId can be either of the following formats:
KMS Key ID: \"1234abcd-12ab-34cd-56ef-1234567890ab\"
Amazon Resource Name (ARN) of a KMS Key: \"arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab\"
ID for the KMS key that Amazon Comprehend uses to encrypt trained custom models. The ModelKmsKeyId can be either of the following formats:
KMS Key ID: \"1234abcd-12ab-34cd-56ef-1234567890ab\"
Amazon Resource Name (ARN) of a KMS Key: \"arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab\"
ID for the KMS key that Amazon Comprehend uses to encrypt the volume.
" } }, @@ -2105,7 +2106,7 @@ "base": null, "refs": { "BatchDetectDominantLanguageItemResult$Languages": "One or more DominantLanguage objects describing the dominant languages in the document.
", - "DetectDominantLanguageResponse$Languages": "The languages that Amazon Comprehend detected in the input text. For each language, the response returns the RFC 5646 language code and the level of confidence that Amazon Comprehend has in the accuracy of its inference. For more information about RFC 5646, see Tags for Identifying Languages on the IETF Tools web site.
" + "DetectDominantLanguageResponse$Languages": "Array of languages that Amazon Comprehend detected in the input text. The array is sorted in descending order of the score (the dominant language is always the first element in the array).
For each language, the response returns the RFC 5646 language code and the level of confidence that Amazon Comprehend has in the accuracy of its inference. For more information about RFC 5646, see Tags for Identifying Languages on the IETF Tools web site.
" } }, "ListOfEntities": { @@ -2401,8 +2402,8 @@ "Policy": { "base": null, "refs": { - "CreateDocumentClassifierRequest$ModelPolicy": "The resource-based policy to attach to your custom document classifier model. You can use this policy to allow another AWS account to import your custom model.
Provide your policy as a JSON body that you enter as a UTF-8 encoded string without line breaks. To provide valid JSON, enclose the attribute names and values in double quotes. If the JSON body is also enclosed in double quotes, then you must escape the double quotes that are inside the policy:
\"{\\\"attribute\\\": \\\"value\\\", \\\"attribute\\\": [\\\"value\\\"]}\"
To avoid escaping quotes, you can use single quotes to enclose the policy and double quotes to enclose the JSON names and values:
'{\"attribute\": \"value\", \"attribute\": [\"value\"]}'
The JSON resource-based policy to attach to your custom entity recognizer model. You can use this policy to allow another AWS account to import your custom model.
Provide your JSON as a UTF-8 encoded string without line breaks. To provide valid JSON for your policy, enclose the attribute names and values in double quotes. If the JSON body is also enclosed in double quotes, then you must escape the double quotes that are inside the policy:
\"{\\\"attribute\\\": \\\"value\\\", \\\"attribute\\\": [\\\"value\\\"]}\"
To avoid escaping quotes, you can use single quotes to enclose the policy and double quotes to enclose the JSON names and values:
'{\"attribute\": \"value\", \"attribute\": [\"value\"]}'
The resource-based policy to attach to your custom document classifier model. You can use this policy to allow another Amazon Web Services account to import your custom model.
Provide your policy as a JSON body that you enter as a UTF-8 encoded string without line breaks. To provide valid JSON, enclose the attribute names and values in double quotes. If the JSON body is also enclosed in double quotes, then you must escape the double quotes that are inside the policy:
\"{\\\"attribute\\\": \\\"value\\\", \\\"attribute\\\": [\\\"value\\\"]}\"
To avoid escaping quotes, you can use single quotes to enclose the policy and double quotes to enclose the JSON names and values:
'{\"attribute\": \"value\", \"attribute\": [\"value\"]}'
The JSON resource-based policy to attach to your custom entity recognizer model. You can use this policy to allow another Amazon Web Services account to import your custom model.
Provide your JSON as a UTF-8 encoded string without line breaks. To provide valid JSON for your policy, enclose the attribute names and values in double quotes. If the JSON body is also enclosed in double quotes, then you must escape the double quotes that are inside the policy:
\"{\\\"attribute\\\": \\\"value\\\", \\\"attribute\\\": [\\\"value\\\"]}\"
To avoid escaping quotes, you can use single quotes to enclose the policy and double quotes to enclose the JSON names and values:
'{\"attribute\": \"value\", \"attribute\": [\"value\"]}'
The JSON body of the resource-based policy.
", "PutResourcePolicyRequest$ResourcePolicy": "The JSON resource-based policy to attach to your custom model. Provide your JSON as a UTF-8 encoded string without line breaks. To provide valid JSON for your policy, enclose the attribute names and values in double quotes. If the JSON body is also enclosed in double quotes, then you must escape the double quotes that are inside the policy:
\"{\\\"attribute\\\": \\\"value\\\", \\\"attribute\\\": [\\\"value\\\"]}\"
To avoid escaping quotes, you can use single quotes to enclose the policy and double quotes to enclose the JSON names and values:
'{\"attribute\": \"value\", \"attribute\": [\"value\"]}'
The Amazon S3 location of the augmented manifest file.
", "DatasetAugmentedManifestsListItem$AnnotationDataS3Uri": "The S3 prefix to the annotation files that are referred in the augmented manifest file.
", "DatasetAugmentedManifestsListItem$SourceDocumentsS3Uri": "The S3 prefix to the source files (PDFs) that are referred to in the augmented manifest file.
", - "DatasetDocumentClassifierInputDataConfig$S3Uri": "The Amazon S3 URI for the input data. The S3 bucket must be in the same region as the API endpoint that you are calling. The URI can point to a single input file or it can provide the prefix for a collection of input files.
For example, if you use the URI S3://bucketName/prefix
, if the prefix is a single file, Amazon Comprehend uses that file as input. If more than one file begins with the prefix, Amazon Comprehend uses all of them as input.
This parameter is required if you set DataFormat
to COMPREHEND_CSV
.
Specifies the Amazon S3 location where the training documents for an entity recognizer are located. The URI must be in the same region as the API endpoint that you are calling.
", + "DatasetDocumentClassifierInputDataConfig$S3Uri": "The Amazon S3 URI for the input data. The S3 bucket must be in the same Region as the API endpoint that you are calling. The URI can point to a single input file or it can provide the prefix for a collection of input files.
For example, if you use the URI S3://bucketName/prefix
, if the prefix is a single file, Amazon Comprehend uses that file as input. If more than one file begins with the prefix, Amazon Comprehend uses all of them as input.
This parameter is required if you set DataFormat
to COMPREHEND_CSV
.
Specifies the Amazon S3 location where the training documents for an entity recognizer are located. The URI must be in the same Region as the API endpoint that you are calling.
", "DatasetEntityRecognizerDocuments$S3Uri": "Specifies the Amazon S3 location where the documents for the dataset are located.
", "DatasetEntityRecognizerEntityList$S3Uri": "Specifies the Amazon S3 location where the entity list is located.
", "DatasetProperties$DatasetS3Uri": "The S3 URI where the dataset is stored.
", - "DocumentClassifierInputDataConfig$S3Uri": "The Amazon S3 URI for the input data. The S3 bucket must be in the same region as the API endpoint that you are calling. The URI can point to a single input file or it can provide the prefix for a collection of input files.
For example, if you use the URI S3://bucketName/prefix
, if the prefix is a single file, Amazon Comprehend uses that file as input. If more than one file begins with the prefix, Amazon Comprehend uses all of them as input.
This parameter is required if you set DataFormat
to COMPREHEND_CSV
.
This specifies the Amazon S3 location where the test annotations for an entity recognizer are located. The URI must be in the same AWS Region as the API endpoint that you are calling.
", - "DocumentClassifierOutputDataConfig$S3Uri": "When you use the OutputDataConfig
object while creating a custom classifier, you specify the Amazon S3 location where you want to write the confusion matrix. The URI must be in the same region as the API endpoint that you are calling. The location is used as the prefix for the actual location of this output file.
When the custom classifier job is finished, the service creates the output file in a directory specific to the job. The S3Uri
field contains the location of the output file, called output.tar.gz
. It is a compressed archive that contains the confusion matrix.
The Amazon S3 URI for the input data. The S3 bucket must be in the same Region as the API endpoint that you are calling. The URI can point to a single input file or it can provide the prefix for a collection of input files.
For example, if you use the URI S3://bucketName/prefix
, if the prefix is a single file, Amazon Comprehend uses that file as input. If more than one file begins with the prefix, Amazon Comprehend uses all of them as input.
This parameter is required if you set DataFormat
to COMPREHEND_CSV
.
This specifies the Amazon S3 location where the test annotations for an entity recognizer are located. The URI must be in the same Amazon Web Services Region as the API endpoint that you are calling.
", + "DocumentClassifierOutputDataConfig$S3Uri": "When you use the OutputDataConfig
object while creating a custom classifier, you specify the Amazon S3 location where you want to write the confusion matrix. The URI must be in the same Region as the API endpoint that you are calling. The location is used as the prefix for the actual location of this output file.
When the custom classifier job is finished, the service creates the output file in a directory specific to the job. The S3Uri
field contains the location of the output file, called output.tar.gz
. It is a compressed archive that contains the confusion matrix.
The Amazon S3 prefix for the data lake location of the flywheel statistics.
", - "EntityRecognizerAnnotations$S3Uri": "Specifies the Amazon S3 location where the annotations for an entity recognizer are located. The URI must be in the same region as the API endpoint that you are calling.
", - "EntityRecognizerAnnotations$TestS3Uri": "Specifies the Amazon S3 location where the test annotations for an entity recognizer are located. The URI must be in the same region as the API endpoint that you are calling.
", - "EntityRecognizerDocuments$S3Uri": "Specifies the Amazon S3 location where the training documents for an entity recognizer are located. The URI must be in the same region as the API endpoint that you are calling.
", - "EntityRecognizerDocuments$TestS3Uri": "Specifies the Amazon S3 location where the test documents for an entity recognizer are located. The URI must be in the same AWS Region as the API endpoint that you are calling.
", - "EntityRecognizerEntityList$S3Uri": "Specifies the Amazon S3 location where the entity list is located. The URI must be in the same region as the API endpoint that you are calling.
", + "EntityRecognizerAnnotations$S3Uri": "Specifies the Amazon S3 location where the annotations for an entity recognizer are located. The URI must be in the same Region as the API endpoint that you are calling.
", + "EntityRecognizerAnnotations$TestS3Uri": "Specifies the Amazon S3 location where the test annotations for an entity recognizer are located. The URI must be in the same Region as the API endpoint that you are calling.
", + "EntityRecognizerDocuments$S3Uri": "Specifies the Amazon S3 location where the training documents for an entity recognizer are located. The URI must be in the same Region as the API endpoint that you are calling.
", + "EntityRecognizerDocuments$TestS3Uri": "Specifies the Amazon S3 location where the test documents for an entity recognizer are located. The URI must be in the same Amazon Web Services Region as the API endpoint that you are calling.
", + "EntityRecognizerEntityList$S3Uri": "Specifies the Amazon S3 location where the entity list is located. The URI must be in the same Region as the API endpoint that you are calling.
", "EntityRecognizerOutputDataConfig$FlywheelStatsS3Prefix": "The Amazon S3 prefix for the data lake location of the flywheel statistics.
", "FlywheelIterationProperties$EvaluationManifestS3Prefix": "", "FlywheelProperties$DataLakeS3Uri": "Amazon S3 URI of the data lake location.
", "FlywheelSummary$DataLakeS3Uri": "Amazon S3 URI of the data lake location.
", - "InputDataConfig$S3Uri": "The Amazon S3 URI for the input data. The URI must be in same region as the API endpoint that you are calling. The URI can point to a single input file or it can provide the prefix for a collection of data files.
For example, if you use the URI S3://bucketName/prefix
, if the prefix is a single file, Amazon Comprehend uses that file as input. If more than one file begins with the prefix, Amazon Comprehend uses all of them as input.
When you use the OutputDataConfig
object with asynchronous operations, you specify the Amazon S3 location where you want to write the output data. The URI must be in the same region as the API endpoint that you are calling. The location is used as the prefix for the actual location of the output file.
When the topic detection job is finished, the service creates an output file in a directory specific to the job. The S3Uri
field contains the location of the output file, called output.tar.gz
. It is a compressed archive that contains the ouput of the operation.
For a PII entity detection job, the output file is plain text, not a compressed archive. The output file name is the same as the input file, with .out
appended at the end.
The Amazon S3 URI for the input data. The URI must be in same Region as the API endpoint that you are calling. The URI can point to a single input file or it can provide the prefix for a collection of data files.
For example, if you use the URI S3://bucketName/prefix
, if the prefix is a single file, Amazon Comprehend uses that file as input. If more than one file begins with the prefix, Amazon Comprehend uses all of them as input.
When you use the OutputDataConfig
object with asynchronous operations, you specify the Amazon S3 location where you want to write the output data. The URI must be in the same Region as the API endpoint that you are calling. The location is used as the prefix for the actual location of the output file.
When the topic detection job is finished, the service creates an output file in a directory specific to the job. The S3Uri
field contains the location of the output file, called output.tar.gz
. It is a compressed archive that contains the ouput of the operation.
For a PII entity detection job, the output file is plain text, not a compressed archive. The output file name is the same as the input file, with .out
appended at the end.
When you use the PiiOutputDataConfig
object with asynchronous operations, you specify the Amazon S3 location where you want to write the output data.
For a PII entity detection job, the output file is plain text, not a compressed archive. The output file name is the same as the input file, with .out
appended at the end.
The ID for each subnet being used in your private VPC. This subnet is a subset of the a range of IPv4 addresses used by the VPC and is specific to a given availability zone in the VPC’s region. This ID number is preceded by \"subnet-\", for instance: \"subnet-04ccf456919e69055\". For more information, see VPCs and Subnets.
" + "VpcConfig$Subnets": "The ID for each subnet being used in your private VPC. This subnet is a subset of the a range of IPv4 addresses used by the VPC and is specific to a given availability zone in the VPC’s Region. This ID number is preceded by \"subnet-\", for instance: \"subnet-04ccf456919e69055\". For more information, see VPCs and Subnets.
" } }, "SyntaxLanguageCode": { @@ -3122,13 +3123,13 @@ "VersionName": { "base": null, "refs": { - "CreateDocumentClassifierRequest$VersionName": "The version name given to the newly created classifier. Version names can have a maximum of 256 characters. Alphanumeric characters, hyphens (-) and underscores (_) are allowed. The version name must be unique among all models with the same classifier name in the account/AWS Region.
", - "CreateEntityRecognizerRequest$VersionName": "The version name given to the newly created recognizer. Version names can be a maximum of 256 characters. Alphanumeric characters, hyphens (-) and underscores (_) are allowed. The version name must be unique among all models with the same recognizer name in the account/ AWS Region.
", + "CreateDocumentClassifierRequest$VersionName": "The version name given to the newly created classifier. Version names can have a maximum of 256 characters. Alphanumeric characters, hyphens (-) and underscores (_) are allowed. The version name must be unique among all models with the same classifier name in the Amazon Web Services account/Amazon Web Services Region.
", + "CreateEntityRecognizerRequest$VersionName": "The version name given to the newly created recognizer. Version names can be a maximum of 256 characters. Alphanumeric characters, hyphens (-) and underscores (_) are allowed. The version name must be unique among all models with the same recognizer name in the account/Region.
", "DocumentClassifierProperties$VersionName": "The version name that you assigned to the document classifier.
", "DocumentClassifierSummary$LatestVersionName": "The version name you assigned to the latest document classifier version.
", "EntityRecognizerProperties$VersionName": "The version name you assigned to the entity recognizer.
", "EntityRecognizerSummary$LatestVersionName": "The version name you assigned to the latest entity recognizer version.
", - "ImportModelRequest$VersionName": "The version name given to the custom model that is created by this import. Version names can have a maximum of 256 characters. Alphanumeric characters, hyphens (-) and underscores (_) are allowed. The version name must be unique among all models with the same classifier name in the account/AWS Region.
" + "ImportModelRequest$VersionName": "The version name given to the custom model that is created by this import. Version names can have a maximum of 256 characters. Alphanumeric characters, hyphens (-) and underscores (_) are allowed. The version name must be unique among all models with the same classifier name in the account/Region.
" } }, "VpcConfig": { diff --git a/models/apis/comprehend/2017-11-27/endpoint-tests-1.json b/models/apis/comprehend/2017-11-27/endpoint-tests-1.json index e75ad3540f7..a63e830fa03 100644 --- a/models/apis/comprehend/2017-11-27/endpoint-tests-1.json +++ b/models/apis/comprehend/2017-11-27/endpoint-tests-1.json @@ -8,9 +8,9 @@ } }, "params": { - "Region": "ap-northeast-1", "UseFIPS": false, - "UseDualStack": false + "UseDualStack": false, + "Region": "ap-northeast-1" } }, { @@ -21,9 +21,9 @@ } }, "params": { - "Region": "ap-northeast-2", "UseFIPS": false, - "UseDualStack": false + "UseDualStack": false, + "Region": "ap-northeast-2" } }, { @@ -34,9 +34,9 @@ } }, "params": { - "Region": "ap-south-1", "UseFIPS": false, - "UseDualStack": false + "UseDualStack": false, + "Region": "ap-south-1" } }, { @@ -47,9 +47,9 @@ } }, "params": { - "Region": "ap-southeast-1", "UseFIPS": false, - "UseDualStack": false + "UseDualStack": false, + "Region": "ap-southeast-1" } }, { @@ -60,9 +60,9 @@ } }, "params": { - "Region": "ap-southeast-2", "UseFIPS": false, - "UseDualStack": false + "UseDualStack": false, + "Region": "ap-southeast-2" } }, { @@ -73,9 +73,9 @@ } }, "params": { - "Region": "ca-central-1", "UseFIPS": false, - "UseDualStack": false + "UseDualStack": false, + "Region": "ca-central-1" } }, { @@ -86,9 +86,9 @@ } }, "params": { - "Region": "eu-central-1", "UseFIPS": false, - "UseDualStack": false + "UseDualStack": false, + "Region": "eu-central-1" } }, { @@ -99,9 +99,9 @@ } }, "params": { - "Region": "eu-west-1", "UseFIPS": false, - "UseDualStack": false + "UseDualStack": false, + "Region": "eu-west-1" } }, { @@ -112,9 +112,9 @@ } }, "params": { - "Region": "eu-west-2", "UseFIPS": false, - "UseDualStack": false + "UseDualStack": false, + "Region": "eu-west-2" } }, { @@ -125,9 +125,9 @@ } }, "params": { - "Region": "us-east-1", "UseFIPS": false, - "UseDualStack": false + "UseDualStack": false, + "Region": "us-east-1" } }, { @@ -138,9 +138,9 @@ } }, "params": { - "Region": "us-east-1", "UseFIPS": true, - "UseDualStack": false + "UseDualStack": false, + "Region": "us-east-1" } }, { @@ -151,9 +151,9 @@ } }, "params": { - "Region": "us-east-2", "UseFIPS": false, - "UseDualStack": false + "UseDualStack": false, + "Region": "us-east-2" } }, { @@ -164,9 +164,9 @@ } }, "params": { - "Region": "us-east-2", "UseFIPS": true, - "UseDualStack": false + "UseDualStack": false, + "Region": "us-east-2" } }, { @@ -177,9 +177,9 @@ } }, "params": { - "Region": "us-west-2", "UseFIPS": false, - "UseDualStack": false + "UseDualStack": false, + "Region": "us-west-2" } }, { @@ -190,9 +190,9 @@ } }, "params": { - "Region": "us-west-2", "UseFIPS": true, - "UseDualStack": false + "UseDualStack": false, + "Region": "us-west-2" } }, { @@ -203,9 +203,9 @@ } }, "params": { - "Region": "us-east-1", "UseFIPS": true, - "UseDualStack": true + "UseDualStack": true, + "Region": "us-east-1" } }, { @@ -216,9 +216,9 @@ } }, "params": { - "Region": "us-east-1", "UseFIPS": false, - "UseDualStack": true + "UseDualStack": true, + "Region": "us-east-1" } }, { @@ -229,9 +229,9 @@ } }, "params": { - "Region": "cn-north-1", "UseFIPS": true, - "UseDualStack": true + "UseDualStack": true, + "Region": "cn-north-1" } }, { @@ -242,9 +242,9 @@ } }, "params": { - "Region": "cn-north-1", "UseFIPS": true, - "UseDualStack": false + "UseDualStack": false, + "Region": "cn-north-1" } }, { @@ -255,9 +255,9 @@ } }, "params": { - "Region": "cn-north-1", "UseFIPS": false, - "UseDualStack": true + "UseDualStack": true, + "Region": "cn-north-1" } }, { @@ -268,9 +268,9 @@ } }, "params": { - "Region": "cn-north-1", "UseFIPS": false, - "UseDualStack": false + "UseDualStack": false, + "Region": "cn-north-1" } }, { @@ -281,9 +281,9 @@ } }, "params": { - "Region": "us-gov-west-1", "UseFIPS": false, - "UseDualStack": false + "UseDualStack": false, + "Region": "us-gov-west-1" } }, { @@ -294,9 +294,9 @@ } }, "params": { - "Region": "us-gov-west-1", "UseFIPS": true, - "UseDualStack": false + "UseDualStack": false, + "Region": "us-gov-west-1" } }, { @@ -307,9 +307,9 @@ } }, "params": { - "Region": "us-gov-east-1", "UseFIPS": true, - "UseDualStack": true + "UseDualStack": true, + "Region": "us-gov-east-1" } }, { @@ -320,9 +320,9 @@ } }, "params": { - "Region": "us-gov-east-1", "UseFIPS": true, - "UseDualStack": false + "UseDualStack": false, + "Region": "us-gov-east-1" } }, { @@ -333,9 +333,9 @@ } }, "params": { - "Region": "us-gov-east-1", "UseFIPS": false, - "UseDualStack": true + "UseDualStack": true, + "Region": "us-gov-east-1" } }, { @@ -346,9 +346,9 @@ } }, "params": { - "Region": "us-gov-east-1", "UseFIPS": false, - "UseDualStack": false + "UseDualStack": false, + "Region": "us-gov-east-1" } }, { @@ -359,9 +359,20 @@ } }, "params": { - "Region": "us-iso-east-1", "UseFIPS": false, - "UseDualStack": false + "UseDualStack": false, + "Region": "us-iso-east-1" + } + }, + { + "documentation": "For region us-iso-east-1 with FIPS enabled and DualStack enabled", + "expect": { + "error": "FIPS and DualStack are enabled, but this partition does not support one or both" + }, + "params": { + "UseFIPS": true, + "UseDualStack": true, + "Region": "us-iso-east-1" } }, { @@ -372,9 +383,31 @@ } }, "params": { - "Region": "us-iso-east-1", "UseFIPS": true, - "UseDualStack": false + "UseDualStack": false, + "Region": "us-iso-east-1" + } + }, + { + "documentation": "For region us-iso-east-1 with FIPS disabled and DualStack enabled", + "expect": { + "error": "DualStack is enabled but this partition does not support DualStack" + }, + "params": { + "UseFIPS": false, + "UseDualStack": true, + "Region": "us-iso-east-1" + } + }, + { + "documentation": "For region us-isob-east-1 with FIPS enabled and DualStack enabled", + "expect": { + "error": "FIPS and DualStack are enabled, but this partition does not support one or both" + }, + "params": { + "UseFIPS": true, + "UseDualStack": true, + "Region": "us-isob-east-1" } }, { @@ -385,9 +418,20 @@ } }, "params": { - "Region": "us-isob-east-1", "UseFIPS": true, - "UseDualStack": false + "UseDualStack": false, + "Region": "us-isob-east-1" + } + }, + { + "documentation": "For region us-isob-east-1 with FIPS disabled and DualStack enabled", + "expect": { + "error": "DualStack is enabled but this partition does not support DualStack" + }, + "params": { + "UseFIPS": false, + "UseDualStack": true, + "Region": "us-isob-east-1" } }, { @@ -398,9 +442,9 @@ } }, "params": { - "Region": "us-isob-east-1", "UseFIPS": false, - "UseDualStack": false + "UseDualStack": false, + "Region": "us-isob-east-1" } }, { @@ -411,9 +455,9 @@ } }, "params": { - "Region": "us-east-1", "UseFIPS": false, "UseDualStack": false, + "Region": "us-east-1", "Endpoint": "https://example.com" } }, @@ -436,9 +480,9 @@ "error": "Invalid Configuration: FIPS and custom endpoint are not supported" }, "params": { - "Region": "us-east-1", "UseFIPS": true, "UseDualStack": false, + "Region": "us-east-1", "Endpoint": "https://example.com" } }, @@ -448,11 +492,17 @@ "error": "Invalid Configuration: Dualstack and custom endpoint are not supported" }, "params": { - "Region": "us-east-1", "UseFIPS": false, "UseDualStack": true, + "Region": "us-east-1", "Endpoint": "https://example.com" } + }, + { + "documentation": "Missing region", + "expect": { + "error": "Invalid Configuration: Missing Region" + } } ], "version": "1.0" diff --git a/models/apis/monitoring/2010-08-01/docs-2.json b/models/apis/monitoring/2010-08-01/docs-2.json index 83cce3ee46d..6121c7a7d10 100644 --- a/models/apis/monitoring/2010-08-01/docs-2.json +++ b/models/apis/monitoring/2010-08-01/docs-2.json @@ -2,7 +2,7 @@ "version": "2.0", "service": "Amazon CloudWatch monitors your Amazon Web Services (Amazon Web Services) resources and the applications you run on Amazon Web Services in real time. You can use CloudWatch to collect and track metrics, which are the variables you want to measure for your resources and applications.
CloudWatch alarms send notifications or automatically change the resources you are monitoring based on rules that you define. For example, you can monitor the CPU usage and disk reads and writes of your Amazon EC2 instances. Then, use this data to determine whether you should launch additional instances to handle increased load. You can also use this data to stop under-used instances to save money.
In addition to monitoring the built-in metrics that come with Amazon Web Services, you can monitor your own custom metrics. With CloudWatch, you gain system-wide visibility into resource utilization, application performance, and operational health.
", "operations": { - "DeleteAlarms": "Deletes the specified alarms. You can delete up to 100 alarms in one operation. However, this total can include no more than one composite alarm. For example, you could delete 99 metric alarms and one composite alarms with one operation, but you can't delete two composite alarms with one operation.
In the event of an error, no alarms are deleted.
It is possible to create a loop or cycle of composite alarms, where composite alarm A depends on composite alarm B, and composite alarm B also depends on composite alarm A. In this scenario, you can't delete any composite alarm that is part of the cycle because there is always still a composite alarm that depends on that alarm that you want to delete.
To get out of such a situation, you must break the cycle by changing the rule of one of the composite alarms in the cycle to remove a dependency that creates the cycle. The simplest change to make to break a cycle is to change the AlarmRule
of one of the alarms to false
.
Additionally, the evaluation of composite alarms stops if CloudWatch detects a cycle in the evaluation path.
Deletes the specified alarms. You can delete up to 100 alarms in one operation. However, this total can include no more than one composite alarm. For example, you could delete 99 metric alarms and one composite alarms with one operation, but you can't delete two composite alarms with one operation.
If you specify an incorrect alarm name or make any other error in the operation, no alarms are deleted. To confirm that alarms were deleted successfully, you can use the DescribeAlarms operation after using DeleteAlarms
.
It is possible to create a loop or cycle of composite alarms, where composite alarm A depends on composite alarm B, and composite alarm B also depends on composite alarm A. In this scenario, you can't delete any composite alarm that is part of the cycle because there is always still a composite alarm that depends on that alarm that you want to delete.
To get out of such a situation, you must break the cycle by changing the rule of one of the composite alarms in the cycle to remove a dependency that creates the cycle. The simplest change to make to break a cycle is to change the AlarmRule
of one of the alarms to false
.
Additionally, the evaluation of composite alarms stops if CloudWatch detects a cycle in the evaluation path.
Deletes the specified anomaly detection model from your account. For more information about how to delete an anomaly detection model, see Deleting an anomaly detection model in the CloudWatch User Guide.
", "DeleteDashboards": "Deletes all dashboards that you specify. You can specify up to 100 dashboards to delete. If there is an error during this call, no dashboards are deleted.
", "DeleteInsightRules": "Permanently deletes the specified Contributor Insights rules.
If you create a rule, delete it, and then re-create it with the same name, historical data from the first time the rule was created might not be available.
", @@ -32,7 +32,7 @@ "PutDashboard": "Creates a dashboard if it does not already exist, or updates an existing dashboard. If you update a dashboard, the entire contents are replaced with what you specify here.
All dashboards in your account are global, not region-specific.
A simple way to create a dashboard using PutDashboard
is to copy an existing dashboard. To copy an existing dashboard using the console, you can load the dashboard and then use the View/edit source command in the Actions menu to display the JSON block for that dashboard. Another way to copy a dashboard is to use GetDashboard
, and then use the data returned within DashboardBody
as the template for the new dashboard when you call PutDashboard
.
When you create a dashboard with PutDashboard
, a good practice is to add a text widget at the top of the dashboard with a message that the dashboard was created by script and should not be changed in the console. This message could also point console users to the location of the DashboardBody
script or the CloudFormation template used to create the dashboard.
Creates a Contributor Insights rule. Rules evaluate log events in a CloudWatch Logs log group, enabling you to find contributor data for the log events in that log group. For more information, see Using Contributor Insights to Analyze High-Cardinality Data.
If you create a rule, delete it, and then re-create it with the same name, historical data from the first time the rule was created might not be available.
", "PutManagedInsightRules": " Creates a managed Contributor Insights rule for a specified Amazon Web Services resource. When you enable a managed rule, you create a Contributor Insights rule that collects data from Amazon Web Services services. You cannot edit these rules with PutInsightRule
. The rules can be enabled, disabled, and deleted using EnableInsightRules
, DisableInsightRules
, and DeleteInsightRules
. If a previously created managed rule is currently disabled, a subsequent call to this API will re-enable it. Use ListManagedInsightRules
to describe all available rules.
Creates or updates an alarm and associates it with the specified metric, metric math expression, anomaly detection model, or Metrics Insights query. For more information about using a Metrics Insights query for an alarm, see Create alarms on Metrics Insights queries.
Alarms based on anomaly detection models cannot have Auto Scaling actions.
When this operation creates an alarm, the alarm state is immediately set to INSUFFICIENT_DATA
. The alarm is then evaluated and its state is set appropriately. Any actions associated with the new state are then executed.
When you update an existing alarm, its state is left unchanged, but the update completely overwrites the previous configuration of the alarm.
If you are an IAM user, you must have Amazon EC2 permissions for some alarm operations:
The iam:CreateServiceLinkedRole
for all alarms with EC2 actions
The iam:CreateServiceLinkedRole
to create an alarm with Systems Manager OpsItem actions.
The first time you create an alarm in the Amazon Web Services Management Console, the CLI, or by using the PutMetricAlarm API, CloudWatch creates the necessary service-linked role for you. The service-linked roles are called AWSServiceRoleForCloudWatchEvents
and AWSServiceRoleForCloudWatchAlarms_ActionSSM
. For more information, see Amazon Web Services service-linked role.
Cross-account alarms
You can set an alarm on metrics in the current account, or in another account. To create a cross-account alarm that watches a metric in a different account, you must have completed the following pre-requisites:
The account where the metrics are located (the sharing account) must already have a sharing role named CloudWatch-CrossAccountSharingRole. If it does not already have this role, you must create it using the instructions in Set up a sharing account in Cross-account cross-Region CloudWatch console. The policy for that role must grant access to the ID of the account where you are creating the alarm.
The account where you are creating the alarm (the monitoring account) must already have a service-linked role named AWSServiceRoleForCloudWatchCrossAccount to allow CloudWatch to assume the sharing role in the sharing account. If it does not, you must create it following the directions in Set up a monitoring account in Cross-account cross-Region CloudWatch console.
Creates or updates an alarm and associates it with the specified metric, metric math expression, anomaly detection model, or Metrics Insights query. For more information about using a Metrics Insights query for an alarm, see Create alarms on Metrics Insights queries.
Alarms based on anomaly detection models cannot have Auto Scaling actions.
When this operation creates an alarm, the alarm state is immediately set to INSUFFICIENT_DATA
. The alarm is then evaluated and its state is set appropriately. Any actions associated with the new state are then executed.
When you update an existing alarm, its state is left unchanged, but the update completely overwrites the previous configuration of the alarm.
If you are an IAM user, you must have Amazon EC2 permissions for some alarm operations:
The iam:CreateServiceLinkedRole
permission for all alarms with EC2 actions
The iam:CreateServiceLinkedRole
permissions to create an alarm with Systems Manager OpsItem or response plan actions.
The first time you create an alarm in the Amazon Web Services Management Console, the CLI, or by using the PutMetricAlarm API, CloudWatch creates the necessary service-linked role for you. The service-linked roles are called AWSServiceRoleForCloudWatchEvents
and AWSServiceRoleForCloudWatchAlarms_ActionSSM
. For more information, see Amazon Web Services service-linked role.
Cross-account alarms
You can set an alarm on metrics in the current account, or in another account. To create a cross-account alarm that watches a metric in a different account, you must have completed the following pre-requisites:
The account where the metrics are located (the sharing account) must already have a sharing role named CloudWatch-CrossAccountSharingRole. If it does not already have this role, you must create it using the instructions in Set up a sharing account in Cross-account cross-Region CloudWatch console. The policy for that role must grant access to the ID of the account where you are creating the alarm.
The account where you are creating the alarm (the monitoring account) must already have a service-linked role named AWSServiceRoleForCloudWatchCrossAccount to allow CloudWatch to assume the sharing role in the sharing account. If it does not, you must create it following the directions in Set up a monitoring account in Cross-account cross-Region CloudWatch console.
Publishes metric data points to Amazon CloudWatch. CloudWatch associates the data points with the specified metric. If the specified metric does not exist, CloudWatch creates the metric. When CloudWatch creates a metric, it can take up to fifteen minutes for the metric to appear in calls to ListMetrics.
You can publish either individual data points in the Value
field, or arrays of values and the number of times each value occurred during the period by using the Values
and Counts
fields in the MetricDatum
structure. Using the Values
and Counts
method enables you to publish up to 150 values per metric with one PutMetricData
request, and supports retrieving percentile statistics on this data.
Each PutMetricData
request is limited to 1 MB in size for HTTP POST requests. You can send a payload compressed by gzip. Each request is also limited to no more than 1000 different metrics.
Although the Value
parameter accepts numbers of type Double
, CloudWatch rejects values that are either too small or too large. Values must be in the range of -2^360 to 2^360. In addition, special values (for example, NaN, +Infinity, -Infinity) are not supported.
You can use up to 30 dimensions per metric to further clarify what data the metric collects. Each dimension consists of a Name and Value pair. For more information about specifying dimensions, see Publishing Metrics in the Amazon CloudWatch User Guide.
You specify the time stamp to be associated with each data point. You can specify time stamps that are as much as two weeks before the current date, and as much as 2 hours after the current day and time.
Data points with time stamps from 24 hours ago or longer can take at least 48 hours to become available for GetMetricData or GetMetricStatistics from the time they are submitted. Data points with time stamps between 3 and 24 hours ago can take as much as 2 hours to become available for for GetMetricData or GetMetricStatistics.
CloudWatch needs raw data points to calculate percentile statistics. If you publish data using a statistic set instead, you can only retrieve percentile statistics for this data if one of the following conditions is true:
The SampleCount
value of the statistic set is 1 and Min
, Max
, and Sum
are all equal.
The Min
and Max
are equal, and Sum
is equal to Min
multiplied by SampleCount
.
Creates or updates a metric stream. Metric streams can automatically stream CloudWatch metrics to Amazon Web Services destinations, including Amazon S3, and to many third-party solutions.
For more information, see Using Metric Streams.
To create a metric stream, you must be signed in to an account that has the iam:PassRole
permission and either the CloudWatchFullAccess
policy or the cloudwatch:PutMetricStream
permission.
When you create or update a metric stream, you choose one of the following:
Stream metrics from all metric namespaces in the account.
Stream metrics from all metric namespaces in the account, except for the namespaces that you list in ExcludeFilters
.
Stream metrics from only the metric namespaces that you list in IncludeFilters
.
By default, a metric stream always sends the MAX
, MIN
, SUM
, and SAMPLECOUNT
statistics for each metric that is streamed. You can use the StatisticsConfigurations
parameter to have the metric stream send additional statistics in the stream. Streaming additional statistics incurs additional costs. For more information, see Amazon CloudWatch Pricing.
When you use PutMetricStream
to create a new metric stream, the stream is created in the running
state. If you use it to update an existing stream, the state of the stream is not changed.
If you are using CloudWatch cross-account observability and you create a metric stream in a monitoring account, you can choose whether to include metrics from source accounts in the stream. For more information, see CloudWatch cross-account observability.
", "SetAlarmState": "Temporarily sets the state of an alarm for testing purposes. When the updated state differs from the previous value, the action configured for the appropriate state is invoked. For example, if your alarm is configured to send an Amazon SNS message when an alarm is triggered, temporarily changing the alarm state to ALARM
sends an SNS message.
Metric alarms returns to their actual state quickly, often within seconds. Because the metric alarm state change happens quickly, it is typically only visible in the alarm's History tab in the Amazon CloudWatch console or through DescribeAlarmHistory.
If you use SetAlarmState
on a composite alarm, the composite alarm is not guaranteed to return to its actual state. It returns to its actual state only once any of its children alarms change state. It is also reevaluated if you update its configuration.
If an alarm triggers EC2 Auto Scaling policies or application Auto Scaling policies, you must include information in the StateReasonData
parameter to enable the policy to take the correct action.
If you use this parameter and specify the name of a metric or composite alarm, the operation returns information about the \"parent\" alarms of the alarm you specify. These are the composite alarms that have AlarmRule
parameters that reference the alarm named in ParentsOfAlarmName
. Information about the alarm that you specify in ParentsOfAlarmName
is not returned.
If you specify ParentsOfAlarmName
, you cannot specify any other parameters in the request except for MaxRecords
and NextToken
. If you do so, you receive a validation error.
Only the Alarm Name and ARN are returned by this operation when you use this parameter. To get complete information about these alarms, perform another DescribeAlarms
operation and specify the parent alarm names in the AlarmNames
parameter.
The name of the alarm.
", "PutCompositeAlarmInput$AlarmName": "The name for the composite alarm. This name must be unique within the Region.
", - "PutMetricAlarmInput$AlarmName": "The name for the alarm. This name must be unique within the Region.
", + "PutMetricAlarmInput$AlarmName": "The name for the alarm. This name must be unique within the Region.
The name must contain only UTF-8 characters, and can't contain ASCII control characters
", "SetAlarmStateInput$AlarmName": "The name of the alarm.
" } }, @@ -131,7 +131,7 @@ "AlarmNames": { "base": null, "refs": { - "DeleteAlarmsInput$AlarmNames": "The alarms to be deleted.
", + "DeleteAlarmsInput$AlarmNames": "The alarms to be deleted. Do not enclose the alarm names in quote marks.
", "DescribeAlarmsInput$AlarmNames": "The names of the alarms to retrieve information about.
", "DisableAlarmActionsInput$AlarmNames": "The names of the alarms.
", "EnableAlarmActionsInput$AlarmNames": "The names of the alarms.
" @@ -532,14 +532,14 @@ "DimensionName": { "base": null, "refs": { - "Dimension$Name": "The name of the dimension. Dimension names must contain only ASCII characters, must include at least one non-whitespace character, and cannot start with a colon (:
).
The name of the dimension. Dimension names must contain only ASCII characters, must include at least one non-whitespace character, and cannot start with a colon (:
). ASCII control characters are not supported as part of dimension names.
The dimension name to be matched.
" } }, "DimensionValue": { "base": null, "refs": { - "Dimension$Value": "The value of the dimension. Dimension values must contain only ASCII characters and must include at least one non-whitespace character.
", + "Dimension$Value": "The value of the dimension. Dimension values must contain only ASCII characters and must include at least one non-whitespace character. ASCII control characters are not supported as part of dimension values.
", "DimensionFilter$Value": "The value of the dimension to be matched.
" } }, @@ -553,7 +553,7 @@ "GetMetricStatisticsInput$Dimensions": "The dimensions. If the metric contains multiple dimensions, you must include a value for each dimension. CloudWatch treats each unique combination of dimensions as a separate metric. If a specific combination of dimensions was not published, you can't retrieve statistics for it. You must specify the same dimensions that were used when the metrics were created. For an example, see Dimension Combinations in the Amazon CloudWatch User Guide. For more information about specifying dimensions, see Publishing Metrics in the Amazon CloudWatch User Guide.
", "Metric$Dimensions": "The dimensions for the metric.
", "MetricAlarm$Dimensions": "The dimensions for the metric associated with the alarm.
", - "MetricDatum$Dimensions": "The dimensions associated with the metric.
", + "MetricDatum$Dimensions": "The dimensions associated with the metric.
", "PutAnomalyDetectorInput$Dimensions": "The metric dimensions to create the anomaly detection model for.
", "PutMetricAlarmInput$Dimensions": "The dimensions for the metric specified in MetricName
.
The metric dimensions to create the anomaly detection model for.
" @@ -1247,7 +1247,7 @@ } }, "MetricStreamFilter": { - "base": "This structure contains the name of one of the metric namespaces that is listed in a filter of a metric stream.
", + "base": "This structure contains the name of one of the metric namespaces that is listed in a filter of a metric stream.
The namespace can contain only ASCII printable characters (ASCII range 32 through 126). It must contain at least one non-whitespace character.
", "refs": { "MetricStreamFilters$member": null } @@ -1369,7 +1369,7 @@ "MetricStreamStatisticsMetric$Namespace": "The namespace of the metric.
", "PutAnomalyDetectorInput$Namespace": "The namespace of the metric to create the anomaly detection model for.
", "PutMetricAlarmInput$Namespace": "The namespace for the metric associated specified in MetricName
.
The namespace for the metric data.
To avoid conflicts with Amazon Web Services service namespaces, you should not specify a namespace that begins with AWS/
The namespace for the metric data. You can use ASCII characters for the namespace, except for control characters which are not supported.
To avoid conflicts with Amazon Web Services service namespaces, you should not specify a namespace that begins with AWS/
The namespace of the metric to create the anomaly detection model for.
" } }, @@ -1521,9 +1521,9 @@ "PutCompositeAlarmInput$AlarmActions": "The actions to execute when this alarm transitions to the ALARM
state from any other state. Each action is specified as an Amazon Resource Name (ARN).
Valid Values: arn:aws:sns:region:account-id:sns-topic-name
| arn:aws:ssm:region:account-id:opsitem:severity
The actions to execute when this alarm transitions to the INSUFFICIENT_DATA
state from any other state. Each action is specified as an Amazon Resource Name (ARN).
Valid Values: arn:aws:sns:region:account-id:sns-topic-name
The actions to execute when this alarm transitions to an OK
state from any other state. Each action is specified as an Amazon Resource Name (ARN).
Valid Values: arn:aws:sns:region:account-id:sns-topic-name
The actions to execute when this alarm transitions to an OK
state from any other state. Each action is specified as an Amazon Resource Name (ARN).
Valid Values: arn:aws:automate:region:ec2:stop
| arn:aws:automate:region:ec2:terminate
| arn:aws:automate:region:ec2:recover
| arn:aws:automate:region:ec2:reboot
| arn:aws:sns:region:account-id:sns-topic-name
| arn:aws:autoscaling:region:account-id:scalingPolicy:policy-id:autoScalingGroupName/group-friendly-name:policyName/policy-friendly-name
Valid Values (for use with IAM roles): arn:aws:swf:region:account-id:action/actions/AWS_EC2.InstanceId.Stop/1.0
| arn:aws:swf:region:account-id:action/actions/AWS_EC2.InstanceId.Terminate/1.0
| arn:aws:swf:region:account-id:action/actions/AWS_EC2.InstanceId.Reboot/1.0
| arn:aws:swf:region:account-id:action/actions/AWS_EC2.InstanceId.Recover/1.0
The actions to execute when this alarm transitions to the ALARM
state from any other state. Each action is specified as an Amazon Resource Name (ARN).
Valid Values: arn:aws:automate:region:ec2:stop
| arn:aws:automate:region:ec2:terminate
| arn:aws:automate:region:ec2:recover
| arn:aws:automate:region:ec2:reboot
| arn:aws:sns:region:account-id:sns-topic-name
| arn:aws:autoscaling:region:account-id:scalingPolicy:policy-id:autoScalingGroupName/group-friendly-name:policyName/policy-friendly-name
| arn:aws:ssm:region:account-id:opsitem:severity
| arn:aws:ssm-incidents::account-id:response-plan:response-plan-name
Valid Values (for use with IAM roles): arn:aws:swf:region:account-id:action/actions/AWS_EC2.InstanceId.Stop/1.0
| arn:aws:swf:region:account-id:action/actions/AWS_EC2.InstanceId.Terminate/1.0
| arn:aws:swf:region:account-id:action/actions/AWS_EC2.InstanceId.Reboot/1.0
| arn:aws:swf:region:account-id:action/actions/AWS_EC2.InstanceId.Recover/1.0
The actions to execute when this alarm transitions to the INSUFFICIENT_DATA
state from any other state. Each action is specified as an Amazon Resource Name (ARN).
Valid Values: arn:aws:automate:region:ec2:stop
| arn:aws:automate:region:ec2:terminate
| arn:aws:automate:region:ec2:recover
| arn:aws:automate:region:ec2:reboot
| arn:aws:sns:region:account-id:sns-topic-name
| arn:aws:autoscaling:region:account-id:scalingPolicy:policy-id:autoScalingGroupName/group-friendly-name:policyName/policy-friendly-name
Valid Values (for use with IAM roles): >arn:aws:swf:region:account-id:action/actions/AWS_EC2.InstanceId.Stop/1.0
| arn:aws:swf:region:account-id:action/actions/AWS_EC2.InstanceId.Terminate/1.0
| arn:aws:swf:region:account-id:action/actions/AWS_EC2.InstanceId.Reboot/1.0
The actions to execute when this alarm transitions to an OK
state from any other state. Each action is specified as an Amazon Resource Name (ARN). Valid values:
EC2 actions:
arn:aws:automate:region:ec2:stop
arn:aws:automate:region:ec2:terminate
arn:aws:automate:region:ec2:reboot
arn:aws:automate:region:ec2:recover
arn:aws:swf:region:account-id:action/actions/AWS_EC2.InstanceId.Stop/1.0
arn:aws:swf:region:account-id:action/actions/AWS_EC2.InstanceId.Terminate/1.0
arn:aws:swf:region:account-id:action/actions/AWS_EC2.InstanceId.Reboot/1.0
arn:aws:swf:region:account-id:action/actions/AWS_EC2.InstanceId.Recover/1.0
Autoscaling action:
arn:aws:autoscaling:region:account-id:scalingPolicy:policy-id:autoScalingGroupName/group-friendly-name:policyName/policy-friendly-name
SSN notification action:
arn:aws:sns:region:account-id:sns-topic-name:autoScalingGroupName/group-friendly-name:policyName/policy-friendly-name
SSM integration actions:
arn:aws:ssm:region:account-id:opsitem:severity#CATEGORY=category-name
arn:aws:ssm-incidents::account-id:responseplan/response-plan-name
The actions to execute when this alarm transitions to the ALARM
state from any other state. Each action is specified as an Amazon Resource Name (ARN). Valid values:
EC2 actions:
arn:aws:automate:region:ec2:stop
arn:aws:automate:region:ec2:terminate
arn:aws:automate:region:ec2:reboot
arn:aws:automate:region:ec2:recover
arn:aws:swf:region:account-id:action/actions/AWS_EC2.InstanceId.Stop/1.0
arn:aws:swf:region:account-id:action/actions/AWS_EC2.InstanceId.Terminate/1.0
arn:aws:swf:region:account-id:action/actions/AWS_EC2.InstanceId.Reboot/1.0
arn:aws:swf:region:account-id:action/actions/AWS_EC2.InstanceId.Recover/1.0
Autoscaling action:
arn:aws:autoscaling:region:account-id:scalingPolicy:policy-id:autoScalingGroupName/group-friendly-name:policyName/policy-friendly-name
SSN notification action:
arn:aws:sns:region:account-id:sns-topic-name:autoScalingGroupName/group-friendly-name:policyName/policy-friendly-name
SSM integration actions:
arn:aws:ssm:region:account-id:opsitem:severity#CATEGORY=category-name
arn:aws:ssm-incidents::account-id:responseplan/response-plan-name
The actions to execute when this alarm transitions to the INSUFFICIENT_DATA
state from any other state. Each action is specified as an Amazon Resource Name (ARN). Valid values:
EC2 actions:
arn:aws:automate:region:ec2:stop
arn:aws:automate:region:ec2:terminate
arn:aws:automate:region:ec2:reboot
arn:aws:automate:region:ec2:recover
arn:aws:swf:region:account-id:action/actions/AWS_EC2.InstanceId.Stop/1.0
arn:aws:swf:region:account-id:action/actions/AWS_EC2.InstanceId.Terminate/1.0
arn:aws:swf:region:account-id:action/actions/AWS_EC2.InstanceId.Reboot/1.0
arn:aws:swf:region:account-id:action/actions/AWS_EC2.InstanceId.Recover/1.0
Autoscaling action:
arn:aws:autoscaling:region:account-id:scalingPolicy:policy-id:autoScalingGroupName/group-friendly-name:policyName/policy-friendly-name
SSN notification action:
arn:aws:sns:region:account-id:sns-topic-name:autoScalingGroupName/group-friendly-name:policyName/policy-friendly-name
SSM integration actions:
arn:aws:ssm:region:account-id:opsitem:severity#CATEGORY=category-name
arn:aws:ssm-incidents::account-id:responseplan/response-plan-name
Copies the specified option group.
", "CreateBlueGreenDeployment": "Creates a blue/green deployment.
A blue/green deployment creates a staging environment that copies the production environment. In a blue/green deployment, the blue environment is the current production environment. The green environment is the staging environment. The staging environment stays in sync with the current production environment using logical replication.
You can make changes to the databases in the green environment without affecting production workloads. For example, you can upgrade the major or minor DB engine version, change database parameters, or make schema changes in the staging environment. You can thoroughly test changes in the green environment. When ready, you can switch over the environments to promote the green environment to be the new production environment. The switchover typically takes under a minute.
For more information, see Using Amazon RDS Blue/Green Deployments for database updates in the Amazon RDS User Guide and Using Amazon RDS Blue/Green Deployments for database updates in the Amazon Aurora User Guide.
", "CreateCustomDBEngineVersion": "Creates a custom DB engine version (CEV).
", - "CreateDBCluster": "Creates a new Amazon Aurora DB cluster or Multi-AZ DB cluster.
You can use the ReplicationSourceIdentifier
parameter to create an Amazon Aurora DB cluster as a read replica of another DB cluster or Amazon RDS MySQL or PostgreSQL DB instance. For more information about Amazon Aurora, see What is Amazon Aurora? in the Amazon Aurora User Guide.
You can also use the ReplicationSourceIdentifier
parameter to create a Multi-AZ DB cluster read replica with an RDS for PostgreSQL DB instance as the source. For more information about Multi-AZ DB clusters, see Multi-AZ DB cluster deployments in the Amazon RDS User Guide.
Creates a new Amazon Aurora DB cluster or Multi-AZ DB cluster.
If you create an Aurora DB cluster, the request creates an empty cluster. You must explicitly create the writer instance for your DB cluster using the CreateDBInstance operation. If you create a Multi-AZ DB cluster, the request creates a writer and two reader DB instances for you, each in a different Availability Zone.
You can use the ReplicationSourceIdentifier
parameter to create an Amazon Aurora DB cluster as a read replica of another DB cluster or Amazon RDS MySQL or PostgreSQL DB instance. For more information about Amazon Aurora, see What is Amazon Aurora? in the Amazon Aurora User Guide.
You can also use the ReplicationSourceIdentifier
parameter to create a Multi-AZ DB cluster read replica with an RDS for PostgreSQL DB instance as the source. For more information about Multi-AZ DB clusters, see Multi-AZ DB cluster deployments in the Amazon RDS User Guide.
Creates a new custom endpoint and associates it with an Amazon Aurora DB cluster.
This action applies only to Aurora DB clusters.
Creates a new DB cluster parameter group.
Parameters in a DB cluster parameter group apply to all of the instances in a DB cluster.
A DB cluster parameter group is initially created with the default parameters for the database engine used by instances in the DB cluster. To provide custom values for any of the parameters, you must modify the group after creating it using ModifyDBClusterParameterGroup
. Once you've created a DB cluster parameter group, you need to associate it with your DB cluster using ModifyDBCluster
.
When you associate a new DB cluster parameter group with a running Aurora DB cluster, reboot the DB instances in the DB cluster without failover for the new DB cluster parameter group and associated settings to take effect.
When you associate a new DB cluster parameter group with a running Multi-AZ DB cluster, reboot the DB cluster without failover for the new DB cluster parameter group and associated settings to take effect.
After you create a DB cluster parameter group, you should wait at least 5 minutes before creating your first DB cluster that uses that DB cluster parameter group as the default parameter group. This allows Amazon RDS to fully complete the create action before the DB cluster parameter group is used as the default for a new DB cluster. This is especially important for parameters that are critical when creating the default database for a DB cluster, such as the character set for the default database defined by the character_set_database
parameter. You can use the Parameter Groups option of the Amazon RDS console or the DescribeDBClusterParameters
operation to verify that your DB cluster parameter group has been created or modified.
For more information on Amazon Aurora, see What is Amazon Aurora? in the Amazon Aurora User Guide.
For more information on Multi-AZ DB clusters, see Multi-AZ DB cluster deployments in the Amazon RDS User Guide.
", "CreateDBClusterSnapshot": "Creates a snapshot of a DB cluster.
For more information on Amazon Aurora, see What is Amazon Aurora? in the Amazon Aurora User Guide.
For more information on Multi-AZ DB clusters, see Multi-AZ DB cluster deployments in the Amazon RDS User Guide.
", @@ -136,7 +136,7 @@ "StartDBCluster": "Starts an Amazon Aurora DB cluster that was stopped using the Amazon Web Services console, the stop-db-cluster CLI command, or the StopDBCluster action.
For more information, see Stopping and Starting an Aurora Cluster in the Amazon Aurora User Guide.
This action only applies to Aurora DB clusters.
Starts an Amazon RDS DB instance that was stopped using the Amazon Web Services console, the stop-db-instance CLI command, or the StopDBInstance action.
For more information, see Starting an Amazon RDS DB instance That Was Previously Stopped in the Amazon RDS User Guide.
This command doesn't apply to RDS Custom, Aurora MySQL, and Aurora PostgreSQL. For Aurora DB clusters, use StartDBCluster
instead.
Enables replication of automated backups to a different Amazon Web Services Region.
This command doesn't apply to RDS Custom.
For more information, see Replicating Automated Backups to Another Amazon Web Services Region in the Amazon RDS User Guide.
", - "StartExportTask": "Starts an export of DB snapshot or DB cluster data to Amazon S3. The provided IAM role must have access to the S3 bucket.
You can't export snapshot data from RDS Custom DB instances.
You can't export cluster data from Multi-AZ DB clusters.
For more information on exporting DB snapshot data, see Exporting DB snapshot data to Amazon S3 in the Amazon RDS User Guide or Exporting DB cluster snapshot data to Amazon S3 in the Amazon Aurora User Guide.
For more information on exporting DB cluster data, see Exporting DB cluster data to Amazon S3 in the Amazon Aurora User Guide.
", + "StartExportTask": "Starts an export of DB snapshot or DB cluster data to Amazon S3. The provided IAM role must have access to the S3 bucket.
You can't export snapshot data from RDS Custom DB instances.
You can't export cluster data from Multi-AZ DB clusters.
For more information on exporting DB snapshot data, see Exporting DB snapshot data to Amazon S3 in the Amazon RDS User Guide or Exporting DB cluster snapshot data to Amazon S3 in the Amazon Aurora User Guide.
For more information on exporting DB cluster data, see Exporting DB cluster data to Amazon S3 in the Amazon Aurora User Guide.
", "StopActivityStream": "Stops a database activity stream that was started using the Amazon Web Services console, the start-activity-stream
CLI command, or the StartActivityStream
action.
For more information, see Monitoring Amazon Aurora with Database Activity Streams in the Amazon Aurora User Guide or Monitoring Amazon RDS with Database Activity Streams in the Amazon RDS User Guide.
", "StopDBCluster": "Stops an Amazon Aurora DB cluster. When you stop a DB cluster, Aurora retains the DB cluster's metadata, including its endpoints and DB parameter groups. Aurora also retains the transaction logs so you can do a point-in-time restore if necessary.
For more information, see Stopping and Starting an Aurora Cluster in the Amazon Aurora User Guide.
This action only applies to Aurora DB clusters.
Stops an Amazon RDS DB instance. When you stop a DB instance, Amazon RDS retains the DB instance's metadata, including its endpoint, DB parameter group, and option group membership. Amazon RDS also retains the transaction logs so you can do a point-in-time restore if necessary.
For more information, see Stopping an Amazon RDS DB Instance Temporarily in the Amazon RDS User Guide.
This command doesn't apply to RDS Custom, Aurora MySQL, and Aurora PostgreSQL. For Aurora clusters, use StopDBCluster
instead.
An error occurred while trying to create the CEV.
", + "refs": { + } + }, "CreateCustomDBEngineVersionMessage": { "base": null, "refs": { @@ -2510,7 +2515,7 @@ "GlobalClusterMemberList": { "base": null, "refs": { - "GlobalCluster$GlobalClusterMembers": "The list of cluster IDs for secondary clusters within the global database cluster. Currently limited to 1 item.
" + "GlobalCluster$GlobalClusterMembers": "The list of primary and secondary clusters within the global database cluster.
" } }, "GlobalClusterNotFoundFault": { @@ -2624,7 +2629,7 @@ "CreateDBClusterMessage$BackupRetentionPeriod": "The number of days for which automated backups are retained.
Default: 1
Constraints:
Must be a value from 1 to 35
Valid for: Aurora DB clusters and Multi-AZ DB clusters
", "CreateDBClusterMessage$Port": "The port number on which the instances in the DB cluster accept connections.
RDS for MySQL and Aurora MySQL
Default: 3306
Valid values: 1150-65535
RDS for PostgreSQL and Aurora PostgreSQL
Default: 5432
Valid values: 1150-65535
Valid for: Aurora DB clusters and Multi-AZ DB clusters
", "CreateDBClusterMessage$AllocatedStorage": "The amount of storage in gibibytes (GiB) to allocate to each DB instance in the Multi-AZ DB cluster.
This setting is required to create a Multi-AZ DB cluster.
Valid for: Multi-AZ DB clusters only
", - "CreateDBClusterMessage$Iops": "The amount of Provisioned IOPS (input/output operations per second) to be initially allocated for each DB instance in the Multi-AZ DB cluster.
For information about valid IOPS values, see Amazon RDS Provisioned IOPS storage in the Amazon RDS User Guide.
This setting is required to create a Multi-AZ DB cluster.
Constraints: Must be a multiple between .5 and 50 of the storage amount for the DB cluster.
Valid for: Multi-AZ DB clusters only
", + "CreateDBClusterMessage$Iops": "The amount of Provisioned IOPS (input/output operations per second) to be initially allocated for each DB instance in the Multi-AZ DB cluster.
For information about valid IOPS values, see Provisioned IOPS storage in the Amazon RDS User Guide.
This setting is required to create a Multi-AZ DB cluster.
Constraints: Must be a multiple between .5 and 50 of the storage amount for the DB cluster.
Valid for: Multi-AZ DB clusters only
", "CreateDBClusterMessage$MonitoringInterval": "The interval, in seconds, between points when Enhanced Monitoring metrics are collected for the DB cluster. To turn off collecting Enhanced Monitoring metrics, specify 0. The default is 0.
If MonitoringRoleArn
is specified, also set MonitoringInterval
to a value other than 0.
Valid Values: 0, 1, 5, 10, 15, 30, 60
Valid for: Multi-AZ DB clusters only
", "CreateDBClusterMessage$PerformanceInsightsRetentionPeriod": "The number of days to retain Performance Insights data. The default is 7 days. The following values are valid:
7
month * 31, where month is a number of months from 1-23
731
For example, the following values are valid:
93 (3 months * 31)
341 (11 months * 31)
589 (19 months * 31)
731
If you specify a retention period such as 94, which isn't a valid value, RDS issues an error.
Valid for: Multi-AZ DB clusters only
", "CreateDBInstanceMessage$AllocatedStorage": "The amount of storage in gibibytes (GiB) to allocate for the DB instance.
Type: Integer
Amazon Aurora
Not applicable. Aurora cluster volumes automatically grow as the amount of data in your database increases, though you are only charged for the space that you use in an Aurora cluster volume.
Amazon RDS Custom
Constraints to the amount of storage for each storage type are the following:
General Purpose (SSD) storage (gp2, gp3): Must be an integer from 40 to 65536 for RDS Custom for Oracle, 16384 for RDS Custom for SQL Server.
Provisioned IOPS storage (io1): Must be an integer from 40 to 65536 for RDS Custom for Oracle, 16384 for RDS Custom for SQL Server.
MySQL
Constraints to the amount of storage for each storage type are the following:
General Purpose (SSD) storage (gp2, gp3): Must be an integer from 20 to 65536.
Provisioned IOPS storage (io1): Must be an integer from 100 to 65536.
Magnetic storage (standard): Must be an integer from 5 to 3072.
MariaDB
Constraints to the amount of storage for each storage type are the following:
General Purpose (SSD) storage (gp2, gp3): Must be an integer from 20 to 65536.
Provisioned IOPS storage (io1): Must be an integer from 100 to 65536.
Magnetic storage (standard): Must be an integer from 5 to 3072.
PostgreSQL
Constraints to the amount of storage for each storage type are the following:
General Purpose (SSD) storage (gp2, gp3): Must be an integer from 20 to 65536.
Provisioned IOPS storage (io1): Must be an integer from 100 to 65536.
Magnetic storage (standard): Must be an integer from 5 to 3072.
Oracle
Constraints to the amount of storage for each storage type are the following:
General Purpose (SSD) storage (gp2, gp3): Must be an integer from 20 to 65536.
Provisioned IOPS storage (io1): Must be an integer from 100 to 65536.
Magnetic storage (standard): Must be an integer from 10 to 3072.
SQL Server
Constraints to the amount of storage for each storage type are the following:
General Purpose (SSD) storage (gp2, gp3):
Enterprise and Standard editions: Must be an integer from 20 to 16384.
Web and Express editions: Must be an integer from 20 to 16384.
Provisioned IOPS storage (io1):
Enterprise and Standard editions: Must be an integer from 100 to 16384.
Web and Express editions: Must be an integer from 100 to 16384.
Magnetic storage (standard):
Enterprise and Standard editions: Must be an integer from 20 to 1024.
Web and Express editions: Must be an integer from 20 to 1024.
A list of log types that this DB instance is configured to export to CloudWatch Logs.
Log types vary by DB engine. For information about the log types for each DB engine, see Amazon RDS Database Log Files in the Amazon RDS User Guide.
", "PendingCloudwatchLogsExports$LogTypesToEnable": "Log types that are in the process of being deactivated. After they are deactivated, these log types aren't exported to CloudWatch Logs.
", "PendingCloudwatchLogsExports$LogTypesToDisable": "Log types that are in the process of being enabled. After they are enabled, these log types are exported to CloudWatch Logs.
", - "RestoreDBClusterFromS3Message$EnableCloudwatchLogsExports": "The list of logs that the restored DB cluster is to export to CloudWatch Logs. The values in the list depend on the DB engine being used.
Aurora MySQL
Possible values are audit
, error
, general
, and slowquery
.
Aurora PostgreSQL
Possible value is postgresql
.
For more information about exporting CloudWatch Logs for Amazon Aurora, see Publishing Database Logs to Amazon CloudWatch Logs in the Amazon Aurora User Guide.
", + "RestoreDBClusterFromS3Message$EnableCloudwatchLogsExports": "The list of logs that the restored DB cluster is to export to CloudWatch Logs. The values in the list depend on the DB engine being used.
Aurora MySQL
Possible values are audit
, error
, general
, and slowquery
.
For more information about exporting CloudWatch Logs for Amazon Aurora, see Publishing Database Logs to Amazon CloudWatch Logs in the Amazon Aurora User Guide.
", "RestoreDBClusterFromSnapshotMessage$EnableCloudwatchLogsExports": "The list of logs that the restored DB cluster is to export to Amazon CloudWatch Logs. The values in the list depend on the DB engine being used.
RDS for MySQL
Possible values are error
, general
, and slowquery
.
RDS for PostgreSQL
Possible values are postgresql
and upgrade
.
Aurora MySQL
Possible values are audit
, error
, general
, and slowquery
.
Aurora PostgreSQL
Possible value is postgresql
.
For more information about exporting CloudWatch Logs for Amazon RDS, see Publishing Database Logs to Amazon CloudWatch Logs in the Amazon RDS User Guide.
For more information about exporting CloudWatch Logs for Amazon Aurora, see Publishing Database Logs to Amazon CloudWatch Logs in the Amazon Aurora User Guide.
Valid for: Aurora DB clusters and Multi-AZ DB clusters
", "RestoreDBClusterToPointInTimeMessage$EnableCloudwatchLogsExports": "The list of logs that the restored DB cluster is to export to CloudWatch Logs. The values in the list depend on the DB engine being used.
RDS for MySQL
Possible values are error
, general
, and slowquery
.
RDS for PostgreSQL
Possible values are postgresql
and upgrade
.
Aurora MySQL
Possible values are audit
, error
, general
, and slowquery
.
Aurora PostgreSQL
Possible value is postgresql
.
For more information about exporting CloudWatch Logs for Amazon RDS, see Publishing Database Logs to Amazon CloudWatch Logs in the Amazon RDS User Guide.
For more information about exporting CloudWatch Logs for Amazon Aurora, see Publishing Database Logs to Amazon CloudWatch Logs in the Amazon Aurora User Guide.
Valid for: Aurora DB clusters and Multi-AZ DB clusters
", "RestoreDBInstanceFromDBSnapshotMessage$EnableCloudwatchLogsExports": "The list of logs that the restored DB instance is to export to CloudWatch Logs. The values in the list depend on the DB engine being used. For more information, see Publishing Database Logs to Amazon CloudWatch Logs in the Amazon RDS User Guide.
This setting doesn't apply to RDS Custom.
", @@ -3992,8 +3997,8 @@ "CreateDBClusterMessage$DBClusterIdentifier": "The DB cluster identifier. This parameter is stored as a lowercase string.
Constraints:
Must contain from 1 to 63 letters, numbers, or hyphens.
First character must be a letter.
Can't end with a hyphen or contain two consecutive hyphens.
Example: my-cluster1
Valid for: Aurora DB clusters and Multi-AZ DB clusters
", "CreateDBClusterMessage$DBClusterParameterGroupName": "The name of the DB cluster parameter group to associate with this DB cluster. If you do not specify a value, then the default DB cluster parameter group for the specified DB engine and version is used.
Constraints:
If supplied, must match the name of an existing DB cluster parameter group.
Valid for: Aurora DB clusters and Multi-AZ DB clusters
", "CreateDBClusterMessage$DBSubnetGroupName": "A DB subnet group to associate with this DB cluster.
This setting is required to create a Multi-AZ DB cluster.
Constraints: Must match the name of an existing DBSubnetGroup. Must not be default.
Example: mydbsubnetgroup
Valid for: Aurora DB clusters and Multi-AZ DB clusters
", - "CreateDBClusterMessage$Engine": "The name of the database engine to be used for this DB cluster.
Valid Values:
aurora
(for MySQL 5.6-compatible Aurora)
aurora-mysql
(for MySQL 5.7-compatible and MySQL 8.0-compatible Aurora)
aurora-postgresql
mysql
postgres
Valid for: Aurora DB clusters and Multi-AZ DB clusters
", - "CreateDBClusterMessage$EngineVersion": "The version number of the database engine to use.
To list all of the available engine versions for MySQL 5.6-compatible Aurora, use the following command:
aws rds describe-db-engine-versions --engine aurora --query \"DBEngineVersions[].EngineVersion\"
To list all of the available engine versions for MySQL 5.7-compatible and MySQL 8.0-compatible Aurora, use the following command:
aws rds describe-db-engine-versions --engine aurora-mysql --query \"DBEngineVersions[].EngineVersion\"
To list all of the available engine versions for Aurora PostgreSQL, use the following command:
aws rds describe-db-engine-versions --engine aurora-postgresql --query \"DBEngineVersions[].EngineVersion\"
To list all of the available engine versions for RDS for MySQL, use the following command:
aws rds describe-db-engine-versions --engine mysql --query \"DBEngineVersions[].EngineVersion\"
To list all of the available engine versions for RDS for PostgreSQL, use the following command:
aws rds describe-db-engine-versions --engine postgres --query \"DBEngineVersions[].EngineVersion\"
Aurora MySQL
For information, see MySQL on Amazon RDS Versions in the Amazon Aurora User Guide.
Aurora PostgreSQL
For information, see Amazon Aurora PostgreSQL releases and engine versions in the Amazon Aurora User Guide.
MySQL
For information, see MySQL on Amazon RDS Versions in the Amazon RDS User Guide.
PostgreSQL
For information, see Amazon RDS for PostgreSQL versions and extensions in the Amazon RDS User Guide.
Valid for: Aurora DB clusters and Multi-AZ DB clusters
", + "CreateDBClusterMessage$Engine": "The name of the database engine to be used for this DB cluster.
Valid Values:
aurora-mysql
aurora-postgresql
mysql
postgres
Valid for: Aurora DB clusters and Multi-AZ DB clusters
", + "CreateDBClusterMessage$EngineVersion": "The version number of the database engine to use.
To list all of the available engine versions for Aurora MySQL version 2 (5.7-compatible) and version 3 (MySQL 8.0-compatible), use the following command:
aws rds describe-db-engine-versions --engine aurora-mysql --query \"DBEngineVersions[].EngineVersion\"
You can supply either 5.7
or 8.0
to use the default engine version for Aurora MySQL version 2 or version 3, respectively.
To list all of the available engine versions for Aurora PostgreSQL, use the following command:
aws rds describe-db-engine-versions --engine aurora-postgresql --query \"DBEngineVersions[].EngineVersion\"
To list all of the available engine versions for RDS for MySQL, use the following command:
aws rds describe-db-engine-versions --engine mysql --query \"DBEngineVersions[].EngineVersion\"
To list all of the available engine versions for RDS for PostgreSQL, use the following command:
aws rds describe-db-engine-versions --engine postgres --query \"DBEngineVersions[].EngineVersion\"
Aurora MySQL
For information, see Database engine updates for Amazon Aurora MySQL in the Amazon Aurora User Guide.
Aurora PostgreSQL
For information, see Amazon Aurora PostgreSQL releases and engine versions in the Amazon Aurora User Guide.
MySQL
For information, see Amazon RDS for MySQL in the Amazon RDS User Guide.
PostgreSQL
For information, see Amazon RDS for PostgreSQL in the Amazon RDS User Guide.
Valid for: Aurora DB clusters and Multi-AZ DB clusters
", "CreateDBClusterMessage$MasterUsername": "The name of the master user for the DB cluster.
Constraints:
Must be 1 to 16 letters or numbers.
First character must be a letter.
Can't be a reserved word for the chosen database engine.
Valid for: Aurora DB clusters and Multi-AZ DB clusters
", "CreateDBClusterMessage$MasterUserPassword": "The password for the master database user. This password can contain any printable ASCII character except \"/\", \"\"\", or \"@\".
Constraints:
Must contain from 8 to 41 characters.
Can't be specified if ManageMasterUserPassword
is turned on.
Valid for: Aurora DB clusters and Multi-AZ DB clusters
", "CreateDBClusterMessage$OptionGroupName": "A value that indicates that the DB cluster should be associated with the specified option group.
DB clusters are associated with a default option group that can't be modified.
", @@ -4002,7 +4007,7 @@ "CreateDBClusterMessage$ReplicationSourceIdentifier": "The Amazon Resource Name (ARN) of the source DB instance or DB cluster if this DB cluster is created as a read replica.
Valid for: Aurora DB clusters and RDS for PostgreSQL Multi-AZ DB clusters
", "CreateDBClusterMessage$KmsKeyId": "The Amazon Web Services KMS key identifier for an encrypted DB cluster.
The Amazon Web Services KMS key identifier is the key ARN, key ID, alias ARN, or alias name for the KMS key. To use a KMS key in a different Amazon Web Services account, specify the key ARN or alias ARN.
When a KMS key isn't specified in KmsKeyId
:
If ReplicationSourceIdentifier
identifies an encrypted source, then Amazon RDS will use the KMS key used to encrypt the source. Otherwise, Amazon RDS will use your default KMS key.
If the StorageEncrypted
parameter is enabled and ReplicationSourceIdentifier
isn't specified, then Amazon RDS will use your default KMS key.
There is a default KMS key for your Amazon Web Services account. Your Amazon Web Services account has a different default KMS key for each Amazon Web Services Region.
If you create a read replica of an encrypted DB cluster in another Amazon Web Services Region, you must set KmsKeyId
to a KMS key identifier that is valid in the destination Amazon Web Services Region. This KMS key is used to encrypt the read replica in that Amazon Web Services Region.
Valid for: Aurora DB clusters and Multi-AZ DB clusters
", "CreateDBClusterMessage$PreSignedUrl": "When you are replicating a DB cluster from one Amazon Web Services GovCloud (US) Region to another, an URL that contains a Signature Version 4 signed request for the CreateDBCluster
operation to be called in the source Amazon Web Services Region where the DB cluster is replicated from. Specify PreSignedUrl
only when you are performing cross-Region replication from an encrypted DB cluster.
The presigned URL must be a valid request for the CreateDBCluster
API operation that can run in the source Amazon Web Services Region that contains the encrypted DB cluster to copy.
The presigned URL request must contain the following parameter values:
KmsKeyId
- The KMS key identifier for the KMS key to use to encrypt the copy of the DB cluster in the destination Amazon Web Services Region. This should refer to the same KMS key for both the CreateDBCluster
operation that is called in the destination Amazon Web Services Region, and the operation contained in the presigned URL.
DestinationRegion
- The name of the Amazon Web Services Region that Aurora read replica will be created in.
ReplicationSourceIdentifier
- The DB cluster identifier for the encrypted DB cluster to be copied. This identifier must be in the Amazon Resource Name (ARN) format for the source Amazon Web Services Region. For example, if you are copying an encrypted DB cluster from the us-west-2 Amazon Web Services Region, then your ReplicationSourceIdentifier
would look like Example: arn:aws:rds:us-west-2:123456789012:cluster:aurora-cluster1
.
To learn how to generate a Signature Version 4 signed request, see Authenticating Requests: Using Query Parameters (Amazon Web Services Signature Version 4) and Signature Version 4 Signing Process.
If you are using an Amazon Web Services SDK tool or the CLI, you can specify SourceRegion
(or --source-region
for the CLI) instead of specifying PreSignedUrl
manually. Specifying SourceRegion
autogenerates a presigned URL that is a valid request for the operation that can run in the source Amazon Web Services Region.
Valid for: Aurora DB clusters only
", - "CreateDBClusterMessage$EngineMode": "The DB engine mode of the DB cluster, either provisioned
, serverless
, parallelquery
, global
, or multimaster
.
The parallelquery
engine mode isn't required for Aurora MySQL version 1.23 and higher 1.x versions, and version 2.09 and higher 2.x versions.
The global
engine mode isn't required for Aurora MySQL version 1.22 and higher 1.x versions, and global
engine mode isn't required for any 2.x versions.
The multimaster
engine mode only applies for DB clusters created with Aurora MySQL version 5.6.10a.
The serverless
engine mode only applies for Aurora Serverless v1 DB clusters.
For Aurora PostgreSQL, the global
engine mode isn't required, and both the parallelquery
and the multimaster
engine modes currently aren't supported.
Limitations and requirements apply to some DB engine modes. For more information, see the following sections in the Amazon Aurora User Guide:
Valid for: Aurora DB clusters only
", + "CreateDBClusterMessage$EngineMode": "The DB engine mode of the DB cluster, either provisioned
or serverless
.
The serverless
engine mode only applies for Aurora Serverless v1 DB clusters.
Limitations and requirements apply to some DB engine modes. For more information, see the following sections in the Amazon Aurora User Guide:
Valid for: Aurora DB clusters only
", "CreateDBClusterMessage$GlobalClusterIdentifier": "The global cluster ID of an Aurora cluster that becomes the primary cluster in the new global database cluster.
Valid for: Aurora DB clusters only
", "CreateDBClusterMessage$Domain": "The Active Directory directory ID to create the DB cluster in.
For Amazon Aurora DB clusters, Amazon RDS can use Kerberos authentication to authenticate users that connect to the DB cluster.
For more information, see Kerberos authentication in the Amazon Aurora User Guide.
Valid for: Aurora DB clusters only
", "CreateDBClusterMessage$DomainIAMRoleName": "Specify the name of the IAM role to be used when making API calls to the Directory Service.
Valid for: Aurora DB clusters only
", @@ -4346,7 +4351,7 @@ "DescribeDBClusterSnapshotsMessage$Marker": "An optional pagination token provided by a previous DescribeDBClusterSnapshots
request. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords
.
The user-supplied DB cluster identifier or the Amazon Resource Name (ARN) of the DB cluster. If this parameter is specified, information from only the specific DB cluster is returned. This parameter isn't case-sensitive.
Constraints:
If supplied, must match an existing DBClusterIdentifier.
An optional pagination token provided by a previous DescribeDBClusters
request. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords
.
The database engine to return.
Valid Values:
aurora
(for MySQL 5.6-compatible Aurora)
aurora-mysql
(for MySQL 5.7-compatible and MySQL 8.0-compatible Aurora)
aurora-postgresql
mariadb
mysql
oracle-ee
oracle-ee-cdb
oracle-se2
oracle-se2-cdb
postgres
sqlserver-ee
sqlserver-se
sqlserver-ex
sqlserver-web
The database engine to return.
Valid Values:
aurora-mysql
(for MySQL 5.7-compatible and MySQL 8.0-compatible Aurora)
aurora-postgresql
mariadb
mysql
oracle-ee
oracle-ee-cdb
oracle-se2
oracle-se2-cdb
postgres
sqlserver-ee
sqlserver-se
sqlserver-ex
sqlserver-web
The database engine version to return.
Example: 5.1.49
The name of a specific DB parameter group family to return details for.
Constraints:
If supplied, must match an existing DBParameterGroupFamily.
An optional pagination token provided by a previous request. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords
.
An optional pagination token provided by a previous DescribeOptionGroups request. If this parameter is specified, the response includes only records beyond the marker, up to the value specified by MaxRecords
.
Filters the list of option groups to only include groups associated with a specific database engine.
Valid Values:
mariadb
mysql
oracle-ee
oracle-ee-cdb
oracle-se2
oracle-se2-cdb
postgres
sqlserver-ee
sqlserver-se
sqlserver-ex
sqlserver-web
Filters the list of option groups to only include groups associated with a specific database engine version. If specified, then EngineName must also be specified.
", - "DescribeOrderableDBInstanceOptionsMessage$Engine": "The name of the engine to retrieve DB instance options for.
Valid Values:
aurora
(for MySQL 5.6-compatible Aurora)
aurora-mysql
(for MySQL 5.7-compatible and MySQL 8.0-compatible Aurora)
aurora-postgresql
mariadb
mysql
oracle-ee
oracle-ee-cdb
oracle-se2
oracle-se2-cdb
postgres
sqlserver-ee
sqlserver-se
sqlserver-ex
sqlserver-web
The name of the engine to retrieve DB instance options for.
Valid Values:
aurora-mysql
(for MySQL 5.7-compatible and MySQL 8.0-compatible Aurora)
aurora-postgresql
mariadb
mysql
oracle-ee
oracle-ee-cdb
oracle-se2
oracle-se2-cdb
postgres
sqlserver-ee
sqlserver-se
sqlserver-ex
sqlserver-web
The engine version filter value. Specify this parameter to show only the available offerings matching the specified engine version.
", "DescribeOrderableDBInstanceOptionsMessage$DBInstanceClass": "The DB instance class filter value. Specify this parameter to show only the available offerings matching the specified DB instance class.
", "DescribeOrderableDBInstanceOptionsMessage$LicenseModel": "The license model filter value. Specify this parameter to show only the available offerings matching the specified license model.
RDS Custom supports only the BYOL licensing model.
", @@ -4518,7 +4523,7 @@ "ModifyDBClusterMessage$OptionGroupName": "A value that indicates that the DB cluster should be associated with the specified option group.
DB clusters are associated with a default option group that can't be modified.
", "ModifyDBClusterMessage$PreferredBackupWindow": "The daily time range during which automated backups are created if automated backups are enabled, using the BackupRetentionPeriod
parameter.
The default is a 30-minute window selected at random from an 8-hour block of time for each Amazon Web Services Region. To view the time blocks available, see Backup window in the Amazon Aurora User Guide.
Constraints:
Must be in the format hh24:mi-hh24:mi
.
Must be in Universal Coordinated Time (UTC).
Must not conflict with the preferred maintenance window.
Must be at least 30 minutes.
Valid for: Aurora DB clusters and Multi-AZ DB clusters
", "ModifyDBClusterMessage$PreferredMaintenanceWindow": "The weekly time range during which system maintenance can occur, in Universal Coordinated Time (UTC).
Format: ddd:hh24:mi-ddd:hh24:mi
The default is a 30-minute window selected at random from an 8-hour block of time for each Amazon Web Services Region, occurring on a random day of the week. To see the time blocks available, see Adjusting the Preferred DB Cluster Maintenance Window in the Amazon Aurora User Guide.
Valid Days: Mon, Tue, Wed, Thu, Fri, Sat, Sun.
Constraints: Minimum 30-minute window.
Valid for: Aurora DB clusters and Multi-AZ DB clusters
", - "ModifyDBClusterMessage$EngineVersion": "The version number of the database engine to which you want to upgrade. Changing this parameter results in an outage. The change is applied during the next maintenance window unless ApplyImmediately
is enabled.
To list all of the available engine versions for MySQL 5.6-compatible Aurora, use the following command:
aws rds describe-db-engine-versions --engine aurora --query \"DBEngineVersions[].EngineVersion\"
To list all of the available engine versions for MySQL 5.7-compatible and MySQL 8.0-compatible Aurora, use the following command:
aws rds describe-db-engine-versions --engine aurora-mysql --query \"DBEngineVersions[].EngineVersion\"
To list all of the available engine versions for Aurora PostgreSQL, use the following command:
aws rds describe-db-engine-versions --engine aurora-postgresql --query \"DBEngineVersions[].EngineVersion\"
To list all of the available engine versions for RDS for MySQL, use the following command:
aws rds describe-db-engine-versions --engine mysql --query \"DBEngineVersions[].EngineVersion\"
To list all of the available engine versions for RDS for PostgreSQL, use the following command:
aws rds describe-db-engine-versions --engine postgres --query \"DBEngineVersions[].EngineVersion\"
Valid for: Aurora DB clusters and Multi-AZ DB clusters
", + "ModifyDBClusterMessage$EngineVersion": "The version number of the database engine to which you want to upgrade. Changing this parameter results in an outage. The change is applied during the next maintenance window unless ApplyImmediately
is enabled.
To list all of the available engine versions for Aurora MySQL version 2 (5.7-compatible) and version 3 (MySQL 8.0-compatible), use the following command:
aws rds describe-db-engine-versions --engine aurora-mysql --query \"DBEngineVersions[].EngineVersion\"
To list all of the available engine versions for Aurora PostgreSQL, use the following command:
aws rds describe-db-engine-versions --engine aurora-postgresql --query \"DBEngineVersions[].EngineVersion\"
To list all of the available engine versions for RDS for MySQL, use the following command:
aws rds describe-db-engine-versions --engine mysql --query \"DBEngineVersions[].EngineVersion\"
To list all of the available engine versions for RDS for PostgreSQL, use the following command:
aws rds describe-db-engine-versions --engine postgres --query \"DBEngineVersions[].EngineVersion\"
Valid for: Aurora DB clusters and Multi-AZ DB clusters
", "ModifyDBClusterMessage$DBInstanceParameterGroupName": "The name of the DB parameter group to apply to all instances of the DB cluster.
When you apply a parameter group using the DBInstanceParameterGroupName
parameter, the DB cluster isn't rebooted automatically. Also, parameter changes are applied immediately rather than during the next maintenance window.
Default: The existing name setting
Constraints:
The DB parameter group must be in the same DB parameter group family as this DB cluster.
The DBInstanceParameterGroupName
parameter is valid in combination with the AllowMajorVersionUpgrade
parameter for a major version upgrade only.
Valid for: Aurora DB clusters only
", "ModifyDBClusterMessage$Domain": "The Active Directory directory ID to move the DB cluster to. Specify none
to remove the cluster from its current domain. The domain must be created prior to this operation.
For more information, see Kerberos Authentication in the Amazon Aurora User Guide.
Valid for: Aurora DB clusters only
", "ModifyDBClusterMessage$DomainIAMRoleName": "Specify the name of the IAM role to be used when making API calls to the Directory Service.
Valid for: Aurora DB clusters only
", @@ -4690,8 +4695,8 @@ "RestoreDBClusterFromS3Message$DBClusterIdentifier": "The name of the DB cluster to create from the source data in the Amazon S3 bucket. This parameter isn't case-sensitive.
Constraints:
Must contain from 1 to 63 letters, numbers, or hyphens.
First character must be a letter.
Can't end with a hyphen or contain two consecutive hyphens.
Example: my-cluster1
The name of the DB cluster parameter group to associate with the restored DB cluster. If this argument is omitted, default.aurora5.6
is used.
Constraints:
If supplied, must match the name of an existing DBClusterParameterGroup.
A DB subnet group to associate with the restored DB cluster.
Constraints: If supplied, must match the name of an existing DBSubnetGroup.
Example: mydbsubnetgroup
The name of the database engine to be used for this DB cluster.
Valid Values: aurora
(for MySQL 5.6-compatible Aurora) and aurora-mysql
(for MySQL 5.7-compatible and MySQL 8.0-compatible Aurora)
The version number of the database engine to use.
To list all of the available engine versions for aurora
(for MySQL 5.6-compatible Aurora), use the following command:
aws rds describe-db-engine-versions --engine aurora --query \"DBEngineVersions[].EngineVersion\"
To list all of the available engine versions for aurora-mysql
(for MySQL 5.7-compatible and MySQL 8.0-compatible Aurora), use the following command:
aws rds describe-db-engine-versions --engine aurora-mysql --query \"DBEngineVersions[].EngineVersion\"
Aurora MySQL
Example: 5.6.10a
, 5.6.mysql_aurora.1.19.2
, 5.7.mysql_aurora.2.07.1
, 8.0.mysql_aurora.3.02.0
The name of the database engine to be used for this DB cluster.
Valid Values: aurora-mysql
(for MySQL 5.7-compatible and MySQL 8.0-compatible Aurora)
The version number of the database engine to use.
To list all of the available engine versions for aurora-mysql
(MySQL 5.7-compatible and MySQL 8.0-compatible Aurora), use the following command:
aws rds describe-db-engine-versions --engine aurora-mysql --query \"DBEngineVersions[].EngineVersion\"
Aurora MySQL
Examples: 5.7.mysql_aurora.2.07.1
, 8.0.mysql_aurora.3.02.0
The name of the master user for the restored DB cluster.
Constraints:
Must be 1 to 16 letters or numbers.
First character must be a letter.
Can't be a reserved word for the chosen database engine.
The password for the master database user. This password can contain any printable ASCII character except \"/\", \"\"\", or \"@\".
Constraints:
Must contain from 8 to 41 characters.
Can't be specified if ManageMasterUserPassword
is turned on.
A value that indicates that the restored DB cluster should be associated with the specified option group.
Permanent options can't be removed from an option group. An option group can't be removed from a DB cluster once it is associated with a DB cluster.
", @@ -4710,7 +4715,7 @@ "RestoreDBClusterFromSnapshotMessage$DBClusterIdentifier": "The name of the DB cluster to create from the DB snapshot or DB cluster snapshot. This parameter isn't case-sensitive.
Constraints:
Must contain from 1 to 63 letters, numbers, or hyphens
First character must be a letter
Can't end with a hyphen or contain two consecutive hyphens
Example: my-snapshot-id
Valid for: Aurora DB clusters and Multi-AZ DB clusters
", "RestoreDBClusterFromSnapshotMessage$SnapshotIdentifier": "The identifier for the DB snapshot or DB cluster snapshot to restore from.
You can use either the name or the Amazon Resource Name (ARN) to specify a DB cluster snapshot. However, you can use only the ARN to specify a DB snapshot.
Constraints:
Must match the identifier of an existing Snapshot.
Valid for: Aurora DB clusters and Multi-AZ DB clusters
", "RestoreDBClusterFromSnapshotMessage$Engine": "The database engine to use for the new DB cluster.
Default: The same as source
Constraint: Must be compatible with the engine of the source
Valid for: Aurora DB clusters and Multi-AZ DB clusters
", - "RestoreDBClusterFromSnapshotMessage$EngineVersion": "The version of the database engine to use for the new DB cluster.
To list all of the available engine versions for MySQL 5.6-compatible Aurora, use the following command:
aws rds describe-db-engine-versions --engine aurora --query \"DBEngineVersions[].EngineVersion\"
To list all of the available engine versions for MySQL 5.7-compatible and MySQL 8.0-compatible Aurora, use the following command:
aws rds describe-db-engine-versions --engine aurora-mysql --query \"DBEngineVersions[].EngineVersion\"
To list all of the available engine versions for Aurora PostgreSQL, use the following command:
aws rds describe-db-engine-versions --engine aurora-postgresql --query \"DBEngineVersions[].EngineVersion\"
To list all of the available engine versions for RDS for MySQL, use the following command:
aws rds describe-db-engine-versions --engine mysql --query \"DBEngineVersions[].EngineVersion\"
To list all of the available engine versions for RDS for PostgreSQL, use the following command:
aws rds describe-db-engine-versions --engine postgres --query \"DBEngineVersions[].EngineVersion\"
Aurora MySQL
See MySQL on Amazon RDS Versions in the Amazon Aurora User Guide.
Aurora PostgreSQL
See Amazon Aurora PostgreSQL releases and engine versions in the Amazon Aurora User Guide.
MySQL
See MySQL on Amazon RDS Versions in the Amazon RDS User Guide.
PostgreSQL
See Amazon RDS for PostgreSQL versions and extensions in the Amazon RDS User Guide.
Valid for: Aurora DB clusters and Multi-AZ DB clusters
", + "RestoreDBClusterFromSnapshotMessage$EngineVersion": "The version of the database engine to use for the new DB cluster. If you don't specify an engine version, the default version for the database engine in the Amazon Web Services Region is used.
To list all of the available engine versions for MySQL 5.7-compatible and MySQL 8.0-compatible Aurora, use the following command:
aws rds describe-db-engine-versions --engine aurora-mysql --query \"DBEngineVersions[].EngineVersion\"
To list all of the available engine versions for Aurora PostgreSQL, use the following command:
aws rds describe-db-engine-versions --engine aurora-postgresql --query \"DBEngineVersions[].EngineVersion\"
To list all of the available engine versions for RDS for MySQL, use the following command:
aws rds describe-db-engine-versions --engine mysql --query \"DBEngineVersions[].EngineVersion\"
To list all of the available engine versions for RDS for PostgreSQL, use the following command:
aws rds describe-db-engine-versions --engine postgres --query \"DBEngineVersions[].EngineVersion\"
Aurora MySQL
See Database engine updates for Amazon Aurora MySQL in the Amazon Aurora User Guide.
Aurora PostgreSQL
See Amazon Aurora PostgreSQL releases and engine versions in the Amazon Aurora User Guide.
MySQL
See Amazon RDS for MySQL in the Amazon RDS User Guide.
PostgreSQL
See Amazon RDS for PostgreSQL versions and extensions in the Amazon RDS User Guide.
Valid for: Aurora DB clusters and Multi-AZ DB clusters
", "RestoreDBClusterFromSnapshotMessage$DBSubnetGroupName": "The name of the DB subnet group to use for the new DB cluster.
Constraints: If supplied, must match the name of an existing DB subnet group.
Example: mydbsubnetgroup
Valid for: Aurora DB clusters and Multi-AZ DB clusters
", "RestoreDBClusterFromSnapshotMessage$DatabaseName": "The database name for the restored DB cluster.
Valid for: Aurora DB clusters and Multi-AZ DB clusters
", "RestoreDBClusterFromSnapshotMessage$OptionGroupName": "The name of the option group to use for the restored DB cluster.
DB clusters are associated with a default option group that can't be modified.
", @@ -4823,7 +4828,7 @@ "StartExportTaskMessage$ExportTaskIdentifier": "A unique identifier for the export task. This ID isn't an identifier for the Amazon S3 bucket where the data is to be exported.
", "StartExportTaskMessage$SourceArn": "The Amazon Resource Name (ARN) of the snapshot or cluster to export to Amazon S3.
", "StartExportTaskMessage$S3BucketName": "The name of the Amazon S3 bucket to export the snapshot or cluster data to.
", - "StartExportTaskMessage$IamRoleArn": "The name of the IAM role to use for writing to the Amazon S3 bucket when exporting a snapshot or cluster.
", + "StartExportTaskMessage$IamRoleArn": "The name of the IAM role to use for writing to the Amazon S3 bucket when exporting a snapshot or cluster.
In the IAM policy attached to your IAM role, include the following required actions to allow the transfer of files from Amazon RDS or Amazon Aurora to an S3 bucket:
s3:PutObject*
s3:GetObject*
s3:ListBucket
s3:DeleteObject*
s3:GetBucketLocation
In the policy, include the resources to identify the S3 bucket and objects in the bucket. The following list of resources shows the Amazon Resource Name (ARN) format for accessing S3:
arn:aws:s3:::your-s3-bucket
arn:aws:s3:::your-s3-bucket/*
The ID of the Amazon Web Services KMS key to use to encrypt the data exported to Amazon S3. The Amazon Web Services KMS key identifier is the key ARN, key ID, alias ARN, or alias name for the KMS key. The caller of this operation must be authorized to run the following operations. These can be set in the Amazon Web Services KMS key policy:
kms:Encrypt
kms:Decrypt
kms:GenerateDataKey
kms:GenerateDataKeyWithoutPlaintext
kms:ReEncryptFrom
kms:ReEncryptTo
kms:CreateGrant
kms:DescribeKey
kms:RetireGrant
The Amazon S3 bucket prefix to use as the file name and path of the exported data.
", "StopActivityStreamRequest$ResourceArn": "The Amazon Resource Name (ARN) of the DB cluster for the database activity stream. For example, arn:aws:rds:us-east-1:12345667890:cluster:das-cluster
.
Contains the details about a blue/green deployment.
For more information, see Using Amazon RDS Blue/Green Deployments for database updates in the Amazon RDS User Guide and Using Amazon RDS Blue/Green Deployments for database updates in the Amazon Aurora User Guide.
", + "base": "Contains the details about a blue/green deployment.
For more information, see Using Amazon RDS Blue/Green Deployments for database updates in the Amazon RDS User Guide and Using Amazon RDS Blue/Green Deployments for database updates in the Amazon Aurora User Guide.
", "refs": { "SwitchoverDetailList$member": null } @@ -4973,7 +4978,7 @@ "SwitchoverDetailStatus": { "base": null, "refs": { - "SwitchoverDetail$Status": "The switchover status of a resource in a blue/green deployment.
Values:
preparing-for-switchover
- The resource is being prepared to switch over.
ready-for-switchover
- The resource is ready to switch over.
switchover-in-progress
- The resource is being switched over.
switchover-completed
- The resource has been switched over.
switchover-failed
- The resource attempted to switch over but failed.
The switchover status of a resource in a blue/green deployment.
Values:
PROVISIONING
- The resource is being prepared to switch over.
AVAILABLE
- The resource is ready to switch over.
SWITCHOVER_IN_PROGRESS
- The resource is being switched over.
SWITCHOVER_COMPLETED
- The resource has been switched over.
SWITCHOVER_FAILED
- The resource attempted to switch over but failed.
MISSING_SOURCE
- The source resource has been deleted.
MISSING_TARGET
- The target resource has been deleted.