From 49521540501bb752b5315b79e989fdb847c05850 Mon Sep 17 00:00:00 2001 From: aws-sdk-go-automation <43143561+aws-sdk-go-automation@users.noreply.github.com> Date: Thu, 11 Feb 2021 11:09:40 -0800 Subject: [PATCH] Release v1.37.9 (2021-02-11) (#3779) Release v1.37.9 (2021-02-11) === ### Service Client Updates * `service/databrew`: Updates service API and documentation * `service/rds`: Updates service API and documentation * Adding support for RDS Aurora Global Database Failover --- CHANGELOG.md | 8 + aws/version.go | 2 +- models/apis/databrew/2017-07-25/api-2.json | 35 ++- models/apis/databrew/2017-07-25/docs-2.json | 39 ++- models/apis/rds/2014-10-31/api-2.json | 67 +++- models/apis/rds/2014-10-31/docs-2.json | 42 ++- service/gluedatabrew/api.go | 151 ++++++++- service/rds/api.go | 323 +++++++++++++++++++- service/rds/rdsiface/interface.go | 4 + 9 files changed, 632 insertions(+), 39 deletions(-) diff --git a/CHANGELOG.md b/CHANGELOG.md index 6e05a8a80bb..20e5e2e47fe 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -1,3 +1,11 @@ +Release v1.37.9 (2021-02-11) +=== + +### Service Client Updates +* `service/databrew`: Updates service API and documentation +* `service/rds`: Updates service API and documentation + * Adding support for RDS Aurora Global Database Failover + Release v1.37.8 (2021-02-09) === diff --git a/aws/version.go b/aws/version.go index c6201b829ea..5c5b2183ef1 100644 --- a/aws/version.go +++ b/aws/version.go @@ -5,4 +5,4 @@ package aws const SDKName = "aws-sdk-go" // SDKVersion is the version of this SDK -const SDKVersion = "1.37.8" +const SDKVersion = "1.37.9" diff --git a/models/apis/databrew/2017-07-25/api-2.json b/models/apis/databrew/2017-07-25/api-2.json index bd344d56990..409a243b4db 100644 --- a/models/apis/databrew/2017-07-25/api-2.json +++ b/models/apis/databrew/2017-07-25/api-2.json @@ -706,7 +706,8 @@ "OutputLocation":{"shape":"S3Location"}, "RoleArn":{"shape":"Arn"}, "Tags":{"shape":"TagMap"}, - "Timeout":{"shape":"Timeout"} + "Timeout":{"shape":"Timeout"}, + "JobSample":{"shape":"JobSample"} } }, "CreateProfileJobResponse":{ @@ -1046,7 +1047,8 @@ "ResourceArn":{"shape":"Arn"}, "RoleArn":{"shape":"Arn"}, "Tags":{"shape":"TagMap"}, - "Timeout":{"shape":"Timeout"} + "Timeout":{"shape":"Timeout"}, + "JobSample":{"shape":"JobSample"} } }, "DescribeJobRunRequest":{ @@ -1085,7 +1087,8 @@ "Outputs":{"shape":"OutputList"}, "RecipeReference":{"shape":"RecipeReference"}, "StartedBy":{"shape":"StartedBy"}, - "StartedOn":{"shape":"Date"} + "StartedOn":{"shape":"Date"}, + "JobSample":{"shape":"JobSample"} } }, "DescribeProjectRequest":{ @@ -1254,7 +1257,8 @@ "ResourceArn":{"shape":"Arn"}, "RoleArn":{"shape":"Arn"}, "Timeout":{"shape":"Timeout"}, - "Tags":{"shape":"TagMap"} + "Tags":{"shape":"TagMap"}, + "JobSample":{"shape":"JobSample"} } }, "JobList":{ @@ -1287,7 +1291,8 @@ "Outputs":{"shape":"OutputList"}, "RecipeReference":{"shape":"RecipeReference"}, "StartedBy":{"shape":"StartedBy"}, - "StartedOn":{"shape":"Date"} + "StartedOn":{"shape":"Date"}, + "JobSample":{"shape":"JobSample"} } }, "JobRunErrorMessage":{"type":"string"}, @@ -1312,6 +1317,14 @@ "TIMEOUT" ] }, + "JobSample":{ + "type":"structure", + "members":{ + "Mode":{"shape":"SampleMode"}, + "Size":{"shape":"JobSize"} + } + }, + "JobSize":{"type":"long"}, "JobType":{ "type":"string", "enum":[ @@ -1626,7 +1639,7 @@ }, "ParameterValue":{ "type":"string", - "max":8192, + "max":12288, "min":1 }, "Preview":{"type":"boolean"}, @@ -1791,6 +1804,13 @@ "Type":{"shape":"SampleType"} } }, + "SampleMode":{ + "type":"string", + "enum":[ + "FULL_DATASET", + "CUSTOM_ROWS" + ] + }, "SampleSize":{ "type":"integer", "max":5000, @@ -2102,7 +2122,8 @@ "MaxRetries":{"shape":"MaxRetries"}, "OutputLocation":{"shape":"S3Location"}, "RoleArn":{"shape":"Arn"}, - "Timeout":{"shape":"Timeout"} + "Timeout":{"shape":"Timeout"}, + "JobSample":{"shape":"JobSample"} } }, "UpdateProfileJobResponse":{ diff --git a/models/apis/databrew/2017-07-25/docs-2.json b/models/apis/databrew/2017-07-25/docs-2.json index d512158729e..902a649bd36 100644 --- a/models/apis/databrew/2017-07-25/docs-2.json +++ b/models/apis/databrew/2017-07-25/docs-2.json @@ -2,7 +2,7 @@ "version": "2.0", "service": "
AWS Glue DataBrew is a visual, cloud-scale data-preparation service. DataBrew simplifies data preparation tasks, targeting data issues that are hard to spot and time-consuming to fix. DataBrew empowers users of all technical levels to visualize the data and perform one-click data transformations, with no coding required.
", "operations": { - "BatchDeleteRecipeVersion": "Deletes one or more versions of a recipe at a time.
The entire request will be rejected if:
The recipe does not exist.
There is an invalid version identifier in the list of versions.
The verision list is empty.
The version list size exceeds 50.
The verison list contains duplicate entries.
The request will complete successfully, but with partial failures, if:
A version does not exist.
A version is being used by a job.
You specify LATEST_WORKING
, but it's being used by a project.
The version fails to be deleted.
The LATEST_WORKING
version will only be deleted if the recipe has no other versions. If you try to delete LATEST_WORKING
while other versions exist (or if they can't be deleted), then LATEST_WORKING
will be listed as partial failure in the response.
Deletes one or more versions of a recipe at a time.
The entire request will be rejected if:
The recipe does not exist.
There is an invalid version identifier in the list of versions.
The version list is empty.
The version list size exceeds 50.
The version list contains duplicate entries.
The request will complete successfully, but with partial failures, if:
A version does not exist.
A version is being used by a job.
You specify LATEST_WORKING
, but it's being used by a project.
The version fails to be deleted.
The LATEST_WORKING
version will only be deleted if the recipe has no other versions. If you try to delete LATEST_WORKING
while other versions exist (or if they can't be deleted), then LATEST_WORKING
will be listed as partial failure in the response.
Creates a new DataBrew dataset.
", "CreateProfileJob": "Creates a new job to analyze a dataset and create its data profile.
", "CreateProject": "Creates a new DataBrew project.
", @@ -496,12 +496,12 @@ "EncryptionMode": { "base": null, "refs": { - "CreateProfileJobRequest$EncryptionMode": "The encryption mode for the job, which can be one of the following:
SSE-KMS
- para>SSE-KMS
- server-side encryption with AWS KMS-managed keys.
SSE-S3
- Server-side encryption with keys managed by Amazon S3.
The encryption mode for the job, which can be one of the following:
SSE-KMS
- Server-side encryption with AWS KMS-managed keys.
SSE-S3
- Server-side encryption with keys managed by Amazon S3.
The encryption mode for the job, which can be one of the following:
SSE-KMS
- Server-side encryption with AWS KMS-managed keys.
SSE-S3
- Server-side encryption with keys managed by Amazon S3.
The encryption mode for the job, which can be one of the following:
SSE-KMS
- SSE-KMS
- Server-side encryption with AWS KMS-managed keys.
SSE-S3
- Server-side encryption with keys managed by Amazon S3.
The encryption mode for the job, which can be one of the following:
SSE-KMS
- Server-side encryption with keys managed by AWS KMS.
SSE-S3
- Server-side encryption with keys managed by Amazon S3.
The encryption mode for the job, which can be one of the following:
SSE-KMS
- Server-side encryption with keys managed by AWS KMS.
SSE-S3
- Server-side encryption with keys managed by Amazon S3.
The encryption mode for the job, which can be one of the following:
SSE-KMS
- Server-side encryption with AWS KMS-managed keys.
SSE-S3
- Server-side encryption with keys managed by Amazon S3.
The encryption mode for the job, which can be one of the following:
SSE-KMS
- Server-side encryption with AWS KMS-managed keys.
SSE-S3
- Server-side encryption with keys managed by Amazon S3.
The encryption mode for the job, which can be one of the following:
SSE-KMS
- Server-side encryption with AWS KMS-managed keys.
SSE-S3
- Server-side encryption with keys managed by Amazon S3.
The encryption mode for the job, which can be one of the following:
SSE-KMS
- Server-side encryption with keys managed by AWS KMS.
SSE-S3
- Server-side encryption with keys managed by Amazon S3.
The encryption mode for the job, which can be one of the following:
SSE-KMS
- Server-side encryption with keys managed by AWS KMS.
SSE-S3
- Server-side encryption with keys managed by Amazon S3.
The amount of time, in seconds, during which a job run consumed resources.
", + "DescribeJobRunResponse$ExecutionTime": "The amount of time, in seconds, during which the job run consumed resources.
", "JobRun$ExecutionTime": "The amount of time, in seconds, during which a job run consumed resources.
" } }, @@ -636,6 +636,23 @@ "JobRun$State": "The current state of the job run entity itself.
" } }, + "JobSample": { + "base": "Sample configuration for Profile Jobs only. Determines the number of rows on which the Profile job will be executed. If a JobSample value is not provided for profile jobs, the default value will be used. The default value is CUSTOM_ROWS for the mode parameter and 20000 for the size parameter.
", + "refs": { + "CreateProfileJobRequest$JobSample": "Sample configuration for profile jobs only. Determines the number of rows on which the profile job will be executed. If a JobSample value is not provided, the default value will be used. The default value is CUSTOM_ROWS for the mode parameter and 20000 for the size parameter.
", + "DescribeJobResponse$JobSample": "Sample configuration for profile jobs only. Determines the number of rows on which the profile job will be executed.
", + "DescribeJobRunResponse$JobSample": "Sample configuration for profile jobs only. Determines the number of rows on which the profile job will be executed. If a JobSample value is not provided, the default value will be used. The default value is CUSTOM_ROWS for the mode parameter and 20000 for the size parameter.
", + "Job$JobSample": "Sample configuration for profile jobs only. Determines the number of rows on which the profile job will be executed. If a JobSample value is not provided, the default value will be used. The default value is CUSTOM_ROWS for the mode parameter and 20000 for the size parameter.
", + "JobRun$JobSample": "Sample configuration for profile jobs only. Determines the number of rows on which the profile job will be executed. If a JobSample value is not provided, the default value will be used. The default value is CUSTOM_ROWS for the mode parameter and 20000 for the size parameter.
", + "UpdateProfileJobRequest$JobSample": "Sample configuration for Profile Jobs only. Determines the number of rows on which the Profile job will be executed. If a JobSample value is not provided for profile jobs, the default value will be used. The default value is CUSTOM_ROWS for the mode parameter and 20000 for the size parameter.
" + } + }, + "JobSize": { + "base": null, + "refs": { + "JobSample$Size": "Size parameter is only required when the mode is CUSTOM_ROWS. Profile job will be executed on the the specified number of rows. The maximum value for size is Long.MAX_VALUE.
Long.MAX_VALUE = 9223372036854775807
" + } + }, "JobType": { "base": null, "refs": { @@ -1111,6 +1128,12 @@ "UpdateProjectRequest$Sample": null } }, + "SampleMode": { + "base": null, + "refs": { + "JobSample$Mode": "Determines whether the profile job will be executed on the entire dataset or on a specified number of rows. Must be one of the following:
FULL_DATASET: Profile job will be executed on the entire dataset.
CUSTOM_ROWS: Profile job will be executed on the number of rows specified in the Size parameter.
The Amazon Resource Name (ARN) of the user who initiated the job run.
", + "DescribeJobRunResponse$StartedBy": "The Amazon Resource Name (ARN) of the user who started the job run.
", "JobRun$StartedBy": "The Amazon Resource Name (ARN) of the user who initiated the job run.
" } }, diff --git a/models/apis/rds/2014-10-31/api-2.json b/models/apis/rds/2014-10-31/api-2.json index aa36c484070..3b8585b1694 100644 --- a/models/apis/rds/2014-10-31/api-2.json +++ b/models/apis/rds/2014-10-31/api-2.json @@ -1401,6 +1401,24 @@ {"shape":"InvalidDBInstanceStateFault"} ] }, + "FailoverGlobalCluster":{ + "name":"FailoverGlobalCluster", + "http":{ + "method":"POST", + "requestUri":"/" + }, + "input":{"shape":"FailoverGlobalClusterMessage"}, + "output":{ + "shape":"FailoverGlobalClusterResult", + "resultWrapper":"FailoverGlobalClusterResult" + }, + "errors":[ + {"shape":"GlobalClusterNotFoundFault"}, + {"shape":"InvalidGlobalClusterStateFault"}, + {"shape":"InvalidDBClusterStateFault"}, + {"shape":"DBClusterNotFoundFault"} + ] + }, "ImportInstallationMedia":{ "name":"ImportInstallationMedia", "http":{ @@ -3385,6 +3403,12 @@ }, "exception":true }, + "DBClusterIdentifier":{ + "type":"string", + "max":255, + "min":1, + "pattern":"[A-Za-z][0-9A-Za-z-:._]*" + }, "DBClusterList":{ "type":"list", "member":{ @@ -5365,6 +5389,40 @@ "DBCluster":{"shape":"DBCluster"} } }, + "FailoverGlobalClusterMessage":{ + "type":"structure", + "required":[ + "GlobalClusterIdentifier", + "TargetDbClusterIdentifier" + ], + "members":{ + "GlobalClusterIdentifier":{"shape":"GlobalClusterIdentifier"}, + "TargetDbClusterIdentifier":{"shape":"DBClusterIdentifier"} + } + }, + "FailoverGlobalClusterResult":{ + "type":"structure", + "members":{ + "GlobalCluster":{"shape":"GlobalCluster"} + } + }, + "FailoverState":{ + "type":"structure", + "members":{ + "Status":{"shape":"FailoverStatus"}, + "FromDbClusterArn":{"shape":"String"}, + "ToDbClusterArn":{"shape":"String"} + }, + "wrapper":true + }, + "FailoverStatus":{ + "type":"string", + "enum":[ + "pending", + "failing-over", + "cancelling" + ] + }, "FeatureNameList":{ "type":"list", "member":{"shape":"String"} @@ -5406,7 +5464,8 @@ "DatabaseName":{"shape":"String"}, "StorageEncrypted":{"shape":"BooleanOptional"}, "DeletionProtection":{"shape":"BooleanOptional"}, - "GlobalClusterMembers":{"shape":"GlobalClusterMemberList"} + "GlobalClusterMembers":{"shape":"GlobalClusterMemberList"}, + "FailoverState":{"shape":"FailoverState"} }, "wrapper":true }, @@ -5421,6 +5480,12 @@ }, "exception":true }, + "GlobalClusterIdentifier":{ + "type":"string", + "max":255, + "min":1, + "pattern":"[A-Za-z][0-9A-Za-z-:._]*" + }, "GlobalClusterList":{ "type":"list", "member":{ diff --git a/models/apis/rds/2014-10-31/docs-2.json b/models/apis/rds/2014-10-31/docs-2.json index dd64478a5a4..ea25a9058b9 100644 --- a/models/apis/rds/2014-10-31/docs-2.json +++ b/models/apis/rds/2014-10-31/docs-2.json @@ -88,6 +88,7 @@ "DescribeValidDBInstanceModifications": "You can call DescribeValidDBInstanceModifications
to learn what modifications you can make to your DB instance. You can use this information when you call ModifyDBInstance
.
Downloads all or a portion of the specified log file, up to 1 MB in size.
", "FailoverDBCluster": "Forces a failover for a DB cluster.
A failover for a DB cluster promotes one of the Aurora Replicas (read-only instances) in the DB cluster to be the primary instance (the cluster writer).
Amazon Aurora will automatically fail over to an Aurora Replica, if one exists, when the primary instance fails. You can force a failover when you want to simulate a failure of a primary instance for testing. Because each instance in a DB cluster has its own endpoint address, you will need to clean up and re-establish any existing connections that use those endpoint addresses when the failover is complete.
For more information on Amazon Aurora, see What Is Amazon Aurora? in the Amazon Aurora User Guide.
This action only applies to Aurora DB clusters.
Initiates the failover process for an Aurora global database (GlobalCluster).
A failover for an Aurora global database promotes one of secondary read-only DB clusters to be the primary DB cluster and demotes the primary DB cluster to being a secondary (read-only) DB cluster. In other words, the role of the current primary DB cluster and the selected (target) DB cluster are switched. The selected secondary DB cluster assumes full read/write capabilities for the Aurora global database.
For more information about failing over an Amazon Aurora global database, see Managed planned failover for Amazon Aurora global databases in the Amazon Aurora User Guide.
This action applies to GlobalCluster (Aurora global databases) only. Use this action only on healthy Aurora global databases with running Aurora DB clusters and no Region-wide outages, to test disaster recovery scenarios or to reconfigure your Aurora global database topology.
Imports the installation media for a DB engine that requires an on-premises customer provided license, such as SQL Server.
", "ListTagsForResource": "Lists all tags on an Amazon RDS resource.
For an overview on tagging an Amazon RDS resource, see Tagging Amazon RDS Resources in the Amazon RDS User Guide.
", "ModifyCertificates": "Override the system-default Secure Sockets Layer/Transport Layer Security (SSL/TLS) certificate for Amazon RDS for new DB instances temporarily, or remove the override.
By using this operation, you can specify an RDS-approved SSL/TLS certificate for new DB instances that is different from the default certificate provided by RDS. You can also use this operation to remove the override, so that new DB instances use the default certificate provided by RDS.
You might need to override the default certificate in the following situations:
You already migrated your applications to support the latest certificate authority (CA) certificate, but the new CA certificate is not yet the RDS default CA certificate for the specified AWS Region.
RDS has already moved to a new default CA certificate for the specified AWS Region, but you are still in the process of supporting the new CA certificate. In this case, you temporarily need additional time to finish your application changes.
For more information about rotating your SSL/TLS certificate for RDS DB engines, see Rotating Your SSL/TLS Certificate in the Amazon RDS User Guide.
For more information about rotating your SSL/TLS certificate for Aurora DB engines, see Rotating Your SSL/TLS Certificate in the Amazon Aurora User Guide.
", @@ -394,7 +395,7 @@ "CreateDBClusterMessage$DeletionProtection": "A value that indicates whether the DB cluster has deletion protection enabled. The database can't be deleted when deletion protection is enabled. By default, deletion protection is disabled.
", "CreateDBClusterMessage$EnableHttpEndpoint": "A value that indicates whether to enable the HTTP endpoint for an Aurora Serverless DB cluster. By default, the HTTP endpoint is disabled.
When enabled, the HTTP endpoint provides a connectionless web service API for running SQL queries on the Aurora Serverless DB cluster. You can also query your database from inside the RDS console with the query editor.
For more information, see Using the Data API for Aurora Serverless in the Amazon Aurora User Guide.
", "CreateDBClusterMessage$CopyTagsToSnapshot": "A value that indicates whether to copy all tags from the DB cluster to snapshots of the DB cluster. The default is not to copy them.
", - "CreateDBClusterMessage$EnableGlobalWriteForwarding": "A value that indicates whether to enable write operations to be forwarded from this cluster to the primary cluster in an Aurora global database. The resulting changes are replicated back to this cluster. This parameter only applies to DB clusters that are secondary clusters in an Aurora global database. By default, Aurora disallows write operations for secondary clusters.
", + "CreateDBClusterMessage$EnableGlobalWriteForwarding": "A value that indicates whether to enable this DB cluster to forward write operations to the primary cluster of an Aurora global database (GlobalCluster). By default, write operations are not allowed on Aurora DB clusters that are secondary clusters in an Aurora global database.
You can set this value only on Aurora DB clusters that are members of an Aurora global database. With this parameter enabled, a secondary cluster can forward writes to the current primary cluster and the resulting changes are replicated back to this cluster. For the primary DB cluster of an Aurora global database, this value is used immediately if the primary is demoted by the FailoverGlobalCluster API operation, but it does nothing until then.
", "CreateDBInstanceMessage$MultiAZ": "A value that indicates whether the DB instance is a Multi-AZ deployment. You can't set the AvailabilityZone
parameter if the DB instance is a Multi-AZ deployment.
A value that indicates whether minor engine upgrades are applied automatically to the DB instance during the maintenance window. By default, minor engine upgrades are applied automatically.
", "CreateDBInstanceMessage$PubliclyAccessible": "A value that indicates whether the DB instance is publicly accessible.
When the DB instance is publicly accessible, its DNS endpoint resolves to the private IP address from within the DB instance's VPC, and to the public IP address from outside of the DB instance's VPC. Access to the DB instance is ultimately controlled by the security group it uses, and that public access is not permitted if the security group assigned to the DB instance doesn't permit it.
When the DB instance isn't publicly accessible, it is an internal DB instance with a DNS name that resolves to a private IP address.
Default: The default behavior varies depending on whether DBSubnetGroupName
is specified.
If DBSubnetGroupName
isn't specified, and PubliclyAccessible
isn't specified, the following applies:
If the default VPC in the target region doesn’t have an Internet gateway attached to it, the DB instance is private.
If the default VPC in the target region has an Internet gateway attached to it, the DB instance is public.
If DBSubnetGroupName
is specified, and PubliclyAccessible
isn't specified, the following applies:
If the subnets are part of a VPC that doesn’t have an Internet gateway attached to it, the DB instance is private.
If the subnets are part of a VPC that has an Internet gateway attached to it, the DB instance is public.
A value that indicates whether the DB cluster has deletion protection enabled. The database can't be deleted when deletion protection is enabled. By default, deletion protection is disabled.
", "ModifyDBClusterMessage$EnableHttpEndpoint": "A value that indicates whether to enable the HTTP endpoint for an Aurora Serverless DB cluster. By default, the HTTP endpoint is disabled.
When enabled, the HTTP endpoint provides a connectionless web service API for running SQL queries on the Aurora Serverless DB cluster. You can also query your database from inside the RDS console with the query editor.
For more information, see Using the Data API for Aurora Serverless in the Amazon Aurora User Guide.
", "ModifyDBClusterMessage$CopyTagsToSnapshot": "A value that indicates whether to copy all tags from the DB cluster to snapshots of the DB cluster. The default is not to copy them.
", - "ModifyDBClusterMessage$EnableGlobalWriteForwarding": "A value that indicates whether to enable write operations to be forwarded from this cluster to the primary cluster in an Aurora global database. The resulting changes are replicated back to this cluster. This parameter only applies to DB clusters that are secondary clusters in an Aurora global database. By default, Aurora disallows write operations for secondary clusters.
", + "ModifyDBClusterMessage$EnableGlobalWriteForwarding": "A value that indicates whether to enable this DB cluster to forward write operations to the primary cluster of an Aurora global database (GlobalCluster). By default, write operations are not allowed on Aurora DB clusters that are secondary clusters in an Aurora global database.
You can set this value only on Aurora DB clusters that are members of an Aurora global database. With this parameter enabled, a secondary cluster can forward writes to the current primary cluster and the resulting changes are replicated back to this cluster. For the primary DB cluster of an Aurora global database, this value is used immediately if the primary is demoted by the FailoverGlobalCluster API operation, but it does nothing until then.
", "ModifyDBInstanceMessage$MultiAZ": "A value that indicates whether the DB instance is a Multi-AZ deployment. Changing this parameter doesn't result in an outage and the change is applied during the next maintenance window unless the ApplyImmediately
parameter is enabled for this request.
A value that indicates whether minor version upgrades are applied automatically to the DB instance during the maintenance window. Changing this parameter doesn't result in an outage except in the following case and the change is asynchronously applied as soon as possible. An outage results if this parameter is enabled during the maintenance window, and a newer minor version is available, and RDS has enabled auto patching for that engine version.
", "ModifyDBInstanceMessage$CopyTagsToSnapshot": "A value that indicates whether to copy all tags from the DB instance to snapshots of the DB instance. By default, tags are not copied.
Amazon Aurora
Not applicable. Copying tags to snapshots is managed by the DB cluster. Setting this value for an Aurora DB instance has no effect on the DB cluster setting. For more information, see ModifyDBCluster
.
Identifier of the secondary Aurora DB cluster that you want to promote to primary for the Aurora global database (GlobalCluster.) Use the Amazon Resource Name (ARN) for the identifier so that Aurora can locate the cluster in its AWS Region.
" + } + }, "DBClusterList": { "base": null, "refs": { @@ -2049,6 +2056,28 @@ "refs": { } }, + "FailoverGlobalClusterMessage": { + "base": null, + "refs": { + } + }, + "FailoverGlobalClusterResult": { + "base": null, + "refs": { + } + }, + "FailoverState": { + "base": "Contains the state of scheduled or in-process failover operations on an Aurora global database (GlobalCluster). This Data type is empty unless a failover operation is scheduled or is currently underway on the Aurora global database.
", + "refs": { + "GlobalCluster$FailoverState": "A data object containing all properties for the current state of an in-process or pending failover process for this Aurora global database. This object is empty unless the FailoverGlobalCluster API operation has been called on this Aurora global database (GlobalCluster).
" + } + }, + "FailoverStatus": { + "base": null, + "refs": { + "FailoverState$Status": "The current status of the Aurora global database (GlobalCluster). Possible values are as follows:
pending A request to fail over the Aurora global database (GlobalCluster) has been received by the service. The GlobalCluster
's primary DB cluster and the specified secondary DB cluster are being verified before the failover process can start.
failing-over This status covers the range of Aurora internal operations that take place during the failover process, such as demoting the primary Aurora DB cluster, promoting the secondary Aurora DB, and synchronizing replicas.
cancelling The request to fail over the Aurora global database (GlobalCluster) was cancelled and the primary Aurora DB cluster and the selected secondary Aurora DB cluster are returning to their previous states.
Identifier of the Aurora global database (GlobalCluster) that should be failed over. The identifier is the unique key assigned by the user when the Aurora global database was created. In other words, it's the name of the Aurora global database that you want to fail over.
Constraints:
Must match the identifier of an existing GlobalCluster (Aurora global database).
A pagination token that can be used in a later DescribeExportTasks
request. A marker is used for pagination to identify the location to begin output for the next response of DescribeExportTasks
.
A DB cluster identifier to force a failover for. This parameter isn't case-sensitive.
Constraints:
Must match the identifier of an existing DBCluster.
The name of the instance to promote to the primary instance.
You must specify the instance identifier for an Aurora Replica in the DB cluster. For example, mydbcluster-replica1
.
The Amazon Resource Name (ARN) of the Aurora DB cluster that is currently being demoted, and which is associated with this state.
", + "FailoverState$ToDbClusterArn": "The Amazon Resource Name (ARN) of the Aurora DB cluster that is currently being promoted, and which is associated with this state.
", "FeatureNameList$member": null, "Filter$Name": "The name of the filter. Filter names are case-sensitive.
", "FilterValueList$member": null, diff --git a/service/gluedatabrew/api.go b/service/gluedatabrew/api.go index 29f08065c4a..ba3dc30cceb 100644 --- a/service/gluedatabrew/api.go +++ b/service/gluedatabrew/api.go @@ -65,11 +65,11 @@ func (c *GlueDataBrew) BatchDeleteRecipeVersionRequest(input *BatchDeleteRecipeV // // * There is an invalid version identifier in the list of versions. // -// * The verision list is empty. +// * The version list is empty. // // * The version list size exceeds 50. // -// * The verison list contains duplicate entries. +// * The version list contains duplicate entries. // // The request will complete successfully, but with partial failures, if: // @@ -4134,12 +4134,17 @@ type CreateProfileJobInput struct { // The encryption mode for the job, which can be one of the following: // - // * SSE-KMS - para>SSE-KMS - server-side encryption with AWS KMS-managed - // keys. + // * SSE-KMS - SSE-KMS - Server-side encryption with AWS KMS-managed keys. // // * SSE-S3 - Server-side encryption with keys managed by Amazon S3. EncryptionMode *string `type:"string" enum:"EncryptionMode"` + // Sample configuration for profile jobs only. Determines the number of rows + // on which the profile job will be executed. If a JobSample value is not provided, + // the default value will be used. The default value is CUSTOM_ROWS for the + // mode parameter and 20000 for the size parameter. + JobSample *JobSample `type:"structure"` + // Enables or disables Amazon CloudWatch logging for the job. If logging is // enabled, CloudWatch writes one log stream for each job run. LogSubscription *string `type:"string" enum:"LogSubscription"` @@ -4247,6 +4252,12 @@ func (s *CreateProfileJobInput) SetEncryptionMode(v string) *CreateProfileJobInp return s } +// SetJobSample sets the JobSample field's value. +func (s *CreateProfileJobInput) SetJobSample(v *JobSample) *CreateProfileJobInput { + s.JobSample = v + return s +} + // SetLogSubscription sets the LogSubscription field's value. func (s *CreateProfileJobInput) SetLogSubscription(v string) *CreateProfileJobInput { s.LogSubscription = &v @@ -4567,7 +4578,7 @@ type CreateRecipeJobInput struct { // The encryption mode for the job, which can be one of the following: // - // * SSE-KMS - Server-side encryption with AWS KMS-managed keys. + // * SSE-KMS - Server-side encryption with keys managed by AWS KMS. // // * SSE-S3 - Server-side encryption with keys managed by Amazon S3. EncryptionMode *string `type:"string" enum:"EncryptionMode"` @@ -5768,11 +5779,15 @@ type DescribeJobOutput struct { // The encryption mode for the job, which can be one of the following: // - // * SSE-KMS - Server-side encryption with AWS KMS-managed keys. + // * SSE-KMS - Server-side encryption with keys managed by AWS KMS. // // * SSE-S3 - Server-side encryption with keys managed by Amazon S3. EncryptionMode *string `type:"string" enum:"EncryptionMode"` + // Sample configuration for profile jobs only. Determines the number of rows + // on which the profile job will be executed. + JobSample *JobSample `type:"structure"` + // The identifier (user name) of the user who last modified the job. LastModifiedBy *string `type:"string"` @@ -5866,6 +5881,12 @@ func (s *DescribeJobOutput) SetEncryptionMode(v string) *DescribeJobOutput { return s } +// SetJobSample sets the JobSample field's value. +func (s *DescribeJobOutput) SetJobSample(v *JobSample) *DescribeJobOutput { + s.JobSample = v + return s +} + // SetLastModifiedBy sets the LastModifiedBy field's value. func (s *DescribeJobOutput) SetLastModifiedBy(v string) *DescribeJobOutput { s.LastModifiedBy = &v @@ -6024,7 +6045,7 @@ type DescribeJobRunOutput struct { // ran. ErrorMessage *string `type:"string"` - // The amount of time, in seconds, during which a job run consumed resources. + // The amount of time, in seconds, during which the job run consumed resources. ExecutionTime *int64 `type:"integer"` // The name of the job being processed during this run. @@ -6032,6 +6053,12 @@ type DescribeJobRunOutput struct { // JobName is a required field JobName *string `min:"1" type:"string" required:"true"` + // Sample configuration for profile jobs only. Determines the number of rows + // on which the profile job will be executed. If a JobSample value is not provided, + // the default value will be used. The default value is CUSTOM_ROWS for the + // mode parameter and 20000 for the size parameter. + JobSample *JobSample `type:"structure"` + // The name of an Amazon CloudWatch log group, where the job writes diagnostic // messages when it runs. LogGroupName *string `min:"1" type:"string"` @@ -6048,7 +6075,7 @@ type DescribeJobRunOutput struct { // The unique identifier of the job run. RunId *string `min:"1" type:"string"` - // The Amazon Resource Name (ARN) of the user who initiated the job run. + // The Amazon Resource Name (ARN) of the user who started the job run. StartedBy *string `type:"string"` // The date and time when the job run began. @@ -6104,6 +6131,12 @@ func (s *DescribeJobRunOutput) SetJobName(v string) *DescribeJobRunOutput { return s } +// SetJobSample sets the JobSample field's value. +func (s *DescribeJobRunOutput) SetJobSample(v *JobSample) *DescribeJobRunOutput { + s.JobSample = v + return s +} + // SetLogGroupName sets the LogGroupName field's value. func (s *DescribeJobRunOutput) SetLogGroupName(v string) *DescribeJobRunOutput { s.LogGroupName = &v @@ -6923,6 +6956,12 @@ type Job struct { // * SSE-S3 - Server-side encryption with keys managed by Amazon S3. EncryptionMode *string `type:"string" enum:"EncryptionMode"` + // Sample configuration for profile jobs only. Determines the number of rows + // on which the profile job will be executed. If a JobSample value is not provided, + // the default value will be used. The default value is CUSTOM_ROWS for the + // mode parameter and 20000 for the size parameter. + JobSample *JobSample `type:"structure"` + // The Amazon Resource Name (ARN) of the user who last modified the job. LastModifiedBy *string `type:"string"` @@ -7021,6 +7060,12 @@ func (s *Job) SetEncryptionMode(v string) *Job { return s } +// SetJobSample sets the JobSample field's value. +func (s *Job) SetJobSample(v *JobSample) *Job { + s.JobSample = v + return s +} + // SetLastModifiedBy sets the LastModifiedBy field's value. func (s *Job) SetLastModifiedBy(v string) *Job { s.LastModifiedBy = &v @@ -7128,6 +7173,12 @@ type JobRun struct { // The name of the job being processed during this run. JobName *string `min:"1" type:"string"` + // Sample configuration for profile jobs only. Determines the number of rows + // on which the profile job will be executed. If a JobSample value is not provided, + // the default value will be used. The default value is CUSTOM_ROWS for the + // mode parameter and 20000 for the size parameter. + JobSample *JobSample `type:"structure"` + // The name of an Amazon CloudWatch log group, where the job writes diagnostic // messages when it runs. LogGroupName *string `min:"1" type:"string"` @@ -7200,6 +7251,12 @@ func (s *JobRun) SetJobName(v string) *JobRun { return s } +// SetJobSample sets the JobSample field's value. +func (s *JobRun) SetJobSample(v *JobSample) *JobRun { + s.JobSample = v + return s +} + // SetLogGroupName sets the LogGroupName field's value. func (s *JobRun) SetLogGroupName(v string) *JobRun { s.LogGroupName = &v @@ -7248,6 +7305,52 @@ func (s *JobRun) SetState(v string) *JobRun { return s } +// Sample configuration for Profile Jobs only. Determines the number of rows +// on which the Profile job will be executed. If a JobSample value is not provided +// for profile jobs, the default value will be used. The default value is CUSTOM_ROWS +// for the mode parameter and 20000 for the size parameter. +type JobSample struct { + _ struct{} `type:"structure"` + + // Determines whether the profile job will be executed on the entire dataset + // or on a specified number of rows. Must be one of the following: + // + // * FULL_DATASET: Profile job will be executed on the entire dataset. + // + // * CUSTOM_ROWS: Profile job will be executed on the number of rows specified + // in the Size parameter. + Mode *string `type:"string" enum:"SampleMode"` + + // Size parameter is only required when the mode is CUSTOM_ROWS. Profile job + // will be executed on the the specified number of rows. The maximum value for + // size is Long.MAX_VALUE. + // + // Long.MAX_VALUE = 9223372036854775807 + Size *int64 `type:"long"` +} + +// String returns the string representation +func (s JobSample) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s JobSample) GoString() string { + return s.String() +} + +// SetMode sets the Mode field's value. +func (s *JobSample) SetMode(v string) *JobSample { + s.Mode = &v + return s +} + +// SetSize sets the Size field's value. +func (s *JobSample) SetSize(v int64) *JobSample { + s.Size = &v + return s +} + // Represents the JSON-specific options that define how input is to be interpreted // by AWS Glue DataBrew. type JsonOptions struct { @@ -9691,11 +9794,17 @@ type UpdateProfileJobInput struct { // The encryption mode for the job, which can be one of the following: // - // * SSE-KMS - Server-side encryption with AWS KMS-managed keys. + // * SSE-KMS - Server-side encryption with keys managed by AWS KMS. // // * SSE-S3 - Server-side encryption with keys managed by Amazon S3. EncryptionMode *string `type:"string" enum:"EncryptionMode"` + // Sample configuration for Profile Jobs only. Determines the number of rows + // on which the Profile job will be executed. If a JobSample value is not provided + // for profile jobs, the default value will be used. The default value is CUSTOM_ROWS + // for the mode parameter and 20000 for the size parameter. + JobSample *JobSample `type:"structure"` + // Enables or disables Amazon CloudWatch logging for the job. If logging is // enabled, CloudWatch writes one log stream for each job run. LogSubscription *string `type:"string" enum:"LogSubscription"` @@ -9784,6 +9893,12 @@ func (s *UpdateProfileJobInput) SetEncryptionMode(v string) *UpdateProfileJobInp return s } +// SetJobSample sets the JobSample field's value. +func (s *UpdateProfileJobInput) SetJobSample(v *JobSample) *UpdateProfileJobInput { + s.JobSample = v + return s +} + // SetLogSubscription sets the LogSubscription field's value. func (s *UpdateProfileJobInput) SetLogSubscription(v string) *UpdateProfileJobInput { s.LogSubscription = &v @@ -10037,7 +10152,7 @@ type UpdateRecipeJobInput struct { // The encryption mode for the job, which can be one of the following: // - // * SSE-KMS - Server-side encryption with AWS KMS-managed keys. + // * SSE-KMS - Server-side encryption with keys managed by AWS KMS. // // * SSE-S3 - Server-side encryption with keys managed by Amazon S3. EncryptionMode *string `type:"string" enum:"EncryptionMode"` @@ -10601,6 +10716,22 @@ func OutputFormat_Values() []string { } } +const ( + // SampleModeFullDataset is a SampleMode enum value + SampleModeFullDataset = "FULL_DATASET" + + // SampleModeCustomRows is a SampleMode enum value + SampleModeCustomRows = "CUSTOM_ROWS" +) + +// SampleMode_Values returns all elements of the SampleMode enum +func SampleMode_Values() []string { + return []string{ + SampleModeFullDataset, + SampleModeCustomRows, + } +} + const ( // SampleTypeFirstN is a SampleType enum value SampleTypeFirstN = "FIRST_N" diff --git a/service/rds/api.go b/service/rds/api.go index 2600aa252a0..11723bf9223 100644 --- a/service/rds/api.go +++ b/service/rds/api.go @@ -9715,6 +9715,112 @@ func (c *RDS) FailoverDBClusterWithContext(ctx aws.Context, input *FailoverDBClu return out, req.Send() } +const opFailoverGlobalCluster = "FailoverGlobalCluster" + +// FailoverGlobalClusterRequest generates a "aws/request.Request" representing the +// client's request for the FailoverGlobalCluster operation. The "output" return +// value will be populated with the request's response once the request completes +// successfully. +// +// Use "Send" method on the returned Request to send the API call to the service. +// the "output" return value is not valid until after Send returns without error. +// +// See FailoverGlobalCluster for more information on using the FailoverGlobalCluster +// API call, and error handling. +// +// This method is useful when you want to inject custom logic or configuration +// into the SDK's request lifecycle. Such as custom headers, or retry logic. +// +// +// // Example sending a request using the FailoverGlobalClusterRequest method. +// req, resp := client.FailoverGlobalClusterRequest(params) +// +// err := req.Send() +// if err == nil { // resp is now filled +// fmt.Println(resp) +// } +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/rds-2014-10-31/FailoverGlobalCluster +func (c *RDS) FailoverGlobalClusterRequest(input *FailoverGlobalClusterInput) (req *request.Request, output *FailoverGlobalClusterOutput) { + op := &request.Operation{ + Name: opFailoverGlobalCluster, + HTTPMethod: "POST", + HTTPPath: "/", + } + + if input == nil { + input = &FailoverGlobalClusterInput{} + } + + output = &FailoverGlobalClusterOutput{} + req = c.newRequest(op, input, output) + return +} + +// FailoverGlobalCluster API operation for Amazon Relational Database Service. +// +// Initiates the failover process for an Aurora global database (GlobalCluster). +// +// A failover for an Aurora global database promotes one of secondary read-only +// DB clusters to be the primary DB cluster and demotes the primary DB cluster +// to being a secondary (read-only) DB cluster. In other words, the role of +// the current primary DB cluster and the selected (target) DB cluster are switched. +// The selected secondary DB cluster assumes full read/write capabilities for +// the Aurora global database. +// +// For more information about failing over an Amazon Aurora global database, +// see Managed planned failover for Amazon Aurora global databases (https://docs.aws.amazon.com/AmazonRDS/latest/AuroraUserGuide/aurora-global-database-disaster-recovery.managed-failover) +// in the Amazon Aurora User Guide. +// +// This action applies to GlobalCluster (Aurora global databases) only. Use +// this action only on healthy Aurora global databases with running Aurora DB +// clusters and no Region-wide outages, to test disaster recovery scenarios +// or to reconfigure your Aurora global database topology. +// +// Returns awserr.Error for service API and SDK errors. Use runtime type assertions +// with awserr.Error's Code and Message methods to get detailed information about +// the error. +// +// See the AWS API reference guide for Amazon Relational Database Service's +// API operation FailoverGlobalCluster for usage and error information. +// +// Returned Error Codes: +// * ErrCodeGlobalClusterNotFoundFault "GlobalClusterNotFoundFault" +// The GlobalClusterIdentifier doesn't refer to an existing global database +// cluster. +// +// * ErrCodeInvalidGlobalClusterStateFault "InvalidGlobalClusterStateFault" +// The global cluster is in an invalid state and can't perform the requested +// operation. +// +// * ErrCodeInvalidDBClusterStateFault "InvalidDBClusterStateFault" +// The requested operation can't be performed while the cluster is in this state. +// +// * ErrCodeDBClusterNotFoundFault "DBClusterNotFoundFault" +// DBClusterIdentifier doesn't refer to an existing DB cluster. +// +// See also, https://docs.aws.amazon.com/goto/WebAPI/rds-2014-10-31/FailoverGlobalCluster +func (c *RDS) FailoverGlobalCluster(input *FailoverGlobalClusterInput) (*FailoverGlobalClusterOutput, error) { + req, out := c.FailoverGlobalClusterRequest(input) + return out, req.Send() +} + +// FailoverGlobalClusterWithContext is the same as FailoverGlobalCluster with the addition of +// the ability to pass a context and additional request options. +// +// See FailoverGlobalCluster for details on how to use this API operation. +// +// The context must be non-nil and will be used for request cancellation. If +// the context is nil a panic will occur. In the future the SDK may create +// sub-contexts for http.Requests. See https://golang.org/pkg/context/ +// for more information on using Contexts. +func (c *RDS) FailoverGlobalClusterWithContext(ctx aws.Context, input *FailoverGlobalClusterInput, opts ...request.Option) (*FailoverGlobalClusterOutput, error) { + req, out := c.FailoverGlobalClusterRequest(input) + req.SetContext(ctx) + req.ApplyOptions(opts...) + return out, req.Send() +} + const opImportInstallationMedia = "ImportInstallationMedia" // ImportInstallationMediaRequest generates a "aws/request.Request" representing the @@ -17316,11 +17422,17 @@ type CreateDBClusterInput struct { // Possible values are postgresql and upgrade. EnableCloudwatchLogsExports []*string `type:"list"` - // A value that indicates whether to enable write operations to be forwarded - // from this cluster to the primary cluster in an Aurora global database. The - // resulting changes are replicated back to this cluster. This parameter only - // applies to DB clusters that are secondary clusters in an Aurora global database. - // By default, Aurora disallows write operations for secondary clusters. + // A value that indicates whether to enable this DB cluster to forward write + // operations to the primary cluster of an Aurora global database (GlobalCluster). + // By default, write operations are not allowed on Aurora DB clusters that are + // secondary clusters in an Aurora global database. + // + // You can set this value only on Aurora DB clusters that are members of an + // Aurora global database. With this parameter enabled, a secondary cluster + // can forward writes to the current primary cluster and the resulting changes + // are replicated back to this cluster. For the primary DB cluster of an Aurora + // global database, this value is used immediately if the primary is demoted + // by the FailoverGlobalCluster API operation, but it does nothing until then. EnableGlobalWriteForwarding *bool `type:"boolean"` // A value that indicates whether to enable the HTTP endpoint for an Aurora @@ -31901,6 +32013,159 @@ func (s *FailoverDBClusterOutput) SetDBCluster(v *DBCluster) *FailoverDBClusterO return s } +type FailoverGlobalClusterInput struct { + _ struct{} `type:"structure"` + + // Identifier of the Aurora global database (GlobalCluster) that should be failed + // over. The identifier is the unique key assigned by the user when the Aurora + // global database was created. In other words, it's the name of the Aurora + // global database that you want to fail over. + // + // Constraints: + // + // * Must match the identifier of an existing GlobalCluster (Aurora global + // database). + // + // GlobalClusterIdentifier is a required field + GlobalClusterIdentifier *string `min:"1" type:"string" required:"true"` + + // Identifier of the secondary Aurora DB cluster that you want to promote to + // primary for the Aurora global database (GlobalCluster.) Use the Amazon Resource + // Name (ARN) for the identifier so that Aurora can locate the cluster in its + // AWS Region. + // + // TargetDbClusterIdentifier is a required field + TargetDbClusterIdentifier *string `min:"1" type:"string" required:"true"` +} + +// String returns the string representation +func (s FailoverGlobalClusterInput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s FailoverGlobalClusterInput) GoString() string { + return s.String() +} + +// Validate inspects the fields of the type to determine if they are valid. +func (s *FailoverGlobalClusterInput) Validate() error { + invalidParams := request.ErrInvalidParams{Context: "FailoverGlobalClusterInput"} + if s.GlobalClusterIdentifier == nil { + invalidParams.Add(request.NewErrParamRequired("GlobalClusterIdentifier")) + } + if s.GlobalClusterIdentifier != nil && len(*s.GlobalClusterIdentifier) < 1 { + invalidParams.Add(request.NewErrParamMinLen("GlobalClusterIdentifier", 1)) + } + if s.TargetDbClusterIdentifier == nil { + invalidParams.Add(request.NewErrParamRequired("TargetDbClusterIdentifier")) + } + if s.TargetDbClusterIdentifier != nil && len(*s.TargetDbClusterIdentifier) < 1 { + invalidParams.Add(request.NewErrParamMinLen("TargetDbClusterIdentifier", 1)) + } + + if invalidParams.Len() > 0 { + return invalidParams + } + return nil +} + +// SetGlobalClusterIdentifier sets the GlobalClusterIdentifier field's value. +func (s *FailoverGlobalClusterInput) SetGlobalClusterIdentifier(v string) *FailoverGlobalClusterInput { + s.GlobalClusterIdentifier = &v + return s +} + +// SetTargetDbClusterIdentifier sets the TargetDbClusterIdentifier field's value. +func (s *FailoverGlobalClusterInput) SetTargetDbClusterIdentifier(v string) *FailoverGlobalClusterInput { + s.TargetDbClusterIdentifier = &v + return s +} + +type FailoverGlobalClusterOutput struct { + _ struct{} `type:"structure"` + + // A data type representing an Aurora global database. + GlobalCluster *GlobalCluster `type:"structure"` +} + +// String returns the string representation +func (s FailoverGlobalClusterOutput) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s FailoverGlobalClusterOutput) GoString() string { + return s.String() +} + +// SetGlobalCluster sets the GlobalCluster field's value. +func (s *FailoverGlobalClusterOutput) SetGlobalCluster(v *GlobalCluster) *FailoverGlobalClusterOutput { + s.GlobalCluster = v + return s +} + +// Contains the state of scheduled or in-process failover operations on an Aurora +// global database (GlobalCluster). This Data type is empty unless a failover +// operation is scheduled or is currently underway on the Aurora global database. +type FailoverState struct { + _ struct{} `type:"structure"` + + // The Amazon Resource Name (ARN) of the Aurora DB cluster that is currently + // being demoted, and which is associated with this state. + FromDbClusterArn *string `type:"string"` + + // The current status of the Aurora global database (GlobalCluster). Possible + // values are as follows: + // + // * pending – A request to fail over the Aurora global database (GlobalCluster) + // has been received by the service. The GlobalCluster's primary DB cluster + // and the specified secondary DB cluster are being verified before the failover + // process can start. + // + // * failing-over – This status covers the range of Aurora internal operations + // that take place during the failover process, such as demoting the primary + // Aurora DB cluster, promoting the secondary Aurora DB, and synchronizing + // replicas. + // + // * cancelling – The request to fail over the Aurora global database (GlobalCluster) + // was cancelled and the primary Aurora DB cluster and the selected secondary + // Aurora DB cluster are returning to their previous states. + Status *string `type:"string" enum:"FailoverStatus"` + + // The Amazon Resource Name (ARN) of the Aurora DB cluster that is currently + // being promoted, and which is associated with this state. + ToDbClusterArn *string `type:"string"` +} + +// String returns the string representation +func (s FailoverState) String() string { + return awsutil.Prettify(s) +} + +// GoString returns the string representation +func (s FailoverState) GoString() string { + return s.String() +} + +// SetFromDbClusterArn sets the FromDbClusterArn field's value. +func (s *FailoverState) SetFromDbClusterArn(v string) *FailoverState { + s.FromDbClusterArn = &v + return s +} + +// SetStatus sets the Status field's value. +func (s *FailoverState) SetStatus(v string) *FailoverState { + s.Status = &v + return s +} + +// SetToDbClusterArn sets the ToDbClusterArn field's value. +func (s *FailoverState) SetToDbClusterArn(v string) *FailoverState { + s.ToDbClusterArn = &v + return s +} + // A filter name and value pair that is used to return a more specific list // of results from a describe operation. Filters can be used to match a set // of resources by specific criteria, such as IDs. The filters supported by @@ -31987,6 +32252,12 @@ type GlobalCluster struct { // Indicates the database engine version. EngineVersion *string `type:"string"` + // A data object containing all properties for the current state of an in-process + // or pending failover process for this Aurora global database. This object + // is empty unless the FailoverGlobalCluster API operation has been called on + // this Aurora global database (GlobalCluster). + FailoverState *FailoverState `type:"structure"` + // The Amazon Resource Name (ARN) for the global database cluster. GlobalClusterArn *string `type:"string"` @@ -32044,6 +32315,12 @@ func (s *GlobalCluster) SetEngineVersion(v string) *GlobalCluster { return s } +// SetFailoverState sets the FailoverState field's value. +func (s *GlobalCluster) SetFailoverState(v *FailoverState) *GlobalCluster { + s.FailoverState = v + return s +} + // SetGlobalClusterArn sets the GlobalClusterArn field's value. func (s *GlobalCluster) SetGlobalClusterArn(v string) *GlobalCluster { s.GlobalClusterArn = &v @@ -33125,11 +33402,17 @@ type ModifyDBClusterInput struct { // Directory Service. DomainIAMRoleName *string `type:"string"` - // A value that indicates whether to enable write operations to be forwarded - // from this cluster to the primary cluster in an Aurora global database. The - // resulting changes are replicated back to this cluster. This parameter only - // applies to DB clusters that are secondary clusters in an Aurora global database. - // By default, Aurora disallows write operations for secondary clusters. + // A value that indicates whether to enable this DB cluster to forward write + // operations to the primary cluster of an Aurora global database (GlobalCluster). + // By default, write operations are not allowed on Aurora DB clusters that are + // secondary clusters in an Aurora global database. + // + // You can set this value only on Aurora DB clusters that are members of an + // Aurora global database. With this parameter enabled, a secondary cluster + // can forward writes to the current primary cluster and the resulting changes + // are replicated back to this cluster. For the primary DB cluster of an Aurora + // global database, this value is used immediately if the primary is demoted + // by the FailoverGlobalCluster API operation, but it does nothing until then. EnableGlobalWriteForwarding *bool `type:"boolean"` // A value that indicates whether to enable the HTTP endpoint for an Aurora @@ -43268,6 +43551,26 @@ func EngineFamily_Values() []string { } } +const ( + // FailoverStatusPending is a FailoverStatus enum value + FailoverStatusPending = "pending" + + // FailoverStatusFailingOver is a FailoverStatus enum value + FailoverStatusFailingOver = "failing-over" + + // FailoverStatusCancelling is a FailoverStatus enum value + FailoverStatusCancelling = "cancelling" +) + +// FailoverStatus_Values returns all elements of the FailoverStatus enum +func FailoverStatus_Values() []string { + return []string{ + FailoverStatusPending, + FailoverStatusFailingOver, + FailoverStatusCancelling, + } +} + const ( // IAMAuthModeDisabled is a IAMAuthMode enum value IAMAuthModeDisabled = "DISABLED" diff --git a/service/rds/rdsiface/interface.go b/service/rds/rdsiface/interface.go index 23415dee77e..a9cc0eec97d 100644 --- a/service/rds/rdsiface/interface.go +++ b/service/rds/rdsiface/interface.go @@ -506,6 +506,10 @@ type RDSAPI interface { FailoverDBClusterWithContext(aws.Context, *rds.FailoverDBClusterInput, ...request.Option) (*rds.FailoverDBClusterOutput, error) FailoverDBClusterRequest(*rds.FailoverDBClusterInput) (*request.Request, *rds.FailoverDBClusterOutput) + FailoverGlobalCluster(*rds.FailoverGlobalClusterInput) (*rds.FailoverGlobalClusterOutput, error) + FailoverGlobalClusterWithContext(aws.Context, *rds.FailoverGlobalClusterInput, ...request.Option) (*rds.FailoverGlobalClusterOutput, error) + FailoverGlobalClusterRequest(*rds.FailoverGlobalClusterInput) (*request.Request, *rds.FailoverGlobalClusterOutput) + ImportInstallationMedia(*rds.ImportInstallationMediaInput) (*rds.ImportInstallationMediaOutput, error) ImportInstallationMediaWithContext(aws.Context, *rds.ImportInstallationMediaInput, ...request.Option) (*rds.ImportInstallationMediaOutput, error) ImportInstallationMediaRequest(*rds.ImportInstallationMediaInput) (*request.Request, *rds.ImportInstallationMediaOutput)