diff --git a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/_models.py b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/_models.py index ef7778d653b5..a056f9732314 100644 --- a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/_models.py +++ b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/_models.py @@ -2283,19 +2283,23 @@ def __init__(self, **kwargs): class StoreWriteSettings(Model): """Connector write settings. + You probably want to use the sub-classes and not this class directly. Known + sub-classes are: FileServerWriteSettings, AzureDataLakeStoreWriteSettings, + AzureBlobFSWriteSettings, AzureBlobStorageWriteSettings + All required parameters must be populated in order to send to Azure. :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] - :param type: Required. The write setting type. - :type type: str :param max_concurrent_connections: The maximum concurrent connection count for the source data store. Type: integer (or Expression with resultType integer). :type max_concurrent_connections: object :param copy_behavior: The type of copy behavior for copy sink. :type copy_behavior: object + :param type: Required. Constant filled by server. + :type type: str """ _validation = { @@ -2304,17 +2308,21 @@ class StoreWriteSettings(Model): _attribute_map = { 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'copy_behavior': {'key': 'copyBehavior', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, + } + + _subtype_map = { + 'type': {'FileServerWriteSettings': 'FileServerWriteSettings', 'AzureDataLakeStoreWriteSettings': 'AzureDataLakeStoreWriteSettings', 'AzureBlobFSWriteSettings': 'AzureBlobFSWriteSettings', 'AzureBlobStorageWriteSettings': 'AzureBlobStorageWriteSettings'} } def __init__(self, **kwargs): super(StoreWriteSettings, self).__init__(**kwargs) self.additional_properties = kwargs.get('additional_properties', None) - self.type = kwargs.get('type', None) self.max_concurrent_connections = kwargs.get('max_concurrent_connections', None) self.copy_behavior = kwargs.get('copy_behavior', None) + self.type = None class AzureBlobFSWriteSettings(StoreWriteSettings): @@ -2325,14 +2333,17 @@ class AzureBlobFSWriteSettings(StoreWriteSettings): :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] - :param type: Required. The write setting type. - :type type: str :param max_concurrent_connections: The maximum concurrent connection count for the source data store. Type: integer (or Expression with resultType integer). :type max_concurrent_connections: object :param copy_behavior: The type of copy behavior for copy sink. :type copy_behavior: object + :param type: Required. Constant filled by server. + :type type: str + :param block_size_in_mb: Indicates the block size(MB) when writing data to + blob. Type: integer (or Expression with resultType integer). + :type block_size_in_mb: object """ _validation = { @@ -2341,13 +2352,16 @@ class AzureBlobFSWriteSettings(StoreWriteSettings): _attribute_map = { 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'copy_behavior': {'key': 'copyBehavior', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, + 'block_size_in_mb': {'key': 'blockSizeInMB', 'type': 'object'}, } def __init__(self, **kwargs): super(AzureBlobFSWriteSettings, self).__init__(**kwargs) + self.block_size_in_mb = kwargs.get('block_size_in_mb', None) + self.type = 'AzureBlobFSWriteSettings' class AzureBlobStorageLinkedService(LinkedService): @@ -2549,14 +2563,17 @@ class AzureBlobStorageWriteSettings(StoreWriteSettings): :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] - :param type: Required. The write setting type. - :type type: str :param max_concurrent_connections: The maximum concurrent connection count for the source data store. Type: integer (or Expression with resultType integer). :type max_concurrent_connections: object :param copy_behavior: The type of copy behavior for copy sink. :type copy_behavior: object + :param type: Required. Constant filled by server. + :type type: str + :param block_size_in_mb: Indicates the block size(MB) when writing data to + blob. Type: integer (or Expression with resultType integer). + :type block_size_in_mb: object """ _validation = { @@ -2565,13 +2582,16 @@ class AzureBlobStorageWriteSettings(StoreWriteSettings): _attribute_map = { 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'copy_behavior': {'key': 'copyBehavior', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, + 'block_size_in_mb': {'key': 'blockSizeInMB', 'type': 'object'}, } def __init__(self, **kwargs): super(AzureBlobStorageWriteSettings, self).__init__(**kwargs) + self.block_size_in_mb = kwargs.get('block_size_in_mb', None) + self.type = 'AzureBlobStorageWriteSettings' class AzureDatabricksLinkedService(LinkedService): @@ -3518,14 +3538,14 @@ class AzureDataLakeStoreWriteSettings(StoreWriteSettings): :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] - :param type: Required. The write setting type. - :type type: str :param max_concurrent_connections: The maximum concurrent connection count for the source data store. Type: integer (or Expression with resultType integer). :type max_concurrent_connections: object :param copy_behavior: The type of copy behavior for copy sink. :type copy_behavior: object + :param type: Required. Constant filled by server. + :type type: str """ _validation = { @@ -3534,13 +3554,14 @@ class AzureDataLakeStoreWriteSettings(StoreWriteSettings): _attribute_map = { 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'copy_behavior': {'key': 'copyBehavior', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, } def __init__(self, **kwargs): super(AzureDataLakeStoreWriteSettings, self).__init__(**kwargs) + self.type = 'AzureDataLakeStoreWriteSettings' class AzureFunctionActivity(ExecutionActivity): @@ -11061,14 +11082,14 @@ class FileServerWriteSettings(StoreWriteSettings): :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] - :param type: Required. The write setting type. - :type type: str :param max_concurrent_connections: The maximum concurrent connection count for the source data store. Type: integer (or Expression with resultType integer). :type max_concurrent_connections: object :param copy_behavior: The type of copy behavior for copy sink. :type copy_behavior: object + :param type: Required. Constant filled by server. + :type type: str """ _validation = { @@ -11077,13 +11098,14 @@ class FileServerWriteSettings(StoreWriteSettings): _attribute_map = { 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'copy_behavior': {'key': 'copyBehavior', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, } def __init__(self, **kwargs): super(FileServerWriteSettings, self).__init__(**kwargs) + self.type = 'FileServerWriteSettings' class FileShareDataset(Dataset): diff --git a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/_models_py3.py b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/_models_py3.py index 295f650553e3..6449023a7260 100644 --- a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/_models_py3.py +++ b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/_models_py3.py @@ -2283,19 +2283,23 @@ def __init__(self, *, additional_properties=None, source_retry_count=None, sourc class StoreWriteSettings(Model): """Connector write settings. + You probably want to use the sub-classes and not this class directly. Known + sub-classes are: FileServerWriteSettings, AzureDataLakeStoreWriteSettings, + AzureBlobFSWriteSettings, AzureBlobStorageWriteSettings + All required parameters must be populated in order to send to Azure. :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] - :param type: Required. The write setting type. - :type type: str :param max_concurrent_connections: The maximum concurrent connection count for the source data store. Type: integer (or Expression with resultType integer). :type max_concurrent_connections: object :param copy_behavior: The type of copy behavior for copy sink. :type copy_behavior: object + :param type: Required. Constant filled by server. + :type type: str """ _validation = { @@ -2304,17 +2308,21 @@ class StoreWriteSettings(Model): _attribute_map = { 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'copy_behavior': {'key': 'copyBehavior', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, + } + + _subtype_map = { + 'type': {'FileServerWriteSettings': 'FileServerWriteSettings', 'AzureDataLakeStoreWriteSettings': 'AzureDataLakeStoreWriteSettings', 'AzureBlobFSWriteSettings': 'AzureBlobFSWriteSettings', 'AzureBlobStorageWriteSettings': 'AzureBlobStorageWriteSettings'} } - def __init__(self, *, type: str, additional_properties=None, max_concurrent_connections=None, copy_behavior=None, **kwargs) -> None: + def __init__(self, *, additional_properties=None, max_concurrent_connections=None, copy_behavior=None, **kwargs) -> None: super(StoreWriteSettings, self).__init__(**kwargs) self.additional_properties = additional_properties - self.type = type self.max_concurrent_connections = max_concurrent_connections self.copy_behavior = copy_behavior + self.type = None class AzureBlobFSWriteSettings(StoreWriteSettings): @@ -2325,14 +2333,17 @@ class AzureBlobFSWriteSettings(StoreWriteSettings): :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] - :param type: Required. The write setting type. - :type type: str :param max_concurrent_connections: The maximum concurrent connection count for the source data store. Type: integer (or Expression with resultType integer). :type max_concurrent_connections: object :param copy_behavior: The type of copy behavior for copy sink. :type copy_behavior: object + :param type: Required. Constant filled by server. + :type type: str + :param block_size_in_mb: Indicates the block size(MB) when writing data to + blob. Type: integer (or Expression with resultType integer). + :type block_size_in_mb: object """ _validation = { @@ -2341,13 +2352,16 @@ class AzureBlobFSWriteSettings(StoreWriteSettings): _attribute_map = { 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'copy_behavior': {'key': 'copyBehavior', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, + 'block_size_in_mb': {'key': 'blockSizeInMB', 'type': 'object'}, } - def __init__(self, *, type: str, additional_properties=None, max_concurrent_connections=None, copy_behavior=None, **kwargs) -> None: - super(AzureBlobFSWriteSettings, self).__init__(additional_properties=additional_properties, type=type, max_concurrent_connections=max_concurrent_connections, copy_behavior=copy_behavior, **kwargs) + def __init__(self, *, additional_properties=None, max_concurrent_connections=None, copy_behavior=None, block_size_in_mb=None, **kwargs) -> None: + super(AzureBlobFSWriteSettings, self).__init__(additional_properties=additional_properties, max_concurrent_connections=max_concurrent_connections, copy_behavior=copy_behavior, **kwargs) + self.block_size_in_mb = block_size_in_mb + self.type = 'AzureBlobFSWriteSettings' class AzureBlobStorageLinkedService(LinkedService): @@ -2549,14 +2563,17 @@ class AzureBlobStorageWriteSettings(StoreWriteSettings): :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] - :param type: Required. The write setting type. - :type type: str :param max_concurrent_connections: The maximum concurrent connection count for the source data store. Type: integer (or Expression with resultType integer). :type max_concurrent_connections: object :param copy_behavior: The type of copy behavior for copy sink. :type copy_behavior: object + :param type: Required. Constant filled by server. + :type type: str + :param block_size_in_mb: Indicates the block size(MB) when writing data to + blob. Type: integer (or Expression with resultType integer). + :type block_size_in_mb: object """ _validation = { @@ -2565,13 +2582,16 @@ class AzureBlobStorageWriteSettings(StoreWriteSettings): _attribute_map = { 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'copy_behavior': {'key': 'copyBehavior', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, + 'block_size_in_mb': {'key': 'blockSizeInMB', 'type': 'object'}, } - def __init__(self, *, type: str, additional_properties=None, max_concurrent_connections=None, copy_behavior=None, **kwargs) -> None: - super(AzureBlobStorageWriteSettings, self).__init__(additional_properties=additional_properties, type=type, max_concurrent_connections=max_concurrent_connections, copy_behavior=copy_behavior, **kwargs) + def __init__(self, *, additional_properties=None, max_concurrent_connections=None, copy_behavior=None, block_size_in_mb=None, **kwargs) -> None: + super(AzureBlobStorageWriteSettings, self).__init__(additional_properties=additional_properties, max_concurrent_connections=max_concurrent_connections, copy_behavior=copy_behavior, **kwargs) + self.block_size_in_mb = block_size_in_mb + self.type = 'AzureBlobStorageWriteSettings' class AzureDatabricksLinkedService(LinkedService): @@ -3518,14 +3538,14 @@ class AzureDataLakeStoreWriteSettings(StoreWriteSettings): :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] - :param type: Required. The write setting type. - :type type: str :param max_concurrent_connections: The maximum concurrent connection count for the source data store. Type: integer (or Expression with resultType integer). :type max_concurrent_connections: object :param copy_behavior: The type of copy behavior for copy sink. :type copy_behavior: object + :param type: Required. Constant filled by server. + :type type: str """ _validation = { @@ -3534,13 +3554,14 @@ class AzureDataLakeStoreWriteSettings(StoreWriteSettings): _attribute_map = { 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'copy_behavior': {'key': 'copyBehavior', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, } - def __init__(self, *, type: str, additional_properties=None, max_concurrent_connections=None, copy_behavior=None, **kwargs) -> None: - super(AzureDataLakeStoreWriteSettings, self).__init__(additional_properties=additional_properties, type=type, max_concurrent_connections=max_concurrent_connections, copy_behavior=copy_behavior, **kwargs) + def __init__(self, *, additional_properties=None, max_concurrent_connections=None, copy_behavior=None, **kwargs) -> None: + super(AzureDataLakeStoreWriteSettings, self).__init__(additional_properties=additional_properties, max_concurrent_connections=max_concurrent_connections, copy_behavior=copy_behavior, **kwargs) + self.type = 'AzureDataLakeStoreWriteSettings' class AzureFunctionActivity(ExecutionActivity): @@ -11061,14 +11082,14 @@ class FileServerWriteSettings(StoreWriteSettings): :param additional_properties: Unmatched properties from the message are deserialized this collection :type additional_properties: dict[str, object] - :param type: Required. The write setting type. - :type type: str :param max_concurrent_connections: The maximum concurrent connection count for the source data store. Type: integer (or Expression with resultType integer). :type max_concurrent_connections: object :param copy_behavior: The type of copy behavior for copy sink. :type copy_behavior: object + :param type: Required. Constant filled by server. + :type type: str """ _validation = { @@ -11077,13 +11098,14 @@ class FileServerWriteSettings(StoreWriteSettings): _attribute_map = { 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'copy_behavior': {'key': 'copyBehavior', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, } - def __init__(self, *, type: str, additional_properties=None, max_concurrent_connections=None, copy_behavior=None, **kwargs) -> None: - super(FileServerWriteSettings, self).__init__(additional_properties=additional_properties, type=type, max_concurrent_connections=max_concurrent_connections, copy_behavior=copy_behavior, **kwargs) + def __init__(self, *, additional_properties=None, max_concurrent_connections=None, copy_behavior=None, **kwargs) -> None: + super(FileServerWriteSettings, self).__init__(additional_properties=additional_properties, max_concurrent_connections=max_concurrent_connections, copy_behavior=copy_behavior, **kwargs) + self.type = 'FileServerWriteSettings' class FileShareDataset(Dataset):