From 0135d39ae1a49132681f2178548e60d16674a68d Mon Sep 17 00:00:00 2001 From: Azure SDK for Python bot Date: Thu, 20 Jun 2019 16:14:10 +0000 Subject: [PATCH 1/2] Generated from b0ddfd5a2aefefdca6d220fd03714b3fdfc779a6 modify swagger --- .../azure/mgmt/datafactory/models/__init__.py | 25 +++--- .../models/azure_search_index_sink.py | 7 +- .../models/azure_search_index_sink_py3.py | 7 +- .../datafactory/models/cassandra_source.py | 9 ++- .../models/cassandra_source_py3.py | 9 ++- .../mgmt/datafactory/models/copy_source.py | 10 +-- .../datafactory/models/copy_source_py3.py | 10 +-- .../data_factory_management_client_enums.py | 38 +++++++++ .../azure/mgmt/datafactory/models/dataset.py | 4 +- .../mgmt/datafactory/models/dataset_py3.py | 4 +- .../mgmt/datafactory/models/dynamics_sink.py | 7 +- .../datafactory/models/dynamics_sink_py3.py | 7 +- .../models/oracle_partition_settings.py | 46 ----------- .../models/oracle_partition_settings_py3.py | 46 ----------- .../mgmt/datafactory/models/oracle_source.py | 11 --- .../datafactory/models/oracle_source_py3.py | 13 +--- .../datafactory/models/salesforce_sink.py | 7 +- .../datafactory/models/salesforce_sink_py3.py | 7 +- .../datafactory/models/salesforce_source.py | 7 +- .../models/salesforce_source_py3.py | 7 +- .../models/sap_cloud_for_customer_sink.py | 7 +- .../models/sap_cloud_for_customer_sink_py3.py | 7 +- .../models/sap_table_partition_settings.py | 47 ----------- .../sap_table_partition_settings_py3.py | 47 ----------- .../datafactory/models/sap_table_source.py | 11 --- .../models/sap_table_source_py3.py | 13 +--- .../models/teradata_linked_service.py | 10 +-- .../models/teradata_linked_service_py3.py | 12 +-- .../models/teradata_partition_settings.py | 42 ---------- .../models/teradata_partition_settings_py3.py | 42 ---------- .../datafactory/models/teradata_source.py | 68 ---------------- .../datafactory/models/teradata_source_py3.py | 68 ---------------- .../models/teradata_table_dataset.py | 77 ------------------- .../models/teradata_table_dataset_py3.py | 77 ------------------- 34 files changed, 123 insertions(+), 686 deletions(-) delete mode 100644 sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_partition_settings.py delete mode 100644 sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_partition_settings_py3.py delete mode 100644 sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_table_partition_settings.py delete mode 100644 sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_table_partition_settings_py3.py delete mode 100644 sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/teradata_partition_settings.py delete mode 100644 sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/teradata_partition_settings_py3.py delete mode 100644 sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/teradata_source.py delete mode 100644 sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/teradata_source_py3.py delete mode 100644 sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/teradata_table_dataset.py delete mode 100644 sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/teradata_table_dataset_py3.py diff --git a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py index 4414f045cddd..70ebe10afa94 100644 --- a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py +++ b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py @@ -241,7 +241,6 @@ from .salesforce_object_dataset_py3 import SalesforceObjectDataset from .relational_table_dataset_py3 import RelationalTableDataset from .azure_my_sql_table_dataset_py3 import AzureMySqlTableDataset - from .teradata_table_dataset_py3 import TeradataTableDataset from .oracle_table_dataset_py3 import OracleTableDataset from .odata_resource_dataset_py3 import ODataResourceDataset from .cosmos_db_mongo_db_api_collection_dataset_py3 import CosmosDbMongoDbApiCollectionDataset @@ -329,9 +328,6 @@ from .mongo_db_source_py3 import MongoDbSource from .cassandra_source_py3 import CassandraSource from .web_source_py3 import WebSource - from .teradata_partition_settings_py3 import TeradataPartitionSettings - from .teradata_source_py3 import TeradataSource - from .oracle_partition_settings_py3 import OraclePartitionSettings from .oracle_source_py3 import OracleSource from .azure_data_explorer_source_py3 import AzureDataExplorerSource from .azure_my_sql_source_py3 import AzureMySqlSource @@ -344,7 +340,6 @@ from .sql_server_source_py3 import SqlServerSource from .sql_source_py3 import SqlSource from .rest_source_py3 import RestSource - from .sap_table_partition_settings_py3 import SapTablePartitionSettings from .sap_table_source_py3 import SapTableSource from .sap_open_hub_source_py3 import SapOpenHubSource from .sap_hana_source_py3 import SapHanaSource @@ -698,7 +693,6 @@ from .salesforce_object_dataset import SalesforceObjectDataset from .relational_table_dataset import RelationalTableDataset from .azure_my_sql_table_dataset import AzureMySqlTableDataset - from .teradata_table_dataset import TeradataTableDataset from .oracle_table_dataset import OracleTableDataset from .odata_resource_dataset import ODataResourceDataset from .cosmos_db_mongo_db_api_collection_dataset import CosmosDbMongoDbApiCollectionDataset @@ -786,9 +780,6 @@ from .mongo_db_source import MongoDbSource from .cassandra_source import CassandraSource from .web_source import WebSource - from .teradata_partition_settings import TeradataPartitionSettings - from .teradata_source import TeradataSource - from .oracle_partition_settings import OraclePartitionSettings from .oracle_source import OracleSource from .azure_data_explorer_source import AzureDataExplorerSource from .azure_my_sql_source import AzureMySqlSource @@ -801,7 +792,6 @@ from .sql_server_source import SqlServerSource from .sql_source import SqlSource from .rest_source import RestSource - from .sap_table_partition_settings import SapTablePartitionSettings from .sap_table_source import SapTableSource from .sap_open_hub_source import SapOpenHubSource from .sap_hana_source import SapHanaSource @@ -974,9 +964,14 @@ SybaseAuthenticationType, AzureFunctionActivityMethod, WebActivityMethod, + CassandraSourceReadConsistencyLevels, StoredProcedureParameterType, + SalesforceSourceReadBehavior, HDInsightActivityDebugInfoOption, + SalesforceSinkWriteBehavior, + AzureSearchIndexWriteBehaviorType, PolybaseSettingsRejectType, + SapCloudForCustomerSinkWriteBehavior, WebHookActivityMethod, IntegrationRuntimeType, SelfHostedIntegrationRuntimeNodeStatus, @@ -1223,7 +1218,6 @@ 'SalesforceObjectDataset', 'RelationalTableDataset', 'AzureMySqlTableDataset', - 'TeradataTableDataset', 'OracleTableDataset', 'ODataResourceDataset', 'CosmosDbMongoDbApiCollectionDataset', @@ -1311,9 +1305,6 @@ 'MongoDbSource', 'CassandraSource', 'WebSource', - 'TeradataPartitionSettings', - 'TeradataSource', - 'OraclePartitionSettings', 'OracleSource', 'AzureDataExplorerSource', 'AzureMySqlSource', @@ -1326,7 +1317,6 @@ 'SqlServerSource', 'SqlSource', 'RestSource', - 'SapTablePartitionSettings', 'SapTableSource', 'SapOpenHubSource', 'SapHanaSource', @@ -1498,9 +1488,14 @@ 'SybaseAuthenticationType', 'AzureFunctionActivityMethod', 'WebActivityMethod', + 'CassandraSourceReadConsistencyLevels', 'StoredProcedureParameterType', + 'SalesforceSourceReadBehavior', 'HDInsightActivityDebugInfoOption', + 'SalesforceSinkWriteBehavior', + 'AzureSearchIndexWriteBehaviorType', 'PolybaseSettingsRejectType', + 'SapCloudForCustomerSinkWriteBehavior', 'WebHookActivityMethod', 'IntegrationRuntimeType', 'SelfHostedIntegrationRuntimeNodeStatus', diff --git a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_search_index_sink.py b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_search_index_sink.py index af2505be7a5c..9aae64af8da0 100644 --- a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_search_index_sink.py +++ b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_search_index_sink.py @@ -41,8 +41,9 @@ class AzureSearchIndexSink(CopySink): :param type: Required. Constant filled by server. :type type: str :param write_behavior: Specify the write behavior when upserting documents - into Azure Search Index. - :type write_behavior: object + into Azure Search Index. Possible values include: 'Merge', 'Upload' + :type write_behavior: str or + ~azure.mgmt.datafactory.models.AzureSearchIndexWriteBehaviorType """ _validation = { @@ -57,7 +58,7 @@ class AzureSearchIndexSink(CopySink): 'sink_retry_wait': {'key': 'sinkRetryWait', 'type': 'object'}, 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'type': {'key': 'type', 'type': 'str'}, - 'write_behavior': {'key': 'writeBehavior', 'type': 'object'}, + 'write_behavior': {'key': 'writeBehavior', 'type': 'str'}, } def __init__(self, **kwargs): diff --git a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_search_index_sink_py3.py b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_search_index_sink_py3.py index 9e57f2f1feb3..3cd887a2512c 100644 --- a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_search_index_sink_py3.py +++ b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_search_index_sink_py3.py @@ -41,8 +41,9 @@ class AzureSearchIndexSink(CopySink): :param type: Required. Constant filled by server. :type type: str :param write_behavior: Specify the write behavior when upserting documents - into Azure Search Index. - :type write_behavior: object + into Azure Search Index. Possible values include: 'Merge', 'Upload' + :type write_behavior: str or + ~azure.mgmt.datafactory.models.AzureSearchIndexWriteBehaviorType """ _validation = { @@ -57,7 +58,7 @@ class AzureSearchIndexSink(CopySink): 'sink_retry_wait': {'key': 'sinkRetryWait', 'type': 'object'}, 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'type': {'key': 'type', 'type': 'str'}, - 'write_behavior': {'key': 'writeBehavior', 'type': 'object'}, + 'write_behavior': {'key': 'writeBehavior', 'type': 'str'}, } def __init__(self, *, additional_properties=None, write_batch_size=None, write_batch_timeout=None, sink_retry_count=None, sink_retry_wait=None, max_concurrent_connections=None, write_behavior=None, **kwargs) -> None: diff --git a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/cassandra_source.py b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/cassandra_source.py index 8a52f03cd5ba..e7ba96c18682 100644 --- a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/cassandra_source.py +++ b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/cassandra_source.py @@ -42,8 +42,11 @@ class CassandraSource(CopySource): the client application. Cassandra checks the specified number of Cassandra servers for data to satisfy the read request. Must be one of cassandraSourceReadConsistencyLevels. The default value is 'ONE'. It is - case-insensitive. - :type consistency_level: object + case-insensitive. Possible values include: 'ALL', 'EACH_QUORUM', 'QUORUM', + 'LOCAL_QUORUM', 'ONE', 'TWO', 'THREE', 'LOCAL_ONE', 'SERIAL', + 'LOCAL_SERIAL' + :type consistency_level: str or + ~azure.mgmt.datafactory.models.CassandraSourceReadConsistencyLevels """ _validation = { @@ -57,7 +60,7 @@ class CassandraSource(CopySource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'type': {'key': 'type', 'type': 'str'}, 'query': {'key': 'query', 'type': 'object'}, - 'consistency_level': {'key': 'consistencyLevel', 'type': 'object'}, + 'consistency_level': {'key': 'consistencyLevel', 'type': 'str'}, } def __init__(self, **kwargs): diff --git a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/cassandra_source_py3.py b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/cassandra_source_py3.py index 6957385bab86..bd95d158b868 100644 --- a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/cassandra_source_py3.py +++ b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/cassandra_source_py3.py @@ -42,8 +42,11 @@ class CassandraSource(CopySource): the client application. Cassandra checks the specified number of Cassandra servers for data to satisfy the read request. Must be one of cassandraSourceReadConsistencyLevels. The default value is 'ONE'. It is - case-insensitive. - :type consistency_level: object + case-insensitive. Possible values include: 'ALL', 'EACH_QUORUM', 'QUORUM', + 'LOCAL_QUORUM', 'ONE', 'TWO', 'THREE', 'LOCAL_ONE', 'SERIAL', + 'LOCAL_SERIAL' + :type consistency_level: str or + ~azure.mgmt.datafactory.models.CassandraSourceReadConsistencyLevels """ _validation = { @@ -57,7 +60,7 @@ class CassandraSource(CopySource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'type': {'key': 'type', 'type': 'str'}, 'query': {'key': 'query', 'type': 'object'}, - 'consistency_level': {'key': 'consistencyLevel', 'type': 'object'}, + 'consistency_level': {'key': 'consistencyLevel', 'type': 'str'}, } def __init__(self, *, additional_properties=None, source_retry_count=None, source_retry_wait=None, max_concurrent_connections=None, query=None, consistency_level=None, **kwargs) -> None: diff --git a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/copy_source.py b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/copy_source.py index 99918f1e245c..c261c385de8d 100644 --- a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/copy_source.py +++ b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/copy_source.py @@ -26,10 +26,10 @@ class CopySource(Model): EloquaSource, DrillSource, CouchbaseSource, ConcurSource, AzurePostgreSqlSource, AmazonMWSSource, HttpSource, AzureBlobFSSource, AzureDataLakeStoreSource, Office365Source, CosmosDbMongoDbApiSource, - MongoDbV2Source, MongoDbSource, CassandraSource, WebSource, TeradataSource, - OracleSource, AzureDataExplorerSource, AzureMySqlSource, HdfsSource, - FileSystemSource, SqlDWSource, AzureSqlSource, SqlServerSource, SqlSource, - RestSource, SapTableSource, SapOpenHubSource, SapHanaSource, SapEccSource, + MongoDbV2Source, MongoDbSource, CassandraSource, WebSource, OracleSource, + AzureDataExplorerSource, AzureMySqlSource, HdfsSource, FileSystemSource, + SqlDWSource, AzureSqlSource, SqlServerSource, SqlSource, RestSource, + SapTableSource, SapOpenHubSource, SapHanaSource, SapEccSource, SapCloudForCustomerSource, SalesforceSource, RelationalSource, DynamicsSource, DocumentDbCollectionSource, BlobSource, AzureTableSource, DelimitedTextSource, ParquetSource @@ -67,7 +67,7 @@ class CopySource(Model): } _subtype_map = { - 'type': {'AmazonRedshiftSource': 'AmazonRedshiftSource', 'GoogleAdWordsSource': 'GoogleAdWordsSource', 'OracleServiceCloudSource': 'OracleServiceCloudSource', 'DynamicsAXSource': 'DynamicsAXSource', 'ResponsysSource': 'ResponsysSource', 'SalesforceMarketingCloudSource': 'SalesforceMarketingCloudSource', 'VerticaSource': 'VerticaSource', 'NetezzaSource': 'NetezzaSource', 'ZohoSource': 'ZohoSource', 'XeroSource': 'XeroSource', 'SquareSource': 'SquareSource', 'SparkSource': 'SparkSource', 'ShopifySource': 'ShopifySource', 'ServiceNowSource': 'ServiceNowSource', 'QuickBooksSource': 'QuickBooksSource', 'PrestoSource': 'PrestoSource', 'PhoenixSource': 'PhoenixSource', 'PaypalSource': 'PaypalSource', 'MarketoSource': 'MarketoSource', 'MariaDBSource': 'MariaDBSource', 'MagentoSource': 'MagentoSource', 'JiraSource': 'JiraSource', 'ImpalaSource': 'ImpalaSource', 'HubspotSource': 'HubspotSource', 'HiveSource': 'HiveSource', 'HBaseSource': 'HBaseSource', 'GreenplumSource': 'GreenplumSource', 'GoogleBigQuerySource': 'GoogleBigQuerySource', 'EloquaSource': 'EloquaSource', 'DrillSource': 'DrillSource', 'CouchbaseSource': 'CouchbaseSource', 'ConcurSource': 'ConcurSource', 'AzurePostgreSqlSource': 'AzurePostgreSqlSource', 'AmazonMWSSource': 'AmazonMWSSource', 'HttpSource': 'HttpSource', 'AzureBlobFSSource': 'AzureBlobFSSource', 'AzureDataLakeStoreSource': 'AzureDataLakeStoreSource', 'Office365Source': 'Office365Source', 'CosmosDbMongoDbApiSource': 'CosmosDbMongoDbApiSource', 'MongoDbV2Source': 'MongoDbV2Source', 'MongoDbSource': 'MongoDbSource', 'CassandraSource': 'CassandraSource', 'WebSource': 'WebSource', 'TeradataSource': 'TeradataSource', 'OracleSource': 'OracleSource', 'AzureDataExplorerSource': 'AzureDataExplorerSource', 'AzureMySqlSource': 'AzureMySqlSource', 'HdfsSource': 'HdfsSource', 'FileSystemSource': 'FileSystemSource', 'SqlDWSource': 'SqlDWSource', 'AzureSqlSource': 'AzureSqlSource', 'SqlServerSource': 'SqlServerSource', 'SqlSource': 'SqlSource', 'RestSource': 'RestSource', 'SapTableSource': 'SapTableSource', 'SapOpenHubSource': 'SapOpenHubSource', 'SapHanaSource': 'SapHanaSource', 'SapEccSource': 'SapEccSource', 'SapCloudForCustomerSource': 'SapCloudForCustomerSource', 'SalesforceSource': 'SalesforceSource', 'RelationalSource': 'RelationalSource', 'DynamicsSource': 'DynamicsSource', 'DocumentDbCollectionSource': 'DocumentDbCollectionSource', 'BlobSource': 'BlobSource', 'AzureTableSource': 'AzureTableSource', 'DelimitedTextSource': 'DelimitedTextSource', 'ParquetSource': 'ParquetSource'} + 'type': {'AmazonRedshiftSource': 'AmazonRedshiftSource', 'GoogleAdWordsSource': 'GoogleAdWordsSource', 'OracleServiceCloudSource': 'OracleServiceCloudSource', 'DynamicsAXSource': 'DynamicsAXSource', 'ResponsysSource': 'ResponsysSource', 'SalesforceMarketingCloudSource': 'SalesforceMarketingCloudSource', 'VerticaSource': 'VerticaSource', 'NetezzaSource': 'NetezzaSource', 'ZohoSource': 'ZohoSource', 'XeroSource': 'XeroSource', 'SquareSource': 'SquareSource', 'SparkSource': 'SparkSource', 'ShopifySource': 'ShopifySource', 'ServiceNowSource': 'ServiceNowSource', 'QuickBooksSource': 'QuickBooksSource', 'PrestoSource': 'PrestoSource', 'PhoenixSource': 'PhoenixSource', 'PaypalSource': 'PaypalSource', 'MarketoSource': 'MarketoSource', 'MariaDBSource': 'MariaDBSource', 'MagentoSource': 'MagentoSource', 'JiraSource': 'JiraSource', 'ImpalaSource': 'ImpalaSource', 'HubspotSource': 'HubspotSource', 'HiveSource': 'HiveSource', 'HBaseSource': 'HBaseSource', 'GreenplumSource': 'GreenplumSource', 'GoogleBigQuerySource': 'GoogleBigQuerySource', 'EloquaSource': 'EloquaSource', 'DrillSource': 'DrillSource', 'CouchbaseSource': 'CouchbaseSource', 'ConcurSource': 'ConcurSource', 'AzurePostgreSqlSource': 'AzurePostgreSqlSource', 'AmazonMWSSource': 'AmazonMWSSource', 'HttpSource': 'HttpSource', 'AzureBlobFSSource': 'AzureBlobFSSource', 'AzureDataLakeStoreSource': 'AzureDataLakeStoreSource', 'Office365Source': 'Office365Source', 'CosmosDbMongoDbApiSource': 'CosmosDbMongoDbApiSource', 'MongoDbV2Source': 'MongoDbV2Source', 'MongoDbSource': 'MongoDbSource', 'CassandraSource': 'CassandraSource', 'WebSource': 'WebSource', 'OracleSource': 'OracleSource', 'AzureDataExplorerSource': 'AzureDataExplorerSource', 'AzureMySqlSource': 'AzureMySqlSource', 'HdfsSource': 'HdfsSource', 'FileSystemSource': 'FileSystemSource', 'SqlDWSource': 'SqlDWSource', 'AzureSqlSource': 'AzureSqlSource', 'SqlServerSource': 'SqlServerSource', 'SqlSource': 'SqlSource', 'RestSource': 'RestSource', 'SapTableSource': 'SapTableSource', 'SapOpenHubSource': 'SapOpenHubSource', 'SapHanaSource': 'SapHanaSource', 'SapEccSource': 'SapEccSource', 'SapCloudForCustomerSource': 'SapCloudForCustomerSource', 'SalesforceSource': 'SalesforceSource', 'RelationalSource': 'RelationalSource', 'DynamicsSource': 'DynamicsSource', 'DocumentDbCollectionSource': 'DocumentDbCollectionSource', 'BlobSource': 'BlobSource', 'AzureTableSource': 'AzureTableSource', 'DelimitedTextSource': 'DelimitedTextSource', 'ParquetSource': 'ParquetSource'} } def __init__(self, **kwargs): diff --git a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/copy_source_py3.py b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/copy_source_py3.py index 9f301bc3211e..5a0b7d807b7f 100644 --- a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/copy_source_py3.py +++ b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/copy_source_py3.py @@ -26,10 +26,10 @@ class CopySource(Model): EloquaSource, DrillSource, CouchbaseSource, ConcurSource, AzurePostgreSqlSource, AmazonMWSSource, HttpSource, AzureBlobFSSource, AzureDataLakeStoreSource, Office365Source, CosmosDbMongoDbApiSource, - MongoDbV2Source, MongoDbSource, CassandraSource, WebSource, TeradataSource, - OracleSource, AzureDataExplorerSource, AzureMySqlSource, HdfsSource, - FileSystemSource, SqlDWSource, AzureSqlSource, SqlServerSource, SqlSource, - RestSource, SapTableSource, SapOpenHubSource, SapHanaSource, SapEccSource, + MongoDbV2Source, MongoDbSource, CassandraSource, WebSource, OracleSource, + AzureDataExplorerSource, AzureMySqlSource, HdfsSource, FileSystemSource, + SqlDWSource, AzureSqlSource, SqlServerSource, SqlSource, RestSource, + SapTableSource, SapOpenHubSource, SapHanaSource, SapEccSource, SapCloudForCustomerSource, SalesforceSource, RelationalSource, DynamicsSource, DocumentDbCollectionSource, BlobSource, AzureTableSource, DelimitedTextSource, ParquetSource @@ -67,7 +67,7 @@ class CopySource(Model): } _subtype_map = { - 'type': {'AmazonRedshiftSource': 'AmazonRedshiftSource', 'GoogleAdWordsSource': 'GoogleAdWordsSource', 'OracleServiceCloudSource': 'OracleServiceCloudSource', 'DynamicsAXSource': 'DynamicsAXSource', 'ResponsysSource': 'ResponsysSource', 'SalesforceMarketingCloudSource': 'SalesforceMarketingCloudSource', 'VerticaSource': 'VerticaSource', 'NetezzaSource': 'NetezzaSource', 'ZohoSource': 'ZohoSource', 'XeroSource': 'XeroSource', 'SquareSource': 'SquareSource', 'SparkSource': 'SparkSource', 'ShopifySource': 'ShopifySource', 'ServiceNowSource': 'ServiceNowSource', 'QuickBooksSource': 'QuickBooksSource', 'PrestoSource': 'PrestoSource', 'PhoenixSource': 'PhoenixSource', 'PaypalSource': 'PaypalSource', 'MarketoSource': 'MarketoSource', 'MariaDBSource': 'MariaDBSource', 'MagentoSource': 'MagentoSource', 'JiraSource': 'JiraSource', 'ImpalaSource': 'ImpalaSource', 'HubspotSource': 'HubspotSource', 'HiveSource': 'HiveSource', 'HBaseSource': 'HBaseSource', 'GreenplumSource': 'GreenplumSource', 'GoogleBigQuerySource': 'GoogleBigQuerySource', 'EloquaSource': 'EloquaSource', 'DrillSource': 'DrillSource', 'CouchbaseSource': 'CouchbaseSource', 'ConcurSource': 'ConcurSource', 'AzurePostgreSqlSource': 'AzurePostgreSqlSource', 'AmazonMWSSource': 'AmazonMWSSource', 'HttpSource': 'HttpSource', 'AzureBlobFSSource': 'AzureBlobFSSource', 'AzureDataLakeStoreSource': 'AzureDataLakeStoreSource', 'Office365Source': 'Office365Source', 'CosmosDbMongoDbApiSource': 'CosmosDbMongoDbApiSource', 'MongoDbV2Source': 'MongoDbV2Source', 'MongoDbSource': 'MongoDbSource', 'CassandraSource': 'CassandraSource', 'WebSource': 'WebSource', 'TeradataSource': 'TeradataSource', 'OracleSource': 'OracleSource', 'AzureDataExplorerSource': 'AzureDataExplorerSource', 'AzureMySqlSource': 'AzureMySqlSource', 'HdfsSource': 'HdfsSource', 'FileSystemSource': 'FileSystemSource', 'SqlDWSource': 'SqlDWSource', 'AzureSqlSource': 'AzureSqlSource', 'SqlServerSource': 'SqlServerSource', 'SqlSource': 'SqlSource', 'RestSource': 'RestSource', 'SapTableSource': 'SapTableSource', 'SapOpenHubSource': 'SapOpenHubSource', 'SapHanaSource': 'SapHanaSource', 'SapEccSource': 'SapEccSource', 'SapCloudForCustomerSource': 'SapCloudForCustomerSource', 'SalesforceSource': 'SalesforceSource', 'RelationalSource': 'RelationalSource', 'DynamicsSource': 'DynamicsSource', 'DocumentDbCollectionSource': 'DocumentDbCollectionSource', 'BlobSource': 'BlobSource', 'AzureTableSource': 'AzureTableSource', 'DelimitedTextSource': 'DelimitedTextSource', 'ParquetSource': 'ParquetSource'} + 'type': {'AmazonRedshiftSource': 'AmazonRedshiftSource', 'GoogleAdWordsSource': 'GoogleAdWordsSource', 'OracleServiceCloudSource': 'OracleServiceCloudSource', 'DynamicsAXSource': 'DynamicsAXSource', 'ResponsysSource': 'ResponsysSource', 'SalesforceMarketingCloudSource': 'SalesforceMarketingCloudSource', 'VerticaSource': 'VerticaSource', 'NetezzaSource': 'NetezzaSource', 'ZohoSource': 'ZohoSource', 'XeroSource': 'XeroSource', 'SquareSource': 'SquareSource', 'SparkSource': 'SparkSource', 'ShopifySource': 'ShopifySource', 'ServiceNowSource': 'ServiceNowSource', 'QuickBooksSource': 'QuickBooksSource', 'PrestoSource': 'PrestoSource', 'PhoenixSource': 'PhoenixSource', 'PaypalSource': 'PaypalSource', 'MarketoSource': 'MarketoSource', 'MariaDBSource': 'MariaDBSource', 'MagentoSource': 'MagentoSource', 'JiraSource': 'JiraSource', 'ImpalaSource': 'ImpalaSource', 'HubspotSource': 'HubspotSource', 'HiveSource': 'HiveSource', 'HBaseSource': 'HBaseSource', 'GreenplumSource': 'GreenplumSource', 'GoogleBigQuerySource': 'GoogleBigQuerySource', 'EloquaSource': 'EloquaSource', 'DrillSource': 'DrillSource', 'CouchbaseSource': 'CouchbaseSource', 'ConcurSource': 'ConcurSource', 'AzurePostgreSqlSource': 'AzurePostgreSqlSource', 'AmazonMWSSource': 'AmazonMWSSource', 'HttpSource': 'HttpSource', 'AzureBlobFSSource': 'AzureBlobFSSource', 'AzureDataLakeStoreSource': 'AzureDataLakeStoreSource', 'Office365Source': 'Office365Source', 'CosmosDbMongoDbApiSource': 'CosmosDbMongoDbApiSource', 'MongoDbV2Source': 'MongoDbV2Source', 'MongoDbSource': 'MongoDbSource', 'CassandraSource': 'CassandraSource', 'WebSource': 'WebSource', 'OracleSource': 'OracleSource', 'AzureDataExplorerSource': 'AzureDataExplorerSource', 'AzureMySqlSource': 'AzureMySqlSource', 'HdfsSource': 'HdfsSource', 'FileSystemSource': 'FileSystemSource', 'SqlDWSource': 'SqlDWSource', 'AzureSqlSource': 'AzureSqlSource', 'SqlServerSource': 'SqlServerSource', 'SqlSource': 'SqlSource', 'RestSource': 'RestSource', 'SapTableSource': 'SapTableSource', 'SapOpenHubSource': 'SapOpenHubSource', 'SapHanaSource': 'SapHanaSource', 'SapEccSource': 'SapEccSource', 'SapCloudForCustomerSource': 'SapCloudForCustomerSource', 'SalesforceSource': 'SalesforceSource', 'RelationalSource': 'RelationalSource', 'DynamicsSource': 'DynamicsSource', 'DocumentDbCollectionSource': 'DocumentDbCollectionSource', 'BlobSource': 'BlobSource', 'AzureTableSource': 'AzureTableSource', 'DelimitedTextSource': 'DelimitedTextSource', 'ParquetSource': 'ParquetSource'} } def __init__(self, *, additional_properties=None, source_retry_count=None, source_retry_wait=None, max_concurrent_connections=None, **kwargs) -> None: diff --git a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/data_factory_management_client_enums.py b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/data_factory_management_client_enums.py index ded527b2602a..f6b7388285d6 100644 --- a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/data_factory_management_client_enums.py +++ b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/data_factory_management_client_enums.py @@ -340,6 +340,20 @@ class WebActivityMethod(str, Enum): delete = "DELETE" +class CassandraSourceReadConsistencyLevels(str, Enum): + + all = "ALL" + each_quorum = "EACH_QUORUM" + quorum = "QUORUM" + local_quorum = "LOCAL_QUORUM" + one = "ONE" + two = "TWO" + three = "THREE" + local_one = "LOCAL_ONE" + serial = "SERIAL" + local_serial = "LOCAL_SERIAL" + + class StoredProcedureParameterType(str, Enum): string = "String" @@ -351,6 +365,12 @@ class StoredProcedureParameterType(str, Enum): date_enum = "Date" +class SalesforceSourceReadBehavior(str, Enum): + + query = "Query" + query_all = "QueryAll" + + class HDInsightActivityDebugInfoOption(str, Enum): none = "None" @@ -358,12 +378,30 @@ class HDInsightActivityDebugInfoOption(str, Enum): failure = "Failure" +class SalesforceSinkWriteBehavior(str, Enum): + + insert = "Insert" + upsert = "Upsert" + + +class AzureSearchIndexWriteBehaviorType(str, Enum): + + merge = "Merge" + upload = "Upload" + + class PolybaseSettingsRejectType(str, Enum): value = "value" percentage = "percentage" +class SapCloudForCustomerSinkWriteBehavior(str, Enum): + + insert = "Insert" + update = "Update" + + class WebHookActivityMethod(str, Enum): post = "POST" diff --git a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset.py b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset.py index c95b5c3da543..76d8375b3da9 100644 --- a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset.py +++ b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset.py @@ -34,7 +34,7 @@ class Dataset(Model): SqlServerTableDataset, SapOpenHubTableDataset, SapHanaTableDataset, SapEccResourceDataset, SapCloudForCustomerResourceDataset, SalesforceObjectDataset, RelationalTableDataset, AzureMySqlTableDataset, - TeradataTableDataset, OracleTableDataset, ODataResourceDataset, + OracleTableDataset, ODataResourceDataset, CosmosDbMongoDbApiCollectionDataset, MongoDbV2CollectionDataset, MongoDbCollectionDataset, FileShareDataset, Office365Dataset, AzureBlobFSDataset, AzureDataLakeStoreDataset, DynamicsEntityDataset, @@ -90,7 +90,7 @@ class Dataset(Model): } _subtype_map = { - 'type': {'GoogleAdWordsObject': 'GoogleAdWordsObjectDataset', 'AzureDataExplorerTable': 'AzureDataExplorerTableDataset', 'OracleServiceCloudObject': 'OracleServiceCloudObjectDataset', 'DynamicsAXResource': 'DynamicsAXResourceDataset', 'ResponsysObject': 'ResponsysObjectDataset', 'SalesforceMarketingCloudObject': 'SalesforceMarketingCloudObjectDataset', 'VerticaTable': 'VerticaTableDataset', 'NetezzaTable': 'NetezzaTableDataset', 'ZohoObject': 'ZohoObjectDataset', 'XeroObject': 'XeroObjectDataset', 'SquareObject': 'SquareObjectDataset', 'SparkObject': 'SparkObjectDataset', 'ShopifyObject': 'ShopifyObjectDataset', 'ServiceNowObject': 'ServiceNowObjectDataset', 'QuickBooksObject': 'QuickBooksObjectDataset', 'PrestoObject': 'PrestoObjectDataset', 'PhoenixObject': 'PhoenixObjectDataset', 'PaypalObject': 'PaypalObjectDataset', 'MarketoObject': 'MarketoObjectDataset', 'MariaDBTable': 'MariaDBTableDataset', 'MagentoObject': 'MagentoObjectDataset', 'JiraObject': 'JiraObjectDataset', 'ImpalaObject': 'ImpalaObjectDataset', 'HubspotObject': 'HubspotObjectDataset', 'HiveObject': 'HiveObjectDataset', 'HBaseObject': 'HBaseObjectDataset', 'GreenplumTable': 'GreenplumTableDataset', 'GoogleBigQueryObject': 'GoogleBigQueryObjectDataset', 'EloquaObject': 'EloquaObjectDataset', 'DrillTable': 'DrillTableDataset', 'CouchbaseTable': 'CouchbaseTableDataset', 'ConcurObject': 'ConcurObjectDataset', 'AzurePostgreSqlTable': 'AzurePostgreSqlTableDataset', 'AmazonMWSObject': 'AmazonMWSObjectDataset', 'HttpFile': 'HttpDataset', 'AzureSearchIndex': 'AzureSearchIndexDataset', 'WebTable': 'WebTableDataset', 'SapTableResource': 'SapTableResourceDataset', 'RestResource': 'RestResourceDataset', 'SqlServerTable': 'SqlServerTableDataset', 'SapOpenHubTable': 'SapOpenHubTableDataset', 'SapHanaTable': 'SapHanaTableDataset', 'SapEccResource': 'SapEccResourceDataset', 'SapCloudForCustomerResource': 'SapCloudForCustomerResourceDataset', 'SalesforceObject': 'SalesforceObjectDataset', 'RelationalTable': 'RelationalTableDataset', 'AzureMySqlTable': 'AzureMySqlTableDataset', 'TeradataTable': 'TeradataTableDataset', 'OracleTable': 'OracleTableDataset', 'ODataResource': 'ODataResourceDataset', 'CosmosDbMongoDbApiCollection': 'CosmosDbMongoDbApiCollectionDataset', 'MongoDbV2Collection': 'MongoDbV2CollectionDataset', 'MongoDbCollection': 'MongoDbCollectionDataset', 'FileShare': 'FileShareDataset', 'Office365Table': 'Office365Dataset', 'AzureBlobFSFile': 'AzureBlobFSDataset', 'AzureDataLakeStoreFile': 'AzureDataLakeStoreDataset', 'DynamicsEntity': 'DynamicsEntityDataset', 'DocumentDbCollection': 'DocumentDbCollectionDataset', 'CustomDataset': 'CustomDataset', 'CassandraTable': 'CassandraTableDataset', 'AzureSqlDWTable': 'AzureSqlDWTableDataset', 'AzureSqlTable': 'AzureSqlTableDataset', 'AzureTable': 'AzureTableDataset', 'AzureBlob': 'AzureBlobDataset', 'DelimitedText': 'DelimitedTextDataset', 'Parquet': 'ParquetDataset', 'AmazonS3Object': 'AmazonS3Dataset'} + 'type': {'GoogleAdWordsObject': 'GoogleAdWordsObjectDataset', 'AzureDataExplorerTable': 'AzureDataExplorerTableDataset', 'OracleServiceCloudObject': 'OracleServiceCloudObjectDataset', 'DynamicsAXResource': 'DynamicsAXResourceDataset', 'ResponsysObject': 'ResponsysObjectDataset', 'SalesforceMarketingCloudObject': 'SalesforceMarketingCloudObjectDataset', 'VerticaTable': 'VerticaTableDataset', 'NetezzaTable': 'NetezzaTableDataset', 'ZohoObject': 'ZohoObjectDataset', 'XeroObject': 'XeroObjectDataset', 'SquareObject': 'SquareObjectDataset', 'SparkObject': 'SparkObjectDataset', 'ShopifyObject': 'ShopifyObjectDataset', 'ServiceNowObject': 'ServiceNowObjectDataset', 'QuickBooksObject': 'QuickBooksObjectDataset', 'PrestoObject': 'PrestoObjectDataset', 'PhoenixObject': 'PhoenixObjectDataset', 'PaypalObject': 'PaypalObjectDataset', 'MarketoObject': 'MarketoObjectDataset', 'MariaDBTable': 'MariaDBTableDataset', 'MagentoObject': 'MagentoObjectDataset', 'JiraObject': 'JiraObjectDataset', 'ImpalaObject': 'ImpalaObjectDataset', 'HubspotObject': 'HubspotObjectDataset', 'HiveObject': 'HiveObjectDataset', 'HBaseObject': 'HBaseObjectDataset', 'GreenplumTable': 'GreenplumTableDataset', 'GoogleBigQueryObject': 'GoogleBigQueryObjectDataset', 'EloquaObject': 'EloquaObjectDataset', 'DrillTable': 'DrillTableDataset', 'CouchbaseTable': 'CouchbaseTableDataset', 'ConcurObject': 'ConcurObjectDataset', 'AzurePostgreSqlTable': 'AzurePostgreSqlTableDataset', 'AmazonMWSObject': 'AmazonMWSObjectDataset', 'HttpFile': 'HttpDataset', 'AzureSearchIndex': 'AzureSearchIndexDataset', 'WebTable': 'WebTableDataset', 'SapTableResource': 'SapTableResourceDataset', 'RestResource': 'RestResourceDataset', 'SqlServerTable': 'SqlServerTableDataset', 'SapOpenHubTable': 'SapOpenHubTableDataset', 'SapHanaTable': 'SapHanaTableDataset', 'SapEccResource': 'SapEccResourceDataset', 'SapCloudForCustomerResource': 'SapCloudForCustomerResourceDataset', 'SalesforceObject': 'SalesforceObjectDataset', 'RelationalTable': 'RelationalTableDataset', 'AzureMySqlTable': 'AzureMySqlTableDataset', 'OracleTable': 'OracleTableDataset', 'ODataResource': 'ODataResourceDataset', 'CosmosDbMongoDbApiCollection': 'CosmosDbMongoDbApiCollectionDataset', 'MongoDbV2Collection': 'MongoDbV2CollectionDataset', 'MongoDbCollection': 'MongoDbCollectionDataset', 'FileShare': 'FileShareDataset', 'Office365Table': 'Office365Dataset', 'AzureBlobFSFile': 'AzureBlobFSDataset', 'AzureDataLakeStoreFile': 'AzureDataLakeStoreDataset', 'DynamicsEntity': 'DynamicsEntityDataset', 'DocumentDbCollection': 'DocumentDbCollectionDataset', 'CustomDataset': 'CustomDataset', 'CassandraTable': 'CassandraTableDataset', 'AzureSqlDWTable': 'AzureSqlDWTableDataset', 'AzureSqlTable': 'AzureSqlTableDataset', 'AzureTable': 'AzureTableDataset', 'AzureBlob': 'AzureBlobDataset', 'DelimitedText': 'DelimitedTextDataset', 'Parquet': 'ParquetDataset', 'AmazonS3Object': 'AmazonS3Dataset'} } def __init__(self, **kwargs): diff --git a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_py3.py b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_py3.py index c0cfdb972550..c793e32f2251 100644 --- a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_py3.py +++ b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_py3.py @@ -34,7 +34,7 @@ class Dataset(Model): SqlServerTableDataset, SapOpenHubTableDataset, SapHanaTableDataset, SapEccResourceDataset, SapCloudForCustomerResourceDataset, SalesforceObjectDataset, RelationalTableDataset, AzureMySqlTableDataset, - TeradataTableDataset, OracleTableDataset, ODataResourceDataset, + OracleTableDataset, ODataResourceDataset, CosmosDbMongoDbApiCollectionDataset, MongoDbV2CollectionDataset, MongoDbCollectionDataset, FileShareDataset, Office365Dataset, AzureBlobFSDataset, AzureDataLakeStoreDataset, DynamicsEntityDataset, @@ -90,7 +90,7 @@ class Dataset(Model): } _subtype_map = { - 'type': {'GoogleAdWordsObject': 'GoogleAdWordsObjectDataset', 'AzureDataExplorerTable': 'AzureDataExplorerTableDataset', 'OracleServiceCloudObject': 'OracleServiceCloudObjectDataset', 'DynamicsAXResource': 'DynamicsAXResourceDataset', 'ResponsysObject': 'ResponsysObjectDataset', 'SalesforceMarketingCloudObject': 'SalesforceMarketingCloudObjectDataset', 'VerticaTable': 'VerticaTableDataset', 'NetezzaTable': 'NetezzaTableDataset', 'ZohoObject': 'ZohoObjectDataset', 'XeroObject': 'XeroObjectDataset', 'SquareObject': 'SquareObjectDataset', 'SparkObject': 'SparkObjectDataset', 'ShopifyObject': 'ShopifyObjectDataset', 'ServiceNowObject': 'ServiceNowObjectDataset', 'QuickBooksObject': 'QuickBooksObjectDataset', 'PrestoObject': 'PrestoObjectDataset', 'PhoenixObject': 'PhoenixObjectDataset', 'PaypalObject': 'PaypalObjectDataset', 'MarketoObject': 'MarketoObjectDataset', 'MariaDBTable': 'MariaDBTableDataset', 'MagentoObject': 'MagentoObjectDataset', 'JiraObject': 'JiraObjectDataset', 'ImpalaObject': 'ImpalaObjectDataset', 'HubspotObject': 'HubspotObjectDataset', 'HiveObject': 'HiveObjectDataset', 'HBaseObject': 'HBaseObjectDataset', 'GreenplumTable': 'GreenplumTableDataset', 'GoogleBigQueryObject': 'GoogleBigQueryObjectDataset', 'EloquaObject': 'EloquaObjectDataset', 'DrillTable': 'DrillTableDataset', 'CouchbaseTable': 'CouchbaseTableDataset', 'ConcurObject': 'ConcurObjectDataset', 'AzurePostgreSqlTable': 'AzurePostgreSqlTableDataset', 'AmazonMWSObject': 'AmazonMWSObjectDataset', 'HttpFile': 'HttpDataset', 'AzureSearchIndex': 'AzureSearchIndexDataset', 'WebTable': 'WebTableDataset', 'SapTableResource': 'SapTableResourceDataset', 'RestResource': 'RestResourceDataset', 'SqlServerTable': 'SqlServerTableDataset', 'SapOpenHubTable': 'SapOpenHubTableDataset', 'SapHanaTable': 'SapHanaTableDataset', 'SapEccResource': 'SapEccResourceDataset', 'SapCloudForCustomerResource': 'SapCloudForCustomerResourceDataset', 'SalesforceObject': 'SalesforceObjectDataset', 'RelationalTable': 'RelationalTableDataset', 'AzureMySqlTable': 'AzureMySqlTableDataset', 'TeradataTable': 'TeradataTableDataset', 'OracleTable': 'OracleTableDataset', 'ODataResource': 'ODataResourceDataset', 'CosmosDbMongoDbApiCollection': 'CosmosDbMongoDbApiCollectionDataset', 'MongoDbV2Collection': 'MongoDbV2CollectionDataset', 'MongoDbCollection': 'MongoDbCollectionDataset', 'FileShare': 'FileShareDataset', 'Office365Table': 'Office365Dataset', 'AzureBlobFSFile': 'AzureBlobFSDataset', 'AzureDataLakeStoreFile': 'AzureDataLakeStoreDataset', 'DynamicsEntity': 'DynamicsEntityDataset', 'DocumentDbCollection': 'DocumentDbCollectionDataset', 'CustomDataset': 'CustomDataset', 'CassandraTable': 'CassandraTableDataset', 'AzureSqlDWTable': 'AzureSqlDWTableDataset', 'AzureSqlTable': 'AzureSqlTableDataset', 'AzureTable': 'AzureTableDataset', 'AzureBlob': 'AzureBlobDataset', 'DelimitedText': 'DelimitedTextDataset', 'Parquet': 'ParquetDataset', 'AmazonS3Object': 'AmazonS3Dataset'} + 'type': {'GoogleAdWordsObject': 'GoogleAdWordsObjectDataset', 'AzureDataExplorerTable': 'AzureDataExplorerTableDataset', 'OracleServiceCloudObject': 'OracleServiceCloudObjectDataset', 'DynamicsAXResource': 'DynamicsAXResourceDataset', 'ResponsysObject': 'ResponsysObjectDataset', 'SalesforceMarketingCloudObject': 'SalesforceMarketingCloudObjectDataset', 'VerticaTable': 'VerticaTableDataset', 'NetezzaTable': 'NetezzaTableDataset', 'ZohoObject': 'ZohoObjectDataset', 'XeroObject': 'XeroObjectDataset', 'SquareObject': 'SquareObjectDataset', 'SparkObject': 'SparkObjectDataset', 'ShopifyObject': 'ShopifyObjectDataset', 'ServiceNowObject': 'ServiceNowObjectDataset', 'QuickBooksObject': 'QuickBooksObjectDataset', 'PrestoObject': 'PrestoObjectDataset', 'PhoenixObject': 'PhoenixObjectDataset', 'PaypalObject': 'PaypalObjectDataset', 'MarketoObject': 'MarketoObjectDataset', 'MariaDBTable': 'MariaDBTableDataset', 'MagentoObject': 'MagentoObjectDataset', 'JiraObject': 'JiraObjectDataset', 'ImpalaObject': 'ImpalaObjectDataset', 'HubspotObject': 'HubspotObjectDataset', 'HiveObject': 'HiveObjectDataset', 'HBaseObject': 'HBaseObjectDataset', 'GreenplumTable': 'GreenplumTableDataset', 'GoogleBigQueryObject': 'GoogleBigQueryObjectDataset', 'EloquaObject': 'EloquaObjectDataset', 'DrillTable': 'DrillTableDataset', 'CouchbaseTable': 'CouchbaseTableDataset', 'ConcurObject': 'ConcurObjectDataset', 'AzurePostgreSqlTable': 'AzurePostgreSqlTableDataset', 'AmazonMWSObject': 'AmazonMWSObjectDataset', 'HttpFile': 'HttpDataset', 'AzureSearchIndex': 'AzureSearchIndexDataset', 'WebTable': 'WebTableDataset', 'SapTableResource': 'SapTableResourceDataset', 'RestResource': 'RestResourceDataset', 'SqlServerTable': 'SqlServerTableDataset', 'SapOpenHubTable': 'SapOpenHubTableDataset', 'SapHanaTable': 'SapHanaTableDataset', 'SapEccResource': 'SapEccResourceDataset', 'SapCloudForCustomerResource': 'SapCloudForCustomerResourceDataset', 'SalesforceObject': 'SalesforceObjectDataset', 'RelationalTable': 'RelationalTableDataset', 'AzureMySqlTable': 'AzureMySqlTableDataset', 'OracleTable': 'OracleTableDataset', 'ODataResource': 'ODataResourceDataset', 'CosmosDbMongoDbApiCollection': 'CosmosDbMongoDbApiCollectionDataset', 'MongoDbV2Collection': 'MongoDbV2CollectionDataset', 'MongoDbCollection': 'MongoDbCollectionDataset', 'FileShare': 'FileShareDataset', 'Office365Table': 'Office365Dataset', 'AzureBlobFSFile': 'AzureBlobFSDataset', 'AzureDataLakeStoreFile': 'AzureDataLakeStoreDataset', 'DynamicsEntity': 'DynamicsEntityDataset', 'DocumentDbCollection': 'DocumentDbCollectionDataset', 'CustomDataset': 'CustomDataset', 'CassandraTable': 'CassandraTableDataset', 'AzureSqlDWTable': 'AzureSqlDWTableDataset', 'AzureSqlTable': 'AzureSqlTableDataset', 'AzureTable': 'AzureTableDataset', 'AzureBlob': 'AzureBlobDataset', 'DelimitedText': 'DelimitedTextDataset', 'Parquet': 'ParquetDataset', 'AmazonS3Object': 'AmazonS3Dataset'} } def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, schema=None, parameters=None, annotations=None, folder=None, **kwargs) -> None: diff --git a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dynamics_sink.py b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dynamics_sink.py index 5afce6ced25b..45bac7b52064 100644 --- a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dynamics_sink.py +++ b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dynamics_sink.py @@ -44,7 +44,8 @@ class DynamicsSink(CopySink): :param type: Required. Constant filled by server. :type type: str :ivar write_behavior: Required. The write behavior for the operation. - :vartype write_behavior: object + Default value: "Upsert" . + :vartype write_behavior: str :param ignore_null_values: The flag indicating whether ignore null values from input dataset (except key fields) during write operation. Default is false. Type: boolean (or Expression with resultType boolean). @@ -64,11 +65,11 @@ class DynamicsSink(CopySink): 'sink_retry_wait': {'key': 'sinkRetryWait', 'type': 'object'}, 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'type': {'key': 'type', 'type': 'str'}, - 'write_behavior': {'key': 'writeBehavior', 'type': 'object'}, + 'write_behavior': {'key': 'writeBehavior', 'type': 'str'}, 'ignore_null_values': {'key': 'ignoreNullValues', 'type': 'object'}, } - write_behavior = None + write_behavior = "Upsert" def __init__(self, **kwargs): super(DynamicsSink, self).__init__(**kwargs) diff --git a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dynamics_sink_py3.py b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dynamics_sink_py3.py index ffdb08363bfd..5f736f9cf658 100644 --- a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dynamics_sink_py3.py +++ b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dynamics_sink_py3.py @@ -44,7 +44,8 @@ class DynamicsSink(CopySink): :param type: Required. Constant filled by server. :type type: str :ivar write_behavior: Required. The write behavior for the operation. - :vartype write_behavior: object + Default value: "Upsert" . + :vartype write_behavior: str :param ignore_null_values: The flag indicating whether ignore null values from input dataset (except key fields) during write operation. Default is false. Type: boolean (or Expression with resultType boolean). @@ -64,11 +65,11 @@ class DynamicsSink(CopySink): 'sink_retry_wait': {'key': 'sinkRetryWait', 'type': 'object'}, 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'type': {'key': 'type', 'type': 'str'}, - 'write_behavior': {'key': 'writeBehavior', 'type': 'object'}, + 'write_behavior': {'key': 'writeBehavior', 'type': 'str'}, 'ignore_null_values': {'key': 'ignoreNullValues', 'type': 'object'}, } - write_behavior = None + write_behavior = "Upsert" def __init__(self, *, additional_properties=None, write_batch_size=None, write_batch_timeout=None, sink_retry_count=None, sink_retry_wait=None, max_concurrent_connections=None, ignore_null_values=None, **kwargs) -> None: super(DynamicsSink, self).__init__(additional_properties=additional_properties, write_batch_size=write_batch_size, write_batch_timeout=write_batch_timeout, sink_retry_count=sink_retry_count, sink_retry_wait=sink_retry_wait, max_concurrent_connections=max_concurrent_connections, **kwargs) diff --git a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_partition_settings.py b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_partition_settings.py deleted file mode 100644 index ed970fd7729a..000000000000 --- a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_partition_settings.py +++ /dev/null @@ -1,46 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for -# license information. -# -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is -# regenerated. -# -------------------------------------------------------------------------- - -from msrest.serialization import Model - - -class OraclePartitionSettings(Model): - """The settings that will be leveraged for oracle source partitioning. - - :param partition_names: Names of the physical partitions of oracle table. - :type partition_names: object - :param partition_column_name: The name of the column in integer type that - will be used for proceeding range partitioning. Type: string (or - Expression with resultType string). - :type partition_column_name: object - :param partition_upper_bound: The maximum value of column specified in - partitionColumnName that will be used for proceeding range partitioning. - Type: string (or Expression with resultType string). - :type partition_upper_bound: object - :param partition_lower_bound: The minimum value of column specified in - partitionColumnName that will be used for proceeding range partitioning. - Type: string (or Expression with resultType string). - :type partition_lower_bound: object - """ - - _attribute_map = { - 'partition_names': {'key': 'partitionNames', 'type': 'object'}, - 'partition_column_name': {'key': 'partitionColumnName', 'type': 'object'}, - 'partition_upper_bound': {'key': 'partitionUpperBound', 'type': 'object'}, - 'partition_lower_bound': {'key': 'partitionLowerBound', 'type': 'object'}, - } - - def __init__(self, **kwargs): - super(OraclePartitionSettings, self).__init__(**kwargs) - self.partition_names = kwargs.get('partition_names', None) - self.partition_column_name = kwargs.get('partition_column_name', None) - self.partition_upper_bound = kwargs.get('partition_upper_bound', None) - self.partition_lower_bound = kwargs.get('partition_lower_bound', None) diff --git a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_partition_settings_py3.py b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_partition_settings_py3.py deleted file mode 100644 index c3d00b09ad90..000000000000 --- a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_partition_settings_py3.py +++ /dev/null @@ -1,46 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for -# license information. -# -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is -# regenerated. -# -------------------------------------------------------------------------- - -from msrest.serialization import Model - - -class OraclePartitionSettings(Model): - """The settings that will be leveraged for oracle source partitioning. - - :param partition_names: Names of the physical partitions of oracle table. - :type partition_names: object - :param partition_column_name: The name of the column in integer type that - will be used for proceeding range partitioning. Type: string (or - Expression with resultType string). - :type partition_column_name: object - :param partition_upper_bound: The maximum value of column specified in - partitionColumnName that will be used for proceeding range partitioning. - Type: string (or Expression with resultType string). - :type partition_upper_bound: object - :param partition_lower_bound: The minimum value of column specified in - partitionColumnName that will be used for proceeding range partitioning. - Type: string (or Expression with resultType string). - :type partition_lower_bound: object - """ - - _attribute_map = { - 'partition_names': {'key': 'partitionNames', 'type': 'object'}, - 'partition_column_name': {'key': 'partitionColumnName', 'type': 'object'}, - 'partition_upper_bound': {'key': 'partitionUpperBound', 'type': 'object'}, - 'partition_lower_bound': {'key': 'partitionLowerBound', 'type': 'object'}, - } - - def __init__(self, *, partition_names=None, partition_column_name=None, partition_upper_bound=None, partition_lower_bound=None, **kwargs) -> None: - super(OraclePartitionSettings, self).__init__(**kwargs) - self.partition_names = partition_names - self.partition_column_name = partition_column_name - self.partition_upper_bound = partition_upper_bound - self.partition_lower_bound = partition_lower_bound diff --git a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_source.py b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_source.py index 84ad79ed19c7..12b3aa31353f 100644 --- a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_source.py +++ b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_source.py @@ -40,13 +40,6 @@ class OracleSource(CopySource): resultType string), pattern: ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type query_timeout: object - :param partition_option: The partition mechanism that will be used for - oracle read in parallel. - :type partition_option: object - :param partition_settings: The settings that will be leveraged for oracle - source partitioning. - :type partition_settings: - ~azure.mgmt.datafactory.models.OraclePartitionSettings """ _validation = { @@ -61,14 +54,10 @@ class OracleSource(CopySource): 'type': {'key': 'type', 'type': 'str'}, 'oracle_reader_query': {'key': 'oracleReaderQuery', 'type': 'object'}, 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'partition_option': {'key': 'partitionOption', 'type': 'object'}, - 'partition_settings': {'key': 'partitionSettings', 'type': 'OraclePartitionSettings'}, } def __init__(self, **kwargs): super(OracleSource, self).__init__(**kwargs) self.oracle_reader_query = kwargs.get('oracle_reader_query', None) self.query_timeout = kwargs.get('query_timeout', None) - self.partition_option = kwargs.get('partition_option', None) - self.partition_settings = kwargs.get('partition_settings', None) self.type = 'OracleSource' diff --git a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_source_py3.py b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_source_py3.py index dfcbd2e0330d..43afe27fda2f 100644 --- a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_source_py3.py +++ b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/oracle_source_py3.py @@ -40,13 +40,6 @@ class OracleSource(CopySource): resultType string), pattern: ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type query_timeout: object - :param partition_option: The partition mechanism that will be used for - oracle read in parallel. - :type partition_option: object - :param partition_settings: The settings that will be leveraged for oracle - source partitioning. - :type partition_settings: - ~azure.mgmt.datafactory.models.OraclePartitionSettings """ _validation = { @@ -61,14 +54,10 @@ class OracleSource(CopySource): 'type': {'key': 'type', 'type': 'str'}, 'oracle_reader_query': {'key': 'oracleReaderQuery', 'type': 'object'}, 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'partition_option': {'key': 'partitionOption', 'type': 'object'}, - 'partition_settings': {'key': 'partitionSettings', 'type': 'OraclePartitionSettings'}, } - def __init__(self, *, additional_properties=None, source_retry_count=None, source_retry_wait=None, max_concurrent_connections=None, oracle_reader_query=None, query_timeout=None, partition_option=None, partition_settings=None, **kwargs) -> None: + def __init__(self, *, additional_properties=None, source_retry_count=None, source_retry_wait=None, max_concurrent_connections=None, oracle_reader_query=None, query_timeout=None, **kwargs) -> None: super(OracleSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, max_concurrent_connections=max_concurrent_connections, **kwargs) self.oracle_reader_query = oracle_reader_query self.query_timeout = query_timeout - self.partition_option = partition_option - self.partition_settings = partition_settings self.type = 'OracleSource' diff --git a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_sink.py b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_sink.py index 4d1a93c08915..9a1291bd4bfe 100644 --- a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_sink.py +++ b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_sink.py @@ -41,8 +41,9 @@ class SalesforceSink(CopySink): :param type: Required. Constant filled by server. :type type: str :param write_behavior: The write behavior for the operation. Default is - Insert. - :type write_behavior: object + Insert. Possible values include: 'Insert', 'Upsert' + :type write_behavior: str or + ~azure.mgmt.datafactory.models.SalesforceSinkWriteBehavior :param external_id_field_name: The name of the external ID field for upsert operation. Default value is 'Id' column. Type: string (or Expression with resultType string). @@ -70,7 +71,7 @@ class SalesforceSink(CopySink): 'sink_retry_wait': {'key': 'sinkRetryWait', 'type': 'object'}, 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'type': {'key': 'type', 'type': 'str'}, - 'write_behavior': {'key': 'writeBehavior', 'type': 'object'}, + 'write_behavior': {'key': 'writeBehavior', 'type': 'str'}, 'external_id_field_name': {'key': 'externalIdFieldName', 'type': 'object'}, 'ignore_null_values': {'key': 'ignoreNullValues', 'type': 'object'}, } diff --git a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_sink_py3.py b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_sink_py3.py index ed7591fbb59b..54a56618d01e 100644 --- a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_sink_py3.py +++ b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_sink_py3.py @@ -41,8 +41,9 @@ class SalesforceSink(CopySink): :param type: Required. Constant filled by server. :type type: str :param write_behavior: The write behavior for the operation. Default is - Insert. - :type write_behavior: object + Insert. Possible values include: 'Insert', 'Upsert' + :type write_behavior: str or + ~azure.mgmt.datafactory.models.SalesforceSinkWriteBehavior :param external_id_field_name: The name of the external ID field for upsert operation. Default value is 'Id' column. Type: string (or Expression with resultType string). @@ -70,7 +71,7 @@ class SalesforceSink(CopySink): 'sink_retry_wait': {'key': 'sinkRetryWait', 'type': 'object'}, 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'type': {'key': 'type', 'type': 'str'}, - 'write_behavior': {'key': 'writeBehavior', 'type': 'object'}, + 'write_behavior': {'key': 'writeBehavior', 'type': 'str'}, 'external_id_field_name': {'key': 'externalIdFieldName', 'type': 'object'}, 'ignore_null_values': {'key': 'ignoreNullValues', 'type': 'object'}, } diff --git a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_source.py b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_source.py index 57a10411f487..4f2590c3ab9d 100644 --- a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_source.py +++ b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_source.py @@ -37,8 +37,9 @@ class SalesforceSource(CopySource): string). :type query: object :param read_behavior: The read behavior for the operation. Default is - Query. - :type read_behavior: object + Query. Possible values include: 'Query', 'QueryAll' + :type read_behavior: str or + ~azure.mgmt.datafactory.models.SalesforceSourceReadBehavior """ _validation = { @@ -52,7 +53,7 @@ class SalesforceSource(CopySource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'type': {'key': 'type', 'type': 'str'}, 'query': {'key': 'query', 'type': 'object'}, - 'read_behavior': {'key': 'readBehavior', 'type': 'object'}, + 'read_behavior': {'key': 'readBehavior', 'type': 'str'}, } def __init__(self, **kwargs): diff --git a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_source_py3.py b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_source_py3.py index 08e6776f5f98..4441e92eaff3 100644 --- a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_source_py3.py +++ b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_source_py3.py @@ -37,8 +37,9 @@ class SalesforceSource(CopySource): string). :type query: object :param read_behavior: The read behavior for the operation. Default is - Query. - :type read_behavior: object + Query. Possible values include: 'Query', 'QueryAll' + :type read_behavior: str or + ~azure.mgmt.datafactory.models.SalesforceSourceReadBehavior """ _validation = { @@ -52,7 +53,7 @@ class SalesforceSource(CopySource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'type': {'key': 'type', 'type': 'str'}, 'query': {'key': 'query', 'type': 'object'}, - 'read_behavior': {'key': 'readBehavior', 'type': 'object'}, + 'read_behavior': {'key': 'readBehavior', 'type': 'str'}, } def __init__(self, *, additional_properties=None, source_retry_count=None, source_retry_wait=None, max_concurrent_connections=None, query=None, read_behavior=None, **kwargs) -> None: diff --git a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_cloud_for_customer_sink.py b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_cloud_for_customer_sink.py index ae99093f277e..e5a37858abb5 100644 --- a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_cloud_for_customer_sink.py +++ b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_cloud_for_customer_sink.py @@ -41,8 +41,9 @@ class SapCloudForCustomerSink(CopySink): :param type: Required. Constant filled by server. :type type: str :param write_behavior: The write behavior for the operation. Default is - 'Insert'. - :type write_behavior: object + 'Insert'. Possible values include: 'Insert', 'Update' + :type write_behavior: str or + ~azure.mgmt.datafactory.models.SapCloudForCustomerSinkWriteBehavior """ _validation = { @@ -57,7 +58,7 @@ class SapCloudForCustomerSink(CopySink): 'sink_retry_wait': {'key': 'sinkRetryWait', 'type': 'object'}, 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'type': {'key': 'type', 'type': 'str'}, - 'write_behavior': {'key': 'writeBehavior', 'type': 'object'}, + 'write_behavior': {'key': 'writeBehavior', 'type': 'str'}, } def __init__(self, **kwargs): diff --git a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_cloud_for_customer_sink_py3.py b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_cloud_for_customer_sink_py3.py index bdbc2cefcbd1..29f01fdd6891 100644 --- a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_cloud_for_customer_sink_py3.py +++ b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_cloud_for_customer_sink_py3.py @@ -41,8 +41,9 @@ class SapCloudForCustomerSink(CopySink): :param type: Required. Constant filled by server. :type type: str :param write_behavior: The write behavior for the operation. Default is - 'Insert'. - :type write_behavior: object + 'Insert'. Possible values include: 'Insert', 'Update' + :type write_behavior: str or + ~azure.mgmt.datafactory.models.SapCloudForCustomerSinkWriteBehavior """ _validation = { @@ -57,7 +58,7 @@ class SapCloudForCustomerSink(CopySink): 'sink_retry_wait': {'key': 'sinkRetryWait', 'type': 'object'}, 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'type': {'key': 'type', 'type': 'str'}, - 'write_behavior': {'key': 'writeBehavior', 'type': 'object'}, + 'write_behavior': {'key': 'writeBehavior', 'type': 'str'}, } def __init__(self, *, additional_properties=None, write_batch_size=None, write_batch_timeout=None, sink_retry_count=None, sink_retry_wait=None, max_concurrent_connections=None, write_behavior=None, **kwargs) -> None: diff --git a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_table_partition_settings.py b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_table_partition_settings.py deleted file mode 100644 index b688fe16683b..000000000000 --- a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_table_partition_settings.py +++ /dev/null @@ -1,47 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for -# license information. -# -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is -# regenerated. -# -------------------------------------------------------------------------- - -from msrest.serialization import Model - - -class SapTablePartitionSettings(Model): - """The settings that will be leveraged for SAP table source partitioning. - - :param partition_column_name: The name of the column that will be used for - proceeding range partitioning. Type: string (or Expression with resultType - string). - :type partition_column_name: object - :param partition_upper_bound: The maximum value of column specified in - partitionColumnName that will be used for proceeding range partitioning. - Type: string (or Expression with resultType string). - :type partition_upper_bound: object - :param partition_lower_bound: The minimum value of column specified in - partitionColumnName that will be used for proceeding range partitioning. - Type: string (or Expression with resultType string). - :type partition_lower_bound: object - :param max_partitions_number: The maximum value of partitions the table - will be split into. Type: integer (or Expression with resultType string). - :type max_partitions_number: object - """ - - _attribute_map = { - 'partition_column_name': {'key': 'partitionColumnName', 'type': 'object'}, - 'partition_upper_bound': {'key': 'partitionUpperBound', 'type': 'object'}, - 'partition_lower_bound': {'key': 'partitionLowerBound', 'type': 'object'}, - 'max_partitions_number': {'key': 'maxPartitionsNumber', 'type': 'object'}, - } - - def __init__(self, **kwargs): - super(SapTablePartitionSettings, self).__init__(**kwargs) - self.partition_column_name = kwargs.get('partition_column_name', None) - self.partition_upper_bound = kwargs.get('partition_upper_bound', None) - self.partition_lower_bound = kwargs.get('partition_lower_bound', None) - self.max_partitions_number = kwargs.get('max_partitions_number', None) diff --git a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_table_partition_settings_py3.py b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_table_partition_settings_py3.py deleted file mode 100644 index 37bdf610ab35..000000000000 --- a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_table_partition_settings_py3.py +++ /dev/null @@ -1,47 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for -# license information. -# -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is -# regenerated. -# -------------------------------------------------------------------------- - -from msrest.serialization import Model - - -class SapTablePartitionSettings(Model): - """The settings that will be leveraged for SAP table source partitioning. - - :param partition_column_name: The name of the column that will be used for - proceeding range partitioning. Type: string (or Expression with resultType - string). - :type partition_column_name: object - :param partition_upper_bound: The maximum value of column specified in - partitionColumnName that will be used for proceeding range partitioning. - Type: string (or Expression with resultType string). - :type partition_upper_bound: object - :param partition_lower_bound: The minimum value of column specified in - partitionColumnName that will be used for proceeding range partitioning. - Type: string (or Expression with resultType string). - :type partition_lower_bound: object - :param max_partitions_number: The maximum value of partitions the table - will be split into. Type: integer (or Expression with resultType string). - :type max_partitions_number: object - """ - - _attribute_map = { - 'partition_column_name': {'key': 'partitionColumnName', 'type': 'object'}, - 'partition_upper_bound': {'key': 'partitionUpperBound', 'type': 'object'}, - 'partition_lower_bound': {'key': 'partitionLowerBound', 'type': 'object'}, - 'max_partitions_number': {'key': 'maxPartitionsNumber', 'type': 'object'}, - } - - def __init__(self, *, partition_column_name=None, partition_upper_bound=None, partition_lower_bound=None, max_partitions_number=None, **kwargs) -> None: - super(SapTablePartitionSettings, self).__init__(**kwargs) - self.partition_column_name = partition_column_name - self.partition_upper_bound = partition_upper_bound - self.partition_lower_bound = partition_lower_bound - self.max_partitions_number = max_partitions_number diff --git a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_table_source.py b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_table_source.py index 1c52db3eb0f8..e1865aa05535 100644 --- a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_table_source.py +++ b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_table_source.py @@ -55,13 +55,6 @@ class SapTableSource(CopySource): function module that will be used to read data from SAP Table. Type: string (or Expression with resultType string). :type custom_rfc_read_table_function_module: object - :param partition_option: The partition mechanism that will be used for SAP - table read in parallel. - :type partition_option: object - :param partition_settings: The settings that will be leveraged for SAP - table source partitioning. - :type partition_settings: - ~azure.mgmt.datafactory.models.SapTablePartitionSettings """ _validation = { @@ -80,8 +73,6 @@ class SapTableSource(CopySource): 'rfc_table_options': {'key': 'rfcTableOptions', 'type': 'object'}, 'batch_size': {'key': 'batchSize', 'type': 'object'}, 'custom_rfc_read_table_function_module': {'key': 'customRfcReadTableFunctionModule', 'type': 'object'}, - 'partition_option': {'key': 'partitionOption', 'type': 'object'}, - 'partition_settings': {'key': 'partitionSettings', 'type': 'SapTablePartitionSettings'}, } def __init__(self, **kwargs): @@ -92,6 +83,4 @@ def __init__(self, **kwargs): self.rfc_table_options = kwargs.get('rfc_table_options', None) self.batch_size = kwargs.get('batch_size', None) self.custom_rfc_read_table_function_module = kwargs.get('custom_rfc_read_table_function_module', None) - self.partition_option = kwargs.get('partition_option', None) - self.partition_settings = kwargs.get('partition_settings', None) self.type = 'SapTableSource' diff --git a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_table_source_py3.py b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_table_source_py3.py index 8e8fbdf12002..73e915a14ed3 100644 --- a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_table_source_py3.py +++ b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_table_source_py3.py @@ -55,13 +55,6 @@ class SapTableSource(CopySource): function module that will be used to read data from SAP Table. Type: string (or Expression with resultType string). :type custom_rfc_read_table_function_module: object - :param partition_option: The partition mechanism that will be used for SAP - table read in parallel. - :type partition_option: object - :param partition_settings: The settings that will be leveraged for SAP - table source partitioning. - :type partition_settings: - ~azure.mgmt.datafactory.models.SapTablePartitionSettings """ _validation = { @@ -80,11 +73,9 @@ class SapTableSource(CopySource): 'rfc_table_options': {'key': 'rfcTableOptions', 'type': 'object'}, 'batch_size': {'key': 'batchSize', 'type': 'object'}, 'custom_rfc_read_table_function_module': {'key': 'customRfcReadTableFunctionModule', 'type': 'object'}, - 'partition_option': {'key': 'partitionOption', 'type': 'object'}, - 'partition_settings': {'key': 'partitionSettings', 'type': 'SapTablePartitionSettings'}, } - def __init__(self, *, additional_properties=None, source_retry_count=None, source_retry_wait=None, max_concurrent_connections=None, row_count=None, row_skips=None, rfc_table_fields=None, rfc_table_options=None, batch_size=None, custom_rfc_read_table_function_module=None, partition_option=None, partition_settings=None, **kwargs) -> None: + def __init__(self, *, additional_properties=None, source_retry_count=None, source_retry_wait=None, max_concurrent_connections=None, row_count=None, row_skips=None, rfc_table_fields=None, rfc_table_options=None, batch_size=None, custom_rfc_read_table_function_module=None, **kwargs) -> None: super(SapTableSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, max_concurrent_connections=max_concurrent_connections, **kwargs) self.row_count = row_count self.row_skips = row_skips @@ -92,6 +83,4 @@ def __init__(self, *, additional_properties=None, source_retry_count=None, sourc self.rfc_table_options = rfc_table_options self.batch_size = batch_size self.custom_rfc_read_table_function_module = custom_rfc_read_table_function_module - self.partition_option = partition_option - self.partition_settings = partition_settings self.type = 'SapTableSource' diff --git a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/teradata_linked_service.py b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/teradata_linked_service.py index 6e02b0d389ab..78b89638b359 100644 --- a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/teradata_linked_service.py +++ b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/teradata_linked_service.py @@ -33,11 +33,8 @@ class TeradataLinkedService(LinkedService): :type annotations: list[object] :param type: Required. Constant filled by server. :type type: str - :param connection_string: Teradata ODBC connection string. Type: string, - SecureString or AzureKeyVaultSecretReference. - :type connection_string: object - :param server: Server name for connection. Type: string (or Expression - with resultType string). + :param server: Required. Server name for connection. Type: string (or + Expression with resultType string). :type server: object :param authentication_type: AuthenticationType to be used for connection. Possible values include: 'Basic', 'Windows' @@ -56,6 +53,7 @@ class TeradataLinkedService(LinkedService): _validation = { 'type': {'required': True}, + 'server': {'required': True}, } _attribute_map = { @@ -65,7 +63,6 @@ class TeradataLinkedService(LinkedService): 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, 'type': {'key': 'type', 'type': 'str'}, - 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, 'server': {'key': 'typeProperties.server', 'type': 'object'}, 'authentication_type': {'key': 'typeProperties.authenticationType', 'type': 'str'}, 'username': {'key': 'typeProperties.username', 'type': 'object'}, @@ -75,7 +72,6 @@ class TeradataLinkedService(LinkedService): def __init__(self, **kwargs): super(TeradataLinkedService, self).__init__(**kwargs) - self.connection_string = kwargs.get('connection_string', None) self.server = kwargs.get('server', None) self.authentication_type = kwargs.get('authentication_type', None) self.username = kwargs.get('username', None) diff --git a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/teradata_linked_service_py3.py b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/teradata_linked_service_py3.py index aac40efe69e0..e80b776454c0 100644 --- a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/teradata_linked_service_py3.py +++ b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/teradata_linked_service_py3.py @@ -33,11 +33,8 @@ class TeradataLinkedService(LinkedService): :type annotations: list[object] :param type: Required. Constant filled by server. :type type: str - :param connection_string: Teradata ODBC connection string. Type: string, - SecureString or AzureKeyVaultSecretReference. - :type connection_string: object - :param server: Server name for connection. Type: string (or Expression - with resultType string). + :param server: Required. Server name for connection. Type: string (or + Expression with resultType string). :type server: object :param authentication_type: AuthenticationType to be used for connection. Possible values include: 'Basic', 'Windows' @@ -56,6 +53,7 @@ class TeradataLinkedService(LinkedService): _validation = { 'type': {'required': True}, + 'server': {'required': True}, } _attribute_map = { @@ -65,7 +63,6 @@ class TeradataLinkedService(LinkedService): 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, 'type': {'key': 'type', 'type': 'str'}, - 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, 'server': {'key': 'typeProperties.server', 'type': 'object'}, 'authentication_type': {'key': 'typeProperties.authenticationType', 'type': 'str'}, 'username': {'key': 'typeProperties.username', 'type': 'object'}, @@ -73,9 +70,8 @@ class TeradataLinkedService(LinkedService): 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, } - def __init__(self, *, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, connection_string=None, server=None, authentication_type=None, username=None, password=None, encrypted_credential=None, **kwargs) -> None: + def __init__(self, *, server, additional_properties=None, connect_via=None, description: str=None, parameters=None, annotations=None, authentication_type=None, username=None, password=None, encrypted_credential=None, **kwargs) -> None: super(TeradataLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) - self.connection_string = connection_string self.server = server self.authentication_type = authentication_type self.username = username diff --git a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/teradata_partition_settings.py b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/teradata_partition_settings.py deleted file mode 100644 index 81f55ffad16c..000000000000 --- a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/teradata_partition_settings.py +++ /dev/null @@ -1,42 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for -# license information. -# -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is -# regenerated. -# -------------------------------------------------------------------------- - -from msrest.serialization import Model - - -class TeradataPartitionSettings(Model): - """The settings that will be leveraged for teradata source partitioning. - - :param partition_column_name: The name of the column in integer type that - will be used for proceeding range or hash partitioning. Type: string (or - Expression with resultType string). - :type partition_column_name: object - :param partition_upper_bound: The maximum value of column specified in - partitionColumnName that will be used for proceeding range partitioning. - Type: string (or Expression with resultType string). - :type partition_upper_bound: object - :param partition_lower_bound: The minimum value of column specified in - partitionColumnName that will be used for proceeding range partitioning. - Type: string (or Expression with resultType string). - :type partition_lower_bound: object - """ - - _attribute_map = { - 'partition_column_name': {'key': 'partitionColumnName', 'type': 'object'}, - 'partition_upper_bound': {'key': 'partitionUpperBound', 'type': 'object'}, - 'partition_lower_bound': {'key': 'partitionLowerBound', 'type': 'object'}, - } - - def __init__(self, **kwargs): - super(TeradataPartitionSettings, self).__init__(**kwargs) - self.partition_column_name = kwargs.get('partition_column_name', None) - self.partition_upper_bound = kwargs.get('partition_upper_bound', None) - self.partition_lower_bound = kwargs.get('partition_lower_bound', None) diff --git a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/teradata_partition_settings_py3.py b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/teradata_partition_settings_py3.py deleted file mode 100644 index b8b4032e8de4..000000000000 --- a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/teradata_partition_settings_py3.py +++ /dev/null @@ -1,42 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for -# license information. -# -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is -# regenerated. -# -------------------------------------------------------------------------- - -from msrest.serialization import Model - - -class TeradataPartitionSettings(Model): - """The settings that will be leveraged for teradata source partitioning. - - :param partition_column_name: The name of the column in integer type that - will be used for proceeding range or hash partitioning. Type: string (or - Expression with resultType string). - :type partition_column_name: object - :param partition_upper_bound: The maximum value of column specified in - partitionColumnName that will be used for proceeding range partitioning. - Type: string (or Expression with resultType string). - :type partition_upper_bound: object - :param partition_lower_bound: The minimum value of column specified in - partitionColumnName that will be used for proceeding range partitioning. - Type: string (or Expression with resultType string). - :type partition_lower_bound: object - """ - - _attribute_map = { - 'partition_column_name': {'key': 'partitionColumnName', 'type': 'object'}, - 'partition_upper_bound': {'key': 'partitionUpperBound', 'type': 'object'}, - 'partition_lower_bound': {'key': 'partitionLowerBound', 'type': 'object'}, - } - - def __init__(self, *, partition_column_name=None, partition_upper_bound=None, partition_lower_bound=None, **kwargs) -> None: - super(TeradataPartitionSettings, self).__init__(**kwargs) - self.partition_column_name = partition_column_name - self.partition_upper_bound = partition_upper_bound - self.partition_lower_bound = partition_lower_bound diff --git a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/teradata_source.py b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/teradata_source.py deleted file mode 100644 index 3e8a9adfab9a..000000000000 --- a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/teradata_source.py +++ /dev/null @@ -1,68 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for -# license information. -# -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is -# regenerated. -# -------------------------------------------------------------------------- - -from .copy_source import CopySource - - -class TeradataSource(CopySource): - """A copy activity Teradata source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are - deserialized this collection - :type additional_properties: dict[str, object] - :param source_retry_count: Source retry count. Type: integer (or - Expression with resultType integer). - :type source_retry_count: object - :param source_retry_wait: Source retry wait. Type: string (or Expression - with resultType string), pattern: - ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: object - :param max_concurrent_connections: The maximum concurrent connection count - for the source data store. Type: integer (or Expression with resultType - integer). - :type max_concurrent_connections: object - :param type: Required. Constant filled by server. - :type type: str - :param query: Teradata query. Type: string (or Expression with resultType - string). - :type query: object - :param partition_option: The partition mechanism that will be used for - teradata read in parallel. - :type partition_option: object - :param partition_settings: The settings that will be leveraged for - teradata source partitioning. - :type partition_settings: - ~azure.mgmt.datafactory.models.TeradataPartitionSettings - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'type': {'key': 'type', 'type': 'str'}, - 'query': {'key': 'query', 'type': 'object'}, - 'partition_option': {'key': 'partitionOption', 'type': 'object'}, - 'partition_settings': {'key': 'partitionSettings', 'type': 'TeradataPartitionSettings'}, - } - - def __init__(self, **kwargs): - super(TeradataSource, self).__init__(**kwargs) - self.query = kwargs.get('query', None) - self.partition_option = kwargs.get('partition_option', None) - self.partition_settings = kwargs.get('partition_settings', None) - self.type = 'TeradataSource' diff --git a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/teradata_source_py3.py b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/teradata_source_py3.py deleted file mode 100644 index f01fb531eb55..000000000000 --- a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/teradata_source_py3.py +++ /dev/null @@ -1,68 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for -# license information. -# -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is -# regenerated. -# -------------------------------------------------------------------------- - -from .copy_source_py3 import CopySource - - -class TeradataSource(CopySource): - """A copy activity Teradata source. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are - deserialized this collection - :type additional_properties: dict[str, object] - :param source_retry_count: Source retry count. Type: integer (or - Expression with resultType integer). - :type source_retry_count: object - :param source_retry_wait: Source retry wait. Type: string (or Expression - with resultType string), pattern: - ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :type source_retry_wait: object - :param max_concurrent_connections: The maximum concurrent connection count - for the source data store. Type: integer (or Expression with resultType - integer). - :type max_concurrent_connections: object - :param type: Required. Constant filled by server. - :type type: str - :param query: Teradata query. Type: string (or Expression with resultType - string). - :type query: object - :param partition_option: The partition mechanism that will be used for - teradata read in parallel. - :type partition_option: object - :param partition_settings: The settings that will be leveraged for - teradata source partitioning. - :type partition_settings: - ~azure.mgmt.datafactory.models.TeradataPartitionSettings - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, - 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, - 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, - 'type': {'key': 'type', 'type': 'str'}, - 'query': {'key': 'query', 'type': 'object'}, - 'partition_option': {'key': 'partitionOption', 'type': 'object'}, - 'partition_settings': {'key': 'partitionSettings', 'type': 'TeradataPartitionSettings'}, - } - - def __init__(self, *, additional_properties=None, source_retry_count=None, source_retry_wait=None, max_concurrent_connections=None, query=None, partition_option=None, partition_settings=None, **kwargs) -> None: - super(TeradataSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, max_concurrent_connections=max_concurrent_connections, **kwargs) - self.query = query - self.partition_option = partition_option - self.partition_settings = partition_settings - self.type = 'TeradataSource' diff --git a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/teradata_table_dataset.py b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/teradata_table_dataset.py deleted file mode 100644 index e396bfd6fb15..000000000000 --- a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/teradata_table_dataset.py +++ /dev/null @@ -1,77 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for -# license information. -# -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is -# regenerated. -# -------------------------------------------------------------------------- - -from .dataset import Dataset - - -class TeradataTableDataset(Dataset): - """The Teradata database dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are - deserialized this collection - :type additional_properties: dict[str, object] - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: - array (or Expression with resultType array), itemType: DatasetDataElement. - :type structure: object - :param schema: Columns that define the physical type schema of the - dataset. Type: array (or Expression with resultType array), itemType: - DatasetSchemaDataElement. - :type schema: object - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.mgmt.datafactory.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.mgmt.datafactory.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the - Dataset. - :type annotations: list[object] - :param folder: The folder that this Dataset is in. If not specified, - Dataset will appear at the root level. - :type folder: ~azure.mgmt.datafactory.models.DatasetFolder - :param type: Required. Constant filled by server. - :type type: str - :param database: The database name of Teradata. Type: string (or - Expression with resultType string). - :type database: object - :param table: The table name of Teradata. Type: string (or Expression with - resultType string). - :type table: object - """ - - _validation = { - 'linked_service_name': {'required': True}, - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'type': {'key': 'type', 'type': 'str'}, - 'database': {'key': 'typeProperties.database', 'type': 'object'}, - 'table': {'key': 'typeProperties.table', 'type': 'object'}, - } - - def __init__(self, **kwargs): - super(TeradataTableDataset, self).__init__(**kwargs) - self.database = kwargs.get('database', None) - self.table = kwargs.get('table', None) - self.type = 'TeradataTable' diff --git a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/teradata_table_dataset_py3.py b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/teradata_table_dataset_py3.py deleted file mode 100644 index 892707b7f133..000000000000 --- a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/teradata_table_dataset_py3.py +++ /dev/null @@ -1,77 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for -# license information. -# -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is -# regenerated. -# -------------------------------------------------------------------------- - -from .dataset_py3 import Dataset - - -class TeradataTableDataset(Dataset): - """The Teradata database dataset. - - All required parameters must be populated in order to send to Azure. - - :param additional_properties: Unmatched properties from the message are - deserialized this collection - :type additional_properties: dict[str, object] - :param description: Dataset description. - :type description: str - :param structure: Columns that define the structure of the dataset. Type: - array (or Expression with resultType array), itemType: DatasetDataElement. - :type structure: object - :param schema: Columns that define the physical type schema of the - dataset. Type: array (or Expression with resultType array), itemType: - DatasetSchemaDataElement. - :type schema: object - :param linked_service_name: Required. Linked service reference. - :type linked_service_name: - ~azure.mgmt.datafactory.models.LinkedServiceReference - :param parameters: Parameters for dataset. - :type parameters: dict[str, - ~azure.mgmt.datafactory.models.ParameterSpecification] - :param annotations: List of tags that can be used for describing the - Dataset. - :type annotations: list[object] - :param folder: The folder that this Dataset is in. If not specified, - Dataset will appear at the root level. - :type folder: ~azure.mgmt.datafactory.models.DatasetFolder - :param type: Required. Constant filled by server. - :type type: str - :param database: The database name of Teradata. Type: string (or - Expression with resultType string). - :type database: object - :param table: The table name of Teradata. Type: string (or Expression with - resultType string). - :type table: object - """ - - _validation = { - 'linked_service_name': {'required': True}, - 'type': {'required': True}, - } - - _attribute_map = { - 'additional_properties': {'key': '', 'type': '{object}'}, - 'description': {'key': 'description', 'type': 'str'}, - 'structure': {'key': 'structure', 'type': 'object'}, - 'schema': {'key': 'schema', 'type': 'object'}, - 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, - 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, - 'annotations': {'key': 'annotations', 'type': '[object]'}, - 'folder': {'key': 'folder', 'type': 'DatasetFolder'}, - 'type': {'key': 'type', 'type': 'str'}, - 'database': {'key': 'typeProperties.database', 'type': 'object'}, - 'table': {'key': 'typeProperties.table', 'type': 'object'}, - } - - def __init__(self, *, linked_service_name, additional_properties=None, description: str=None, structure=None, schema=None, parameters=None, annotations=None, folder=None, database=None, table=None, **kwargs) -> None: - super(TeradataTableDataset, self).__init__(additional_properties=additional_properties, description=description, structure=structure, schema=schema, linked_service_name=linked_service_name, parameters=parameters, annotations=annotations, folder=folder, **kwargs) - self.database = database - self.table = table - self.type = 'TeradataTable' From 6afadc1e386febeaaccc4c31456dd8909ecf542b Mon Sep 17 00:00:00 2001 From: Azure SDK for Python bot Date: Tue, 25 Jun 2019 05:10:58 +0000 Subject: [PATCH 2/2] Generated from 76032c5b6d424dceb3a9b03b7df79e009eb5c183 Change XxxSetting to XxxSettings in private swagger --- .../azure/mgmt/datafactory/models/__init__.py | 114 +++++++++--------- ..._setting.py => amazon_s3_read_settings.py} | 6 +- ..._py3.py => amazon_s3_read_settings_py3.py} | 6 +- ...ting.py => azure_blob_fs_read_settings.py} | 6 +- ....py => azure_blob_fs_read_settings_py3.py} | 6 +- ...ing.py => azure_blob_fs_write_settings.py} | 6 +- ...py => azure_blob_fs_write_settings_py3.py} | 8 +- ...py => azure_blob_storage_read_settings.py} | 6 +- ...> azure_blob_storage_read_settings_py3.py} | 6 +- ...y => azure_blob_storage_write_settings.py} | 6 +- ... azure_blob_storage_write_settings_py3.py} | 6 +- ...=> azure_data_lake_store_read_settings.py} | 6 +- ...zure_data_lake_store_read_settings_py3.py} | 6 +- ...> azure_data_lake_store_write_settings.py} | 6 +- ...ure_data_lake_store_write_settings_py3.py} | 6 +- ...ing.py => delimited_text_read_settings.py} | 6 +- ...py => delimited_text_read_settings_py3.py} | 6 +- .../datafactory/models/delimited_text_sink.py | 8 +- .../models/delimited_text_sink_py3.py | 8 +- .../models/delimited_text_source.py | 8 +- .../models/delimited_text_source_py3.py | 8 +- ...ng.py => delimited_text_write_settings.py} | 6 +- ...y => delimited_text_write_settings_py3.py} | 6 +- ...etting.py => file_server_read_settings.py} | 6 +- ...y3.py => file_server_read_settings_py3.py} | 6 +- ...tting.py => file_server_write_settings.py} | 6 +- ...3.py => file_server_write_settings_py3.py} | 8 +- ...ead_setting.py => format_read_settings.py} | 4 +- ...ing_py3.py => format_read_settings_py3.py} | 4 +- ...te_setting.py => format_write_settings.py} | 4 +- ...ng_py3.py => format_write_settings_py3.py} | 4 +- ...p_read_setting.py => ftp_read_settings.py} | 6 +- ...etting_py3.py => ftp_read_settings_py3.py} | 6 +- ..._read_setting.py => hdfs_read_settings.py} | 6 +- ...tting_py3.py => hdfs_read_settings_py3.py} | 6 +- ..._read_setting.py => http_read_settings.py} | 6 +- ...tting_py3.py => http_read_settings_py3.py} | 6 +- .../mgmt/datafactory/models/parquet_sink.py | 4 +- .../datafactory/models/parquet_sink_py3.py | 4 +- .../mgmt/datafactory/models/parquet_source.py | 4 +- .../datafactory/models/parquet_source_py3.py | 4 +- ..._read_setting.py => sftp_read_settings.py} | 6 +- ...tting_py3.py => sftp_read_settings_py3.py} | 6 +- ...read_setting.py => store_read_settings.py} | 4 +- ...ting_py3.py => store_read_settings_py3.py} | 4 +- ...ite_setting.py => store_write_settings.py} | 4 +- ...ing_py3.py => store_write_settings_py3.py} | 4 +- 47 files changed, 189 insertions(+), 189 deletions(-) rename sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/{amazon_s3_read_setting.py => amazon_s3_read_settings.py} (95%) rename sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/{amazon_s3_read_setting_py3.py => amazon_s3_read_settings_py3.py} (92%) rename sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/{azure_blob_fs_read_setting.py => azure_blob_fs_read_settings.py} (95%) rename sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/{azure_blob_fs_read_setting_py3.py => azure_blob_fs_read_settings_py3.py} (92%) rename sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/{azure_blob_fs_write_setting.py => azure_blob_fs_write_settings.py} (89%) rename sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/{file_server_write_setting_py3.py => azure_blob_fs_write_settings_py3.py} (81%) rename sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/{azure_blob_storage_read_setting.py => azure_blob_storage_read_settings.py} (94%) rename sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/{azure_blob_storage_read_setting_py3.py => azure_blob_storage_read_settings_py3.py} (92%) rename sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/{azure_blob_storage_write_setting.py => azure_blob_storage_write_settings.py} (89%) rename sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/{azure_blob_storage_write_setting_py3.py => azure_blob_storage_write_settings_py3.py} (83%) rename sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/{azure_data_lake_store_read_setting.py => azure_data_lake_store_read_settings.py} (94%) rename sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/{azure_data_lake_store_read_setting_py3.py => azure_data_lake_store_read_settings_py3.py} (91%) rename sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/{azure_data_lake_store_write_setting.py => azure_data_lake_store_write_settings.py} (89%) rename sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/{azure_data_lake_store_write_setting_py3.py => azure_data_lake_store_write_settings_py3.py} (83%) rename sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/{delimited_text_read_setting.py => delimited_text_read_settings.py} (89%) rename sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/{delimited_text_read_setting_py3.py => delimited_text_read_settings_py3.py} (86%) rename sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/{delimited_text_write_setting.py => delimited_text_write_settings.py} (90%) rename sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/{delimited_text_write_setting_py3.py => delimited_text_write_settings_py3.py} (88%) rename sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/{file_server_read_setting.py => file_server_read_settings.py} (95%) rename sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/{file_server_read_setting_py3.py => file_server_read_settings_py3.py} (92%) rename sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/{file_server_write_setting.py => file_server_write_settings.py} (89%) rename sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/{azure_blob_fs_write_setting_py3.py => file_server_write_settings_py3.py} (88%) rename sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/{format_read_setting.py => format_read_settings.py} (93%) rename sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/{format_read_setting_py3.py => format_read_settings_py3.py} (93%) rename sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/{format_write_setting.py => format_write_settings.py} (92%) rename sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/{format_write_setting_py3.py => format_write_settings_py3.py} (92%) rename sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/{ftp_read_setting.py => ftp_read_settings.py} (94%) rename sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/{ftp_read_setting_py3.py => ftp_read_settings_py3.py} (90%) rename sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/{hdfs_read_setting.py => hdfs_read_settings.py} (95%) rename sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/{hdfs_read_setting_py3.py => hdfs_read_settings_py3.py} (93%) rename sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/{http_read_setting.py => http_read_settings.py} (94%) rename sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/{http_read_setting_py3.py => http_read_settings_py3.py} (90%) rename sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/{sftp_read_setting.py => sftp_read_settings.py} (94%) rename sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/{sftp_read_setting_py3.py => sftp_read_settings_py3.py} (91%) rename sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/{connector_read_setting.py => store_read_settings.py} (94%) rename sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/{connector_read_setting_py3.py => store_read_settings_py3.py} (94%) rename sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/{connector_write_setting.py => store_write_settings.py} (94%) rename sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/{connector_write_setting_py3.py => store_write_settings_py3.py} (95%) diff --git a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py index 70ebe10afa94..27c22b42e671 100644 --- a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py +++ b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py @@ -351,18 +351,18 @@ from .document_db_collection_source_py3 import DocumentDbCollectionSource from .blob_source_py3 import BlobSource from .azure_table_source_py3 import AzureTableSource - from .format_read_setting_py3 import FormatReadSetting - from .delimited_text_read_setting_py3 import DelimitedTextReadSetting - from .hdfs_read_setting_py3 import HdfsReadSetting - from .http_read_setting_py3 import HttpReadSetting - from .sftp_read_setting_py3 import SftpReadSetting - from .ftp_read_setting_py3 import FtpReadSetting - from .file_server_read_setting_py3 import FileServerReadSetting - from .amazon_s3_read_setting_py3 import AmazonS3ReadSetting - from .azure_data_lake_store_read_setting_py3 import AzureDataLakeStoreReadSetting - from .azure_blob_fs_read_setting_py3 import AzureBlobFSReadSetting - from .azure_blob_storage_read_setting_py3 import AzureBlobStorageReadSetting - from .connector_read_setting_py3 import ConnectorReadSetting + from .format_read_settings_py3 import FormatReadSettings + from .delimited_text_read_settings_py3 import DelimitedTextReadSettings + from .hdfs_read_settings_py3 import HdfsReadSettings + from .http_read_settings_py3 import HttpReadSettings + from .sftp_read_settings_py3 import SftpReadSettings + from .ftp_read_settings_py3 import FtpReadSettings + from .file_server_read_settings_py3 import FileServerReadSettings + from .amazon_s3_read_settings_py3 import AmazonS3ReadSettings + from .azure_data_lake_store_read_settings_py3 import AzureDataLakeStoreReadSettings + from .azure_blob_fs_read_settings_py3 import AzureBlobFSReadSettings + from .azure_blob_storage_read_settings_py3 import AzureBlobStorageReadSettings + from .store_read_settings_py3 import StoreReadSettings from .delimited_text_source_py3 import DelimitedTextSource from .parquet_source_py3 import ParquetSource from .copy_source_py3 import CopySource @@ -401,17 +401,17 @@ from .document_db_collection_sink_py3 import DocumentDbCollectionSink from .file_system_sink_py3 import FileSystemSink from .blob_sink_py3 import BlobSink - from .file_server_write_setting_py3 import FileServerWriteSetting - from .azure_data_lake_store_write_setting_py3 import AzureDataLakeStoreWriteSetting - from .azure_blob_fs_write_setting_py3 import AzureBlobFSWriteSetting - from .azure_blob_storage_write_setting_py3 import AzureBlobStorageWriteSetting - from .connector_write_setting_py3 import ConnectorWriteSetting + from .file_server_write_settings_py3 import FileServerWriteSettings + from .azure_data_lake_store_write_settings_py3 import AzureDataLakeStoreWriteSettings + from .azure_blob_fs_write_settings_py3 import AzureBlobFSWriteSettings + from .azure_blob_storage_write_settings_py3 import AzureBlobStorageWriteSettings + from .store_write_settings_py3 import StoreWriteSettings from .parquet_sink_py3 import ParquetSink from .azure_table_sink_py3 import AzureTableSink from .azure_queue_sink_py3 import AzureQueueSink from .sap_cloud_for_customer_sink_py3 import SapCloudForCustomerSink - from .format_write_setting_py3 import FormatWriteSetting - from .delimited_text_write_setting_py3 import DelimitedTextWriteSetting + from .format_write_settings_py3 import FormatWriteSettings + from .delimited_text_write_settings_py3 import DelimitedTextWriteSettings from .delimited_text_sink_py3 import DelimitedTextSink from .copy_sink_py3 import CopySink from .copy_activity_py3 import CopyActivity @@ -803,18 +803,18 @@ from .document_db_collection_source import DocumentDbCollectionSource from .blob_source import BlobSource from .azure_table_source import AzureTableSource - from .format_read_setting import FormatReadSetting - from .delimited_text_read_setting import DelimitedTextReadSetting - from .hdfs_read_setting import HdfsReadSetting - from .http_read_setting import HttpReadSetting - from .sftp_read_setting import SftpReadSetting - from .ftp_read_setting import FtpReadSetting - from .file_server_read_setting import FileServerReadSetting - from .amazon_s3_read_setting import AmazonS3ReadSetting - from .azure_data_lake_store_read_setting import AzureDataLakeStoreReadSetting - from .azure_blob_fs_read_setting import AzureBlobFSReadSetting - from .azure_blob_storage_read_setting import AzureBlobStorageReadSetting - from .connector_read_setting import ConnectorReadSetting + from .format_read_settings import FormatReadSettings + from .delimited_text_read_settings import DelimitedTextReadSettings + from .hdfs_read_settings import HdfsReadSettings + from .http_read_settings import HttpReadSettings + from .sftp_read_settings import SftpReadSettings + from .ftp_read_settings import FtpReadSettings + from .file_server_read_settings import FileServerReadSettings + from .amazon_s3_read_settings import AmazonS3ReadSettings + from .azure_data_lake_store_read_settings import AzureDataLakeStoreReadSettings + from .azure_blob_fs_read_settings import AzureBlobFSReadSettings + from .azure_blob_storage_read_settings import AzureBlobStorageReadSettings + from .store_read_settings import StoreReadSettings from .delimited_text_source import DelimitedTextSource from .parquet_source import ParquetSource from .copy_source import CopySource @@ -853,17 +853,17 @@ from .document_db_collection_sink import DocumentDbCollectionSink from .file_system_sink import FileSystemSink from .blob_sink import BlobSink - from .file_server_write_setting import FileServerWriteSetting - from .azure_data_lake_store_write_setting import AzureDataLakeStoreWriteSetting - from .azure_blob_fs_write_setting import AzureBlobFSWriteSetting - from .azure_blob_storage_write_setting import AzureBlobStorageWriteSetting - from .connector_write_setting import ConnectorWriteSetting + from .file_server_write_settings import FileServerWriteSettings + from .azure_data_lake_store_write_settings import AzureDataLakeStoreWriteSettings + from .azure_blob_fs_write_settings import AzureBlobFSWriteSettings + from .azure_blob_storage_write_settings import AzureBlobStorageWriteSettings + from .store_write_settings import StoreWriteSettings from .parquet_sink import ParquetSink from .azure_table_sink import AzureTableSink from .azure_queue_sink import AzureQueueSink from .sap_cloud_for_customer_sink import SapCloudForCustomerSink - from .format_write_setting import FormatWriteSetting - from .delimited_text_write_setting import DelimitedTextWriteSetting + from .format_write_settings import FormatWriteSettings + from .delimited_text_write_settings import DelimitedTextWriteSettings from .delimited_text_sink import DelimitedTextSink from .copy_sink import CopySink from .copy_activity import CopyActivity @@ -1328,18 +1328,18 @@ 'DocumentDbCollectionSource', 'BlobSource', 'AzureTableSource', - 'FormatReadSetting', - 'DelimitedTextReadSetting', - 'HdfsReadSetting', - 'HttpReadSetting', - 'SftpReadSetting', - 'FtpReadSetting', - 'FileServerReadSetting', - 'AmazonS3ReadSetting', - 'AzureDataLakeStoreReadSetting', - 'AzureBlobFSReadSetting', - 'AzureBlobStorageReadSetting', - 'ConnectorReadSetting', + 'FormatReadSettings', + 'DelimitedTextReadSettings', + 'HdfsReadSettings', + 'HttpReadSettings', + 'SftpReadSettings', + 'FtpReadSettings', + 'FileServerReadSettings', + 'AmazonS3ReadSettings', + 'AzureDataLakeStoreReadSettings', + 'AzureBlobFSReadSettings', + 'AzureBlobStorageReadSettings', + 'StoreReadSettings', 'DelimitedTextSource', 'ParquetSource', 'CopySource', @@ -1378,17 +1378,17 @@ 'DocumentDbCollectionSink', 'FileSystemSink', 'BlobSink', - 'FileServerWriteSetting', - 'AzureDataLakeStoreWriteSetting', - 'AzureBlobFSWriteSetting', - 'AzureBlobStorageWriteSetting', - 'ConnectorWriteSetting', + 'FileServerWriteSettings', + 'AzureDataLakeStoreWriteSettings', + 'AzureBlobFSWriteSettings', + 'AzureBlobStorageWriteSettings', + 'StoreWriteSettings', 'ParquetSink', 'AzureTableSink', 'AzureQueueSink', 'SapCloudForCustomerSink', - 'FormatWriteSetting', - 'DelimitedTextWriteSetting', + 'FormatWriteSettings', + 'DelimitedTextWriteSettings', 'DelimitedTextSink', 'CopySink', 'CopyActivity', diff --git a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_s3_read_setting.py b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_s3_read_settings.py similarity index 95% rename from sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_s3_read_setting.py rename to sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_s3_read_settings.py index 4de7e0ebb7b9..e83910136070 100644 --- a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_s3_read_setting.py +++ b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_s3_read_settings.py @@ -9,10 +9,10 @@ # regenerated. # -------------------------------------------------------------------------- -from .connector_read_setting import ConnectorReadSetting +from .store_read_settings import StoreReadSettings -class AmazonS3ReadSetting(ConnectorReadSetting): +class AmazonS3ReadSettings(StoreReadSettings): """Azure data lake store read settings. All required parameters must be populated in order to send to Azure. @@ -68,7 +68,7 @@ class AmazonS3ReadSetting(ConnectorReadSetting): } def __init__(self, **kwargs): - super(AmazonS3ReadSetting, self).__init__(**kwargs) + super(AmazonS3ReadSettings, self).__init__(**kwargs) self.recursive = kwargs.get('recursive', None) self.wildcard_folder_path = kwargs.get('wildcard_folder_path', None) self.wildcard_file_name = kwargs.get('wildcard_file_name', None) diff --git a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_s3_read_setting_py3.py b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_s3_read_settings_py3.py similarity index 92% rename from sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_s3_read_setting_py3.py rename to sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_s3_read_settings_py3.py index deda331ea561..79645a869ac8 100644 --- a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_s3_read_setting_py3.py +++ b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_s3_read_settings_py3.py @@ -9,10 +9,10 @@ # regenerated. # -------------------------------------------------------------------------- -from .connector_read_setting_py3 import ConnectorReadSetting +from .store_read_settings_py3 import StoreReadSettings -class AmazonS3ReadSetting(ConnectorReadSetting): +class AmazonS3ReadSettings(StoreReadSettings): """Azure data lake store read settings. All required parameters must be populated in order to send to Azure. @@ -68,7 +68,7 @@ class AmazonS3ReadSetting(ConnectorReadSetting): } def __init__(self, *, type: str, additional_properties=None, max_concurrent_connections=None, recursive=None, wildcard_folder_path=None, wildcard_file_name=None, prefix=None, enable_partition_discovery: bool=None, modified_datetime_start=None, modified_datetime_end=None, **kwargs) -> None: - super(AmazonS3ReadSetting, self).__init__(additional_properties=additional_properties, type=type, max_concurrent_connections=max_concurrent_connections, **kwargs) + super(AmazonS3ReadSettings, self).__init__(additional_properties=additional_properties, type=type, max_concurrent_connections=max_concurrent_connections, **kwargs) self.recursive = recursive self.wildcard_folder_path = wildcard_folder_path self.wildcard_file_name = wildcard_file_name diff --git a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_blob_fs_read_setting.py b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_blob_fs_read_settings.py similarity index 95% rename from sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_blob_fs_read_setting.py rename to sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_blob_fs_read_settings.py index 11490a288417..6d80ce72ea57 100644 --- a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_blob_fs_read_setting.py +++ b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_blob_fs_read_settings.py @@ -9,10 +9,10 @@ # regenerated. # -------------------------------------------------------------------------- -from .connector_read_setting import ConnectorReadSetting +from .store_read_settings import StoreReadSettings -class AzureBlobFSReadSetting(ConnectorReadSetting): +class AzureBlobFSReadSettings(StoreReadSettings): """Azure blobFS read settings. All required parameters must be populated in order to send to Azure. @@ -64,7 +64,7 @@ class AzureBlobFSReadSetting(ConnectorReadSetting): } def __init__(self, **kwargs): - super(AzureBlobFSReadSetting, self).__init__(**kwargs) + super(AzureBlobFSReadSettings, self).__init__(**kwargs) self.recursive = kwargs.get('recursive', None) self.wildcard_folder_path = kwargs.get('wildcard_folder_path', None) self.wildcard_file_name = kwargs.get('wildcard_file_name', None) diff --git a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_blob_fs_read_setting_py3.py b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_blob_fs_read_settings_py3.py similarity index 92% rename from sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_blob_fs_read_setting_py3.py rename to sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_blob_fs_read_settings_py3.py index 28f3b4f7ceb4..af4746e84f8e 100644 --- a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_blob_fs_read_setting_py3.py +++ b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_blob_fs_read_settings_py3.py @@ -9,10 +9,10 @@ # regenerated. # -------------------------------------------------------------------------- -from .connector_read_setting_py3 import ConnectorReadSetting +from .store_read_settings_py3 import StoreReadSettings -class AzureBlobFSReadSetting(ConnectorReadSetting): +class AzureBlobFSReadSettings(StoreReadSettings): """Azure blobFS read settings. All required parameters must be populated in order to send to Azure. @@ -64,7 +64,7 @@ class AzureBlobFSReadSetting(ConnectorReadSetting): } def __init__(self, *, type: str, additional_properties=None, max_concurrent_connections=None, recursive=None, wildcard_folder_path=None, wildcard_file_name=None, enable_partition_discovery: bool=None, modified_datetime_start=None, modified_datetime_end=None, **kwargs) -> None: - super(AzureBlobFSReadSetting, self).__init__(additional_properties=additional_properties, type=type, max_concurrent_connections=max_concurrent_connections, **kwargs) + super(AzureBlobFSReadSettings, self).__init__(additional_properties=additional_properties, type=type, max_concurrent_connections=max_concurrent_connections, **kwargs) self.recursive = recursive self.wildcard_folder_path = wildcard_folder_path self.wildcard_file_name = wildcard_file_name diff --git a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_blob_fs_write_setting.py b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_blob_fs_write_settings.py similarity index 89% rename from sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_blob_fs_write_setting.py rename to sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_blob_fs_write_settings.py index d5b2d850da58..5de93c10a1f8 100644 --- a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_blob_fs_write_setting.py +++ b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_blob_fs_write_settings.py @@ -9,10 +9,10 @@ # regenerated. # -------------------------------------------------------------------------- -from .connector_write_setting import ConnectorWriteSetting +from .store_write_settings import StoreWriteSettings -class AzureBlobFSWriteSetting(ConnectorWriteSetting): +class AzureBlobFSWriteSettings(StoreWriteSettings): """Azure blobFS write settings. All required parameters must be populated in order to send to Azure. @@ -42,4 +42,4 @@ class AzureBlobFSWriteSetting(ConnectorWriteSetting): } def __init__(self, **kwargs): - super(AzureBlobFSWriteSetting, self).__init__(**kwargs) + super(AzureBlobFSWriteSettings, self).__init__(**kwargs) diff --git a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/file_server_write_setting_py3.py b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_blob_fs_write_settings_py3.py similarity index 81% rename from sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/file_server_write_setting_py3.py rename to sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_blob_fs_write_settings_py3.py index 1ed4bf220417..be0008fd8733 100644 --- a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/file_server_write_setting_py3.py +++ b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_blob_fs_write_settings_py3.py @@ -9,11 +9,11 @@ # regenerated. # -------------------------------------------------------------------------- -from .connector_write_setting_py3 import ConnectorWriteSetting +from .store_write_settings_py3 import StoreWriteSettings -class FileServerWriteSetting(ConnectorWriteSetting): - """File server write settings. +class AzureBlobFSWriteSettings(StoreWriteSettings): + """Azure blobFS write settings. All required parameters must be populated in order to send to Azure. @@ -42,4 +42,4 @@ class FileServerWriteSetting(ConnectorWriteSetting): } def __init__(self, *, type: str, additional_properties=None, max_concurrent_connections=None, copy_behavior=None, **kwargs) -> None: - super(FileServerWriteSetting, self).__init__(additional_properties=additional_properties, type=type, max_concurrent_connections=max_concurrent_connections, copy_behavior=copy_behavior, **kwargs) + super(AzureBlobFSWriteSettings, self).__init__(additional_properties=additional_properties, type=type, max_concurrent_connections=max_concurrent_connections, copy_behavior=copy_behavior, **kwargs) diff --git a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_blob_storage_read_setting.py b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_blob_storage_read_settings.py similarity index 94% rename from sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_blob_storage_read_setting.py rename to sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_blob_storage_read_settings.py index ee07a3576f29..42b11cc6de16 100644 --- a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_blob_storage_read_setting.py +++ b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_blob_storage_read_settings.py @@ -9,10 +9,10 @@ # regenerated. # -------------------------------------------------------------------------- -from .connector_read_setting import ConnectorReadSetting +from .store_read_settings import StoreReadSettings -class AzureBlobStorageReadSetting(ConnectorReadSetting): +class AzureBlobStorageReadSettings(StoreReadSettings): """Azure blob read settings. All required parameters must be populated in order to send to Azure. @@ -64,7 +64,7 @@ class AzureBlobStorageReadSetting(ConnectorReadSetting): } def __init__(self, **kwargs): - super(AzureBlobStorageReadSetting, self).__init__(**kwargs) + super(AzureBlobStorageReadSettings, self).__init__(**kwargs) self.recursive = kwargs.get('recursive', None) self.wildcard_folder_path = kwargs.get('wildcard_folder_path', None) self.wildcard_file_name = kwargs.get('wildcard_file_name', None) diff --git a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_blob_storage_read_setting_py3.py b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_blob_storage_read_settings_py3.py similarity index 92% rename from sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_blob_storage_read_setting_py3.py rename to sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_blob_storage_read_settings_py3.py index 3e3d35774a46..495ea16afd98 100644 --- a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_blob_storage_read_setting_py3.py +++ b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_blob_storage_read_settings_py3.py @@ -9,10 +9,10 @@ # regenerated. # -------------------------------------------------------------------------- -from .connector_read_setting_py3 import ConnectorReadSetting +from .store_read_settings_py3 import StoreReadSettings -class AzureBlobStorageReadSetting(ConnectorReadSetting): +class AzureBlobStorageReadSettings(StoreReadSettings): """Azure blob read settings. All required parameters must be populated in order to send to Azure. @@ -64,7 +64,7 @@ class AzureBlobStorageReadSetting(ConnectorReadSetting): } def __init__(self, *, type: str, additional_properties=None, max_concurrent_connections=None, recursive=None, wildcard_folder_path=None, wildcard_file_name=None, enable_partition_discovery: bool=None, modified_datetime_start=None, modified_datetime_end=None, **kwargs) -> None: - super(AzureBlobStorageReadSetting, self).__init__(additional_properties=additional_properties, type=type, max_concurrent_connections=max_concurrent_connections, **kwargs) + super(AzureBlobStorageReadSettings, self).__init__(additional_properties=additional_properties, type=type, max_concurrent_connections=max_concurrent_connections, **kwargs) self.recursive = recursive self.wildcard_folder_path = wildcard_folder_path self.wildcard_file_name = wildcard_file_name diff --git a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_blob_storage_write_setting.py b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_blob_storage_write_settings.py similarity index 89% rename from sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_blob_storage_write_setting.py rename to sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_blob_storage_write_settings.py index a6499dfda798..08becf208a3d 100644 --- a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_blob_storage_write_setting.py +++ b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_blob_storage_write_settings.py @@ -9,10 +9,10 @@ # regenerated. # -------------------------------------------------------------------------- -from .connector_write_setting import ConnectorWriteSetting +from .store_write_settings import StoreWriteSettings -class AzureBlobStorageWriteSetting(ConnectorWriteSetting): +class AzureBlobStorageWriteSettings(StoreWriteSettings): """Azure blob write settings. All required parameters must be populated in order to send to Azure. @@ -42,4 +42,4 @@ class AzureBlobStorageWriteSetting(ConnectorWriteSetting): } def __init__(self, **kwargs): - super(AzureBlobStorageWriteSetting, self).__init__(**kwargs) + super(AzureBlobStorageWriteSettings, self).__init__(**kwargs) diff --git a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_blob_storage_write_setting_py3.py b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_blob_storage_write_settings_py3.py similarity index 83% rename from sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_blob_storage_write_setting_py3.py rename to sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_blob_storage_write_settings_py3.py index 9abb68c06055..40f2e0103693 100644 --- a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_blob_storage_write_setting_py3.py +++ b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_blob_storage_write_settings_py3.py @@ -9,10 +9,10 @@ # regenerated. # -------------------------------------------------------------------------- -from .connector_write_setting_py3 import ConnectorWriteSetting +from .store_write_settings_py3 import StoreWriteSettings -class AzureBlobStorageWriteSetting(ConnectorWriteSetting): +class AzureBlobStorageWriteSettings(StoreWriteSettings): """Azure blob write settings. All required parameters must be populated in order to send to Azure. @@ -42,4 +42,4 @@ class AzureBlobStorageWriteSetting(ConnectorWriteSetting): } def __init__(self, *, type: str, additional_properties=None, max_concurrent_connections=None, copy_behavior=None, **kwargs) -> None: - super(AzureBlobStorageWriteSetting, self).__init__(additional_properties=additional_properties, type=type, max_concurrent_connections=max_concurrent_connections, copy_behavior=copy_behavior, **kwargs) + super(AzureBlobStorageWriteSettings, self).__init__(additional_properties=additional_properties, type=type, max_concurrent_connections=max_concurrent_connections, copy_behavior=copy_behavior, **kwargs) diff --git a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_data_lake_store_read_setting.py b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_data_lake_store_read_settings.py similarity index 94% rename from sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_data_lake_store_read_setting.py rename to sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_data_lake_store_read_settings.py index 0f0dfe7f7c58..213d69966baf 100644 --- a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_data_lake_store_read_setting.py +++ b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_data_lake_store_read_settings.py @@ -9,10 +9,10 @@ # regenerated. # -------------------------------------------------------------------------- -from .connector_read_setting import ConnectorReadSetting +from .store_read_settings import StoreReadSettings -class AzureDataLakeStoreReadSetting(ConnectorReadSetting): +class AzureDataLakeStoreReadSettings(StoreReadSettings): """Azure data lake store read settings. All required parameters must be populated in order to send to Azure. @@ -64,7 +64,7 @@ class AzureDataLakeStoreReadSetting(ConnectorReadSetting): } def __init__(self, **kwargs): - super(AzureDataLakeStoreReadSetting, self).__init__(**kwargs) + super(AzureDataLakeStoreReadSettings, self).__init__(**kwargs) self.recursive = kwargs.get('recursive', None) self.wildcard_folder_path = kwargs.get('wildcard_folder_path', None) self.wildcard_file_name = kwargs.get('wildcard_file_name', None) diff --git a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_data_lake_store_read_setting_py3.py b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_data_lake_store_read_settings_py3.py similarity index 91% rename from sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_data_lake_store_read_setting_py3.py rename to sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_data_lake_store_read_settings_py3.py index b9159463d681..b4bccc5e78a3 100644 --- a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_data_lake_store_read_setting_py3.py +++ b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_data_lake_store_read_settings_py3.py @@ -9,10 +9,10 @@ # regenerated. # -------------------------------------------------------------------------- -from .connector_read_setting_py3 import ConnectorReadSetting +from .store_read_settings_py3 import StoreReadSettings -class AzureDataLakeStoreReadSetting(ConnectorReadSetting): +class AzureDataLakeStoreReadSettings(StoreReadSettings): """Azure data lake store read settings. All required parameters must be populated in order to send to Azure. @@ -64,7 +64,7 @@ class AzureDataLakeStoreReadSetting(ConnectorReadSetting): } def __init__(self, *, type: str, additional_properties=None, max_concurrent_connections=None, recursive=None, wildcard_folder_path=None, wildcard_file_name=None, enable_partition_discovery: bool=None, modified_datetime_start=None, modified_datetime_end=None, **kwargs) -> None: - super(AzureDataLakeStoreReadSetting, self).__init__(additional_properties=additional_properties, type=type, max_concurrent_connections=max_concurrent_connections, **kwargs) + super(AzureDataLakeStoreReadSettings, self).__init__(additional_properties=additional_properties, type=type, max_concurrent_connections=max_concurrent_connections, **kwargs) self.recursive = recursive self.wildcard_folder_path = wildcard_folder_path self.wildcard_file_name = wildcard_file_name diff --git a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_data_lake_store_write_setting.py b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_data_lake_store_write_settings.py similarity index 89% rename from sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_data_lake_store_write_setting.py rename to sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_data_lake_store_write_settings.py index d7875f545e77..a372606d86ec 100644 --- a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_data_lake_store_write_setting.py +++ b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_data_lake_store_write_settings.py @@ -9,10 +9,10 @@ # regenerated. # -------------------------------------------------------------------------- -from .connector_write_setting import ConnectorWriteSetting +from .store_write_settings import StoreWriteSettings -class AzureDataLakeStoreWriteSetting(ConnectorWriteSetting): +class AzureDataLakeStoreWriteSettings(StoreWriteSettings): """Azure data lake store write settings. All required parameters must be populated in order to send to Azure. @@ -42,4 +42,4 @@ class AzureDataLakeStoreWriteSetting(ConnectorWriteSetting): } def __init__(self, **kwargs): - super(AzureDataLakeStoreWriteSetting, self).__init__(**kwargs) + super(AzureDataLakeStoreWriteSettings, self).__init__(**kwargs) diff --git a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_data_lake_store_write_setting_py3.py b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_data_lake_store_write_settings_py3.py similarity index 83% rename from sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_data_lake_store_write_setting_py3.py rename to sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_data_lake_store_write_settings_py3.py index e05ddcbaeaac..a48cade879c6 100644 --- a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_data_lake_store_write_setting_py3.py +++ b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_data_lake_store_write_settings_py3.py @@ -9,10 +9,10 @@ # regenerated. # -------------------------------------------------------------------------- -from .connector_write_setting_py3 import ConnectorWriteSetting +from .store_write_settings_py3 import StoreWriteSettings -class AzureDataLakeStoreWriteSetting(ConnectorWriteSetting): +class AzureDataLakeStoreWriteSettings(StoreWriteSettings): """Azure data lake store write settings. All required parameters must be populated in order to send to Azure. @@ -42,4 +42,4 @@ class AzureDataLakeStoreWriteSetting(ConnectorWriteSetting): } def __init__(self, *, type: str, additional_properties=None, max_concurrent_connections=None, copy_behavior=None, **kwargs) -> None: - super(AzureDataLakeStoreWriteSetting, self).__init__(additional_properties=additional_properties, type=type, max_concurrent_connections=max_concurrent_connections, copy_behavior=copy_behavior, **kwargs) + super(AzureDataLakeStoreWriteSettings, self).__init__(additional_properties=additional_properties, type=type, max_concurrent_connections=max_concurrent_connections, copy_behavior=copy_behavior, **kwargs) diff --git a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/delimited_text_read_setting.py b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/delimited_text_read_settings.py similarity index 89% rename from sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/delimited_text_read_setting.py rename to sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/delimited_text_read_settings.py index 004eb595a05e..364b103c426a 100644 --- a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/delimited_text_read_setting.py +++ b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/delimited_text_read_settings.py @@ -9,10 +9,10 @@ # regenerated. # -------------------------------------------------------------------------- -from .format_read_setting import FormatReadSetting +from .format_read_settings import FormatReadSettings -class DelimitedTextReadSetting(FormatReadSetting): +class DelimitedTextReadSettings(FormatReadSettings): """Delimited text read settings. All required parameters must be populated in order to send to Azure. @@ -39,5 +39,5 @@ class DelimitedTextReadSetting(FormatReadSetting): } def __init__(self, **kwargs): - super(DelimitedTextReadSetting, self).__init__(**kwargs) + super(DelimitedTextReadSettings, self).__init__(**kwargs) self.skip_line_count = kwargs.get('skip_line_count', None) diff --git a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/delimited_text_read_setting_py3.py b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/delimited_text_read_settings_py3.py similarity index 86% rename from sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/delimited_text_read_setting_py3.py rename to sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/delimited_text_read_settings_py3.py index 87915fcb3db7..62aa0327cfb9 100644 --- a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/delimited_text_read_setting_py3.py +++ b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/delimited_text_read_settings_py3.py @@ -9,10 +9,10 @@ # regenerated. # -------------------------------------------------------------------------- -from .format_read_setting_py3 import FormatReadSetting +from .format_read_settings_py3 import FormatReadSettings -class DelimitedTextReadSetting(FormatReadSetting): +class DelimitedTextReadSettings(FormatReadSettings): """Delimited text read settings. All required parameters must be populated in order to send to Azure. @@ -39,5 +39,5 @@ class DelimitedTextReadSetting(FormatReadSetting): } def __init__(self, *, type: str, additional_properties=None, skip_line_count=None, **kwargs) -> None: - super(DelimitedTextReadSetting, self).__init__(additional_properties=additional_properties, type=type, **kwargs) + super(DelimitedTextReadSettings, self).__init__(additional_properties=additional_properties, type=type, **kwargs) self.skip_line_count = skip_line_count diff --git a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/delimited_text_sink.py b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/delimited_text_sink.py index ae93f209c8b3..15e0e590b4ee 100644 --- a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/delimited_text_sink.py +++ b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/delimited_text_sink.py @@ -41,10 +41,10 @@ class DelimitedTextSink(CopySink): :param type: Required. Constant filled by server. :type type: str :param store_settings: DelimitedText store settings. - :type store_settings: ~azure.mgmt.datafactory.models.ConnectorWriteSetting + :type store_settings: ~azure.mgmt.datafactory.models.StoreWriteSettings :param format_settings: DelimitedText format settings. :type format_settings: - ~azure.mgmt.datafactory.models.DelimitedTextWriteSetting + ~azure.mgmt.datafactory.models.DelimitedTextWriteSettings """ _validation = { @@ -59,8 +59,8 @@ class DelimitedTextSink(CopySink): 'sink_retry_wait': {'key': 'sinkRetryWait', 'type': 'object'}, 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'type': {'key': 'type', 'type': 'str'}, - 'store_settings': {'key': 'storeSettings', 'type': 'ConnectorWriteSetting'}, - 'format_settings': {'key': 'formatSettings', 'type': 'DelimitedTextWriteSetting'}, + 'store_settings': {'key': 'storeSettings', 'type': 'StoreWriteSettings'}, + 'format_settings': {'key': 'formatSettings', 'type': 'DelimitedTextWriteSettings'}, } def __init__(self, **kwargs): diff --git a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/delimited_text_sink_py3.py b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/delimited_text_sink_py3.py index a1ba953a2662..6481f8021527 100644 --- a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/delimited_text_sink_py3.py +++ b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/delimited_text_sink_py3.py @@ -41,10 +41,10 @@ class DelimitedTextSink(CopySink): :param type: Required. Constant filled by server. :type type: str :param store_settings: DelimitedText store settings. - :type store_settings: ~azure.mgmt.datafactory.models.ConnectorWriteSetting + :type store_settings: ~azure.mgmt.datafactory.models.StoreWriteSettings :param format_settings: DelimitedText format settings. :type format_settings: - ~azure.mgmt.datafactory.models.DelimitedTextWriteSetting + ~azure.mgmt.datafactory.models.DelimitedTextWriteSettings """ _validation = { @@ -59,8 +59,8 @@ class DelimitedTextSink(CopySink): 'sink_retry_wait': {'key': 'sinkRetryWait', 'type': 'object'}, 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'type': {'key': 'type', 'type': 'str'}, - 'store_settings': {'key': 'storeSettings', 'type': 'ConnectorWriteSetting'}, - 'format_settings': {'key': 'formatSettings', 'type': 'DelimitedTextWriteSetting'}, + 'store_settings': {'key': 'storeSettings', 'type': 'StoreWriteSettings'}, + 'format_settings': {'key': 'formatSettings', 'type': 'DelimitedTextWriteSettings'}, } def __init__(self, *, additional_properties=None, write_batch_size=None, write_batch_timeout=None, sink_retry_count=None, sink_retry_wait=None, max_concurrent_connections=None, store_settings=None, format_settings=None, **kwargs) -> None: diff --git a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/delimited_text_source.py b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/delimited_text_source.py index 9f2067d24b9c..10a842ca374a 100644 --- a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/delimited_text_source.py +++ b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/delimited_text_source.py @@ -34,10 +34,10 @@ class DelimitedTextSource(CopySource): :param type: Required. Constant filled by server. :type type: str :param store_settings: DelimitedText store settings. - :type store_settings: ~azure.mgmt.datafactory.models.ConnectorReadSetting + :type store_settings: ~azure.mgmt.datafactory.models.StoreReadSettings :param format_settings: DelimitedText format settings. :type format_settings: - ~azure.mgmt.datafactory.models.DelimitedTextReadSetting + ~azure.mgmt.datafactory.models.DelimitedTextReadSettings """ _validation = { @@ -50,8 +50,8 @@ class DelimitedTextSource(CopySource): 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'type': {'key': 'type', 'type': 'str'}, - 'store_settings': {'key': 'storeSettings', 'type': 'ConnectorReadSetting'}, - 'format_settings': {'key': 'formatSettings', 'type': 'DelimitedTextReadSetting'}, + 'store_settings': {'key': 'storeSettings', 'type': 'StoreReadSettings'}, + 'format_settings': {'key': 'formatSettings', 'type': 'DelimitedTextReadSettings'}, } def __init__(self, **kwargs): diff --git a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/delimited_text_source_py3.py b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/delimited_text_source_py3.py index b158f97bde81..e551e32c847e 100644 --- a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/delimited_text_source_py3.py +++ b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/delimited_text_source_py3.py @@ -34,10 +34,10 @@ class DelimitedTextSource(CopySource): :param type: Required. Constant filled by server. :type type: str :param store_settings: DelimitedText store settings. - :type store_settings: ~azure.mgmt.datafactory.models.ConnectorReadSetting + :type store_settings: ~azure.mgmt.datafactory.models.StoreReadSettings :param format_settings: DelimitedText format settings. :type format_settings: - ~azure.mgmt.datafactory.models.DelimitedTextReadSetting + ~azure.mgmt.datafactory.models.DelimitedTextReadSettings """ _validation = { @@ -50,8 +50,8 @@ class DelimitedTextSource(CopySource): 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'type': {'key': 'type', 'type': 'str'}, - 'store_settings': {'key': 'storeSettings', 'type': 'ConnectorReadSetting'}, - 'format_settings': {'key': 'formatSettings', 'type': 'DelimitedTextReadSetting'}, + 'store_settings': {'key': 'storeSettings', 'type': 'StoreReadSettings'}, + 'format_settings': {'key': 'formatSettings', 'type': 'DelimitedTextReadSettings'}, } def __init__(self, *, additional_properties=None, source_retry_count=None, source_retry_wait=None, max_concurrent_connections=None, store_settings=None, format_settings=None, **kwargs) -> None: diff --git a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/delimited_text_write_setting.py b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/delimited_text_write_settings.py similarity index 90% rename from sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/delimited_text_write_setting.py rename to sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/delimited_text_write_settings.py index 21fe168f1316..5e0d8db319e5 100644 --- a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/delimited_text_write_setting.py +++ b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/delimited_text_write_settings.py @@ -9,10 +9,10 @@ # regenerated. # -------------------------------------------------------------------------- -from .format_write_setting import FormatWriteSetting +from .format_write_settings import FormatWriteSettings -class DelimitedTextWriteSetting(FormatWriteSetting): +class DelimitedTextWriteSettings(FormatWriteSettings): """Delimited text write settings. All required parameters must be populated in order to send to Azure. @@ -44,6 +44,6 @@ class DelimitedTextWriteSetting(FormatWriteSetting): } def __init__(self, **kwargs): - super(DelimitedTextWriteSetting, self).__init__(**kwargs) + super(DelimitedTextWriteSettings, self).__init__(**kwargs) self.quote_all_text = kwargs.get('quote_all_text', None) self.file_extension = kwargs.get('file_extension', None) diff --git a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/delimited_text_write_setting_py3.py b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/delimited_text_write_settings_py3.py similarity index 88% rename from sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/delimited_text_write_setting_py3.py rename to sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/delimited_text_write_settings_py3.py index ac0e3b2d00cc..2be019ab1e6a 100644 --- a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/delimited_text_write_setting_py3.py +++ b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/delimited_text_write_settings_py3.py @@ -9,10 +9,10 @@ # regenerated. # -------------------------------------------------------------------------- -from .format_write_setting_py3 import FormatWriteSetting +from .format_write_settings_py3 import FormatWriteSettings -class DelimitedTextWriteSetting(FormatWriteSetting): +class DelimitedTextWriteSettings(FormatWriteSettings): """Delimited text write settings. All required parameters must be populated in order to send to Azure. @@ -44,6 +44,6 @@ class DelimitedTextWriteSetting(FormatWriteSetting): } def __init__(self, *, type: str, file_extension, additional_properties=None, quote_all_text=None, **kwargs) -> None: - super(DelimitedTextWriteSetting, self).__init__(additional_properties=additional_properties, type=type, **kwargs) + super(DelimitedTextWriteSettings, self).__init__(additional_properties=additional_properties, type=type, **kwargs) self.quote_all_text = quote_all_text self.file_extension = file_extension diff --git a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/file_server_read_setting.py b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/file_server_read_settings.py similarity index 95% rename from sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/file_server_read_setting.py rename to sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/file_server_read_settings.py index 6ba2a5f56b79..da9d0809e03a 100644 --- a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/file_server_read_setting.py +++ b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/file_server_read_settings.py @@ -9,10 +9,10 @@ # regenerated. # -------------------------------------------------------------------------- -from .connector_read_setting import ConnectorReadSetting +from .store_read_settings import StoreReadSettings -class FileServerReadSetting(ConnectorReadSetting): +class FileServerReadSettings(StoreReadSettings): """File server read settings. All required parameters must be populated in order to send to Azure. @@ -64,7 +64,7 @@ class FileServerReadSetting(ConnectorReadSetting): } def __init__(self, **kwargs): - super(FileServerReadSetting, self).__init__(**kwargs) + super(FileServerReadSettings, self).__init__(**kwargs) self.recursive = kwargs.get('recursive', None) self.wildcard_folder_path = kwargs.get('wildcard_folder_path', None) self.wildcard_file_name = kwargs.get('wildcard_file_name', None) diff --git a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/file_server_read_setting_py3.py b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/file_server_read_settings_py3.py similarity index 92% rename from sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/file_server_read_setting_py3.py rename to sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/file_server_read_settings_py3.py index 4393692d63f3..1fadb49b1795 100644 --- a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/file_server_read_setting_py3.py +++ b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/file_server_read_settings_py3.py @@ -9,10 +9,10 @@ # regenerated. # -------------------------------------------------------------------------- -from .connector_read_setting_py3 import ConnectorReadSetting +from .store_read_settings_py3 import StoreReadSettings -class FileServerReadSetting(ConnectorReadSetting): +class FileServerReadSettings(StoreReadSettings): """File server read settings. All required parameters must be populated in order to send to Azure. @@ -64,7 +64,7 @@ class FileServerReadSetting(ConnectorReadSetting): } def __init__(self, *, type: str, additional_properties=None, max_concurrent_connections=None, recursive=None, wildcard_folder_path=None, wildcard_file_name=None, enable_partition_discovery: bool=None, modified_datetime_start=None, modified_datetime_end=None, **kwargs) -> None: - super(FileServerReadSetting, self).__init__(additional_properties=additional_properties, type=type, max_concurrent_connections=max_concurrent_connections, **kwargs) + super(FileServerReadSettings, self).__init__(additional_properties=additional_properties, type=type, max_concurrent_connections=max_concurrent_connections, **kwargs) self.recursive = recursive self.wildcard_folder_path = wildcard_folder_path self.wildcard_file_name = wildcard_file_name diff --git a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/file_server_write_setting.py b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/file_server_write_settings.py similarity index 89% rename from sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/file_server_write_setting.py rename to sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/file_server_write_settings.py index 9342210abdfb..f254e46452de 100644 --- a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/file_server_write_setting.py +++ b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/file_server_write_settings.py @@ -9,10 +9,10 @@ # regenerated. # -------------------------------------------------------------------------- -from .connector_write_setting import ConnectorWriteSetting +from .store_write_settings import StoreWriteSettings -class FileServerWriteSetting(ConnectorWriteSetting): +class FileServerWriteSettings(StoreWriteSettings): """File server write settings. All required parameters must be populated in order to send to Azure. @@ -42,4 +42,4 @@ class FileServerWriteSetting(ConnectorWriteSetting): } def __init__(self, **kwargs): - super(FileServerWriteSetting, self).__init__(**kwargs) + super(FileServerWriteSettings, self).__init__(**kwargs) diff --git a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_blob_fs_write_setting_py3.py b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/file_server_write_settings_py3.py similarity index 88% rename from sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_blob_fs_write_setting_py3.py rename to sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/file_server_write_settings_py3.py index 62196ff73838..26a48aca46f6 100644 --- a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_blob_fs_write_setting_py3.py +++ b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/file_server_write_settings_py3.py @@ -9,11 +9,11 @@ # regenerated. # -------------------------------------------------------------------------- -from .connector_write_setting_py3 import ConnectorWriteSetting +from .store_write_settings_py3 import StoreWriteSettings -class AzureBlobFSWriteSetting(ConnectorWriteSetting): - """Azure blobFS write settings. +class FileServerWriteSettings(StoreWriteSettings): + """File server write settings. All required parameters must be populated in order to send to Azure. @@ -42,4 +42,4 @@ class AzureBlobFSWriteSetting(ConnectorWriteSetting): } def __init__(self, *, type: str, additional_properties=None, max_concurrent_connections=None, copy_behavior=None, **kwargs) -> None: - super(AzureBlobFSWriteSetting, self).__init__(additional_properties=additional_properties, type=type, max_concurrent_connections=max_concurrent_connections, copy_behavior=copy_behavior, **kwargs) + super(FileServerWriteSettings, self).__init__(additional_properties=additional_properties, type=type, max_concurrent_connections=max_concurrent_connections, copy_behavior=copy_behavior, **kwargs) diff --git a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/format_read_setting.py b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/format_read_settings.py similarity index 93% rename from sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/format_read_setting.py rename to sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/format_read_settings.py index 730cec9f525f..d5213138b96a 100644 --- a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/format_read_setting.py +++ b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/format_read_settings.py @@ -12,7 +12,7 @@ from msrest.serialization import Model -class FormatReadSetting(Model): +class FormatReadSettings(Model): """Format read settings. All required parameters must be populated in order to send to Azure. @@ -34,6 +34,6 @@ class FormatReadSetting(Model): } def __init__(self, **kwargs): - super(FormatReadSetting, self).__init__(**kwargs) + super(FormatReadSettings, self).__init__(**kwargs) self.additional_properties = kwargs.get('additional_properties', None) self.type = kwargs.get('type', None) diff --git a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/format_read_setting_py3.py b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/format_read_settings_py3.py similarity index 93% rename from sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/format_read_setting_py3.py rename to sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/format_read_settings_py3.py index ed68bf35f009..326da0277b89 100644 --- a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/format_read_setting_py3.py +++ b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/format_read_settings_py3.py @@ -12,7 +12,7 @@ from msrest.serialization import Model -class FormatReadSetting(Model): +class FormatReadSettings(Model): """Format read settings. All required parameters must be populated in order to send to Azure. @@ -34,6 +34,6 @@ class FormatReadSetting(Model): } def __init__(self, *, type: str, additional_properties=None, **kwargs) -> None: - super(FormatReadSetting, self).__init__(**kwargs) + super(FormatReadSettings, self).__init__(**kwargs) self.additional_properties = additional_properties self.type = type diff --git a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/format_write_setting.py b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/format_write_settings.py similarity index 92% rename from sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/format_write_setting.py rename to sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/format_write_settings.py index 0fd6966859d5..2100c6055d0d 100644 --- a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/format_write_setting.py +++ b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/format_write_settings.py @@ -12,7 +12,7 @@ from msrest.serialization import Model -class FormatWriteSetting(Model): +class FormatWriteSettings(Model): """Format write settings. All required parameters must be populated in order to send to Azure. @@ -34,6 +34,6 @@ class FormatWriteSetting(Model): } def __init__(self, **kwargs): - super(FormatWriteSetting, self).__init__(**kwargs) + super(FormatWriteSettings, self).__init__(**kwargs) self.additional_properties = kwargs.get('additional_properties', None) self.type = kwargs.get('type', None) diff --git a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/format_write_setting_py3.py b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/format_write_settings_py3.py similarity index 92% rename from sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/format_write_setting_py3.py rename to sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/format_write_settings_py3.py index 3e5609066208..4150eceffc1c 100644 --- a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/format_write_setting_py3.py +++ b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/format_write_settings_py3.py @@ -12,7 +12,7 @@ from msrest.serialization import Model -class FormatWriteSetting(Model): +class FormatWriteSettings(Model): """Format write settings. All required parameters must be populated in order to send to Azure. @@ -34,6 +34,6 @@ class FormatWriteSetting(Model): } def __init__(self, *, type: str, additional_properties=None, **kwargs) -> None: - super(FormatWriteSetting, self).__init__(**kwargs) + super(FormatWriteSettings, self).__init__(**kwargs) self.additional_properties = additional_properties self.type = type diff --git a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/ftp_read_setting.py b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/ftp_read_settings.py similarity index 94% rename from sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/ftp_read_setting.py rename to sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/ftp_read_settings.py index 137a56948deb..e023f9ae91f7 100644 --- a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/ftp_read_setting.py +++ b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/ftp_read_settings.py @@ -9,10 +9,10 @@ # regenerated. # -------------------------------------------------------------------------- -from .connector_read_setting import ConnectorReadSetting +from .store_read_settings import StoreReadSettings -class FtpReadSetting(ConnectorReadSetting): +class FtpReadSettings(StoreReadSettings): """Ftp read settings. All required parameters must be populated in order to send to Azure. @@ -56,7 +56,7 @@ class FtpReadSetting(ConnectorReadSetting): } def __init__(self, **kwargs): - super(FtpReadSetting, self).__init__(**kwargs) + super(FtpReadSettings, self).__init__(**kwargs) self.recursive = kwargs.get('recursive', None) self.wildcard_folder_path = kwargs.get('wildcard_folder_path', None) self.wildcard_file_name = kwargs.get('wildcard_file_name', None) diff --git a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/ftp_read_setting_py3.py b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/ftp_read_settings_py3.py similarity index 90% rename from sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/ftp_read_setting_py3.py rename to sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/ftp_read_settings_py3.py index 5294301e4fd8..748d306307ac 100644 --- a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/ftp_read_setting_py3.py +++ b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/ftp_read_settings_py3.py @@ -9,10 +9,10 @@ # regenerated. # -------------------------------------------------------------------------- -from .connector_read_setting_py3 import ConnectorReadSetting +from .store_read_settings_py3 import StoreReadSettings -class FtpReadSetting(ConnectorReadSetting): +class FtpReadSettings(StoreReadSettings): """Ftp read settings. All required parameters must be populated in order to send to Azure. @@ -56,7 +56,7 @@ class FtpReadSetting(ConnectorReadSetting): } def __init__(self, *, type: str, additional_properties=None, max_concurrent_connections=None, recursive=None, wildcard_folder_path=None, wildcard_file_name=None, use_binary_transfer: bool=None, **kwargs) -> None: - super(FtpReadSetting, self).__init__(additional_properties=additional_properties, type=type, max_concurrent_connections=max_concurrent_connections, **kwargs) + super(FtpReadSettings, self).__init__(additional_properties=additional_properties, type=type, max_concurrent_connections=max_concurrent_connections, **kwargs) self.recursive = recursive self.wildcard_folder_path = wildcard_folder_path self.wildcard_file_name = wildcard_file_name diff --git a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hdfs_read_setting.py b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hdfs_read_settings.py similarity index 95% rename from sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hdfs_read_setting.py rename to sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hdfs_read_settings.py index 4fdadbc2fcd0..ec4b98c50385 100644 --- a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hdfs_read_setting.py +++ b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hdfs_read_settings.py @@ -9,10 +9,10 @@ # regenerated. # -------------------------------------------------------------------------- -from .connector_read_setting import ConnectorReadSetting +from .store_read_settings import StoreReadSettings -class HdfsReadSetting(ConnectorReadSetting): +class HdfsReadSettings(StoreReadSettings): """HDFS read settings. All required parameters must be populated in order to send to Azure. @@ -67,7 +67,7 @@ class HdfsReadSetting(ConnectorReadSetting): } def __init__(self, **kwargs): - super(HdfsReadSetting, self).__init__(**kwargs) + super(HdfsReadSettings, self).__init__(**kwargs) self.recursive = kwargs.get('recursive', None) self.wildcard_folder_path = kwargs.get('wildcard_folder_path', None) self.wildcard_file_name = kwargs.get('wildcard_file_name', None) diff --git a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hdfs_read_setting_py3.py b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hdfs_read_settings_py3.py similarity index 93% rename from sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hdfs_read_setting_py3.py rename to sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hdfs_read_settings_py3.py index 164a6f497e52..c37a045ec93c 100644 --- a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hdfs_read_setting_py3.py +++ b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hdfs_read_settings_py3.py @@ -9,10 +9,10 @@ # regenerated. # -------------------------------------------------------------------------- -from .connector_read_setting_py3 import ConnectorReadSetting +from .store_read_settings_py3 import StoreReadSettings -class HdfsReadSetting(ConnectorReadSetting): +class HdfsReadSettings(StoreReadSettings): """HDFS read settings. All required parameters must be populated in order to send to Azure. @@ -67,7 +67,7 @@ class HdfsReadSetting(ConnectorReadSetting): } def __init__(self, *, type: str, additional_properties=None, max_concurrent_connections=None, recursive=None, wildcard_folder_path=None, wildcard_file_name=None, enable_partition_discovery: bool=None, modified_datetime_start=None, modified_datetime_end=None, distcp_settings=None, **kwargs) -> None: - super(HdfsReadSetting, self).__init__(additional_properties=additional_properties, type=type, max_concurrent_connections=max_concurrent_connections, **kwargs) + super(HdfsReadSettings, self).__init__(additional_properties=additional_properties, type=type, max_concurrent_connections=max_concurrent_connections, **kwargs) self.recursive = recursive self.wildcard_folder_path = wildcard_folder_path self.wildcard_file_name = wildcard_file_name diff --git a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/http_read_setting.py b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/http_read_settings.py similarity index 94% rename from sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/http_read_setting.py rename to sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/http_read_settings.py index 696a9fdb3faf..a7c175da3489 100644 --- a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/http_read_setting.py +++ b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/http_read_settings.py @@ -9,10 +9,10 @@ # regenerated. # -------------------------------------------------------------------------- -from .connector_read_setting import ConnectorReadSetting +from .store_read_settings import StoreReadSettings -class HttpReadSetting(ConnectorReadSetting): +class HttpReadSettings(StoreReadSettings): """Sftp read settings. All required parameters must be populated in order to send to Azure. @@ -56,7 +56,7 @@ class HttpReadSetting(ConnectorReadSetting): } def __init__(self, **kwargs): - super(HttpReadSetting, self).__init__(**kwargs) + super(HttpReadSettings, self).__init__(**kwargs) self.request_method = kwargs.get('request_method', None) self.request_body = kwargs.get('request_body', None) self.additional_headers = kwargs.get('additional_headers', None) diff --git a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/http_read_setting_py3.py b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/http_read_settings_py3.py similarity index 90% rename from sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/http_read_setting_py3.py rename to sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/http_read_settings_py3.py index 3d5d75a80785..7cea9207c996 100644 --- a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/http_read_setting_py3.py +++ b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/http_read_settings_py3.py @@ -9,10 +9,10 @@ # regenerated. # -------------------------------------------------------------------------- -from .connector_read_setting_py3 import ConnectorReadSetting +from .store_read_settings_py3 import StoreReadSettings -class HttpReadSetting(ConnectorReadSetting): +class HttpReadSettings(StoreReadSettings): """Sftp read settings. All required parameters must be populated in order to send to Azure. @@ -56,7 +56,7 @@ class HttpReadSetting(ConnectorReadSetting): } def __init__(self, *, type: str, additional_properties=None, max_concurrent_connections=None, request_method=None, request_body=None, additional_headers=None, request_timeout=None, **kwargs) -> None: - super(HttpReadSetting, self).__init__(additional_properties=additional_properties, type=type, max_concurrent_connections=max_concurrent_connections, **kwargs) + super(HttpReadSettings, self).__init__(additional_properties=additional_properties, type=type, max_concurrent_connections=max_concurrent_connections, **kwargs) self.request_method = request_method self.request_body = request_body self.additional_headers = additional_headers diff --git a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/parquet_sink.py b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/parquet_sink.py index 38c634ed10dd..dea3e0f8fc52 100644 --- a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/parquet_sink.py +++ b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/parquet_sink.py @@ -41,7 +41,7 @@ class ParquetSink(CopySink): :param type: Required. Constant filled by server. :type type: str :param store_settings: Parquet store settings. - :type store_settings: ~azure.mgmt.datafactory.models.ConnectorWriteSetting + :type store_settings: ~azure.mgmt.datafactory.models.StoreWriteSettings """ _validation = { @@ -56,7 +56,7 @@ class ParquetSink(CopySink): 'sink_retry_wait': {'key': 'sinkRetryWait', 'type': 'object'}, 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'type': {'key': 'type', 'type': 'str'}, - 'store_settings': {'key': 'storeSettings', 'type': 'ConnectorWriteSetting'}, + 'store_settings': {'key': 'storeSettings', 'type': 'StoreWriteSettings'}, } def __init__(self, **kwargs): diff --git a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/parquet_sink_py3.py b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/parquet_sink_py3.py index 96c0c1b57926..463044fef83f 100644 --- a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/parquet_sink_py3.py +++ b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/parquet_sink_py3.py @@ -41,7 +41,7 @@ class ParquetSink(CopySink): :param type: Required. Constant filled by server. :type type: str :param store_settings: Parquet store settings. - :type store_settings: ~azure.mgmt.datafactory.models.ConnectorWriteSetting + :type store_settings: ~azure.mgmt.datafactory.models.StoreWriteSettings """ _validation = { @@ -56,7 +56,7 @@ class ParquetSink(CopySink): 'sink_retry_wait': {'key': 'sinkRetryWait', 'type': 'object'}, 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'type': {'key': 'type', 'type': 'str'}, - 'store_settings': {'key': 'storeSettings', 'type': 'ConnectorWriteSetting'}, + 'store_settings': {'key': 'storeSettings', 'type': 'StoreWriteSettings'}, } def __init__(self, *, additional_properties=None, write_batch_size=None, write_batch_timeout=None, sink_retry_count=None, sink_retry_wait=None, max_concurrent_connections=None, store_settings=None, **kwargs) -> None: diff --git a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/parquet_source.py b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/parquet_source.py index 02e74641d506..ab888c7361a2 100644 --- a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/parquet_source.py +++ b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/parquet_source.py @@ -34,7 +34,7 @@ class ParquetSource(CopySource): :param type: Required. Constant filled by server. :type type: str :param store_settings: Parquet store settings. - :type store_settings: ~azure.mgmt.datafactory.models.ConnectorReadSetting + :type store_settings: ~azure.mgmt.datafactory.models.StoreReadSettings """ _validation = { @@ -47,7 +47,7 @@ class ParquetSource(CopySource): 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'type': {'key': 'type', 'type': 'str'}, - 'store_settings': {'key': 'storeSettings', 'type': 'ConnectorReadSetting'}, + 'store_settings': {'key': 'storeSettings', 'type': 'StoreReadSettings'}, } def __init__(self, **kwargs): diff --git a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/parquet_source_py3.py b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/parquet_source_py3.py index bfe077dd9999..332a7b9b8c5e 100644 --- a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/parquet_source_py3.py +++ b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/parquet_source_py3.py @@ -34,7 +34,7 @@ class ParquetSource(CopySource): :param type: Required. Constant filled by server. :type type: str :param store_settings: Parquet store settings. - :type store_settings: ~azure.mgmt.datafactory.models.ConnectorReadSetting + :type store_settings: ~azure.mgmt.datafactory.models.StoreReadSettings """ _validation = { @@ -47,7 +47,7 @@ class ParquetSource(CopySource): 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'type': {'key': 'type', 'type': 'str'}, - 'store_settings': {'key': 'storeSettings', 'type': 'ConnectorReadSetting'}, + 'store_settings': {'key': 'storeSettings', 'type': 'StoreReadSettings'}, } def __init__(self, *, additional_properties=None, source_retry_count=None, source_retry_wait=None, max_concurrent_connections=None, store_settings=None, **kwargs) -> None: diff --git a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sftp_read_setting.py b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sftp_read_settings.py similarity index 94% rename from sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sftp_read_setting.py rename to sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sftp_read_settings.py index e0cd7ea8fda1..5e7b4faf77ad 100644 --- a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sftp_read_setting.py +++ b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sftp_read_settings.py @@ -9,10 +9,10 @@ # regenerated. # -------------------------------------------------------------------------- -from .connector_read_setting import ConnectorReadSetting +from .store_read_settings import StoreReadSettings -class SftpReadSetting(ConnectorReadSetting): +class SftpReadSettings(StoreReadSettings): """Sftp read settings. All required parameters must be populated in order to send to Azure. @@ -60,7 +60,7 @@ class SftpReadSetting(ConnectorReadSetting): } def __init__(self, **kwargs): - super(SftpReadSetting, self).__init__(**kwargs) + super(SftpReadSettings, self).__init__(**kwargs) self.recursive = kwargs.get('recursive', None) self.wildcard_folder_path = kwargs.get('wildcard_folder_path', None) self.wildcard_file_name = kwargs.get('wildcard_file_name', None) diff --git a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sftp_read_setting_py3.py b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sftp_read_settings_py3.py similarity index 91% rename from sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sftp_read_setting_py3.py rename to sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sftp_read_settings_py3.py index 39beb756905a..e6c27e3ad08a 100644 --- a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sftp_read_setting_py3.py +++ b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sftp_read_settings_py3.py @@ -9,10 +9,10 @@ # regenerated. # -------------------------------------------------------------------------- -from .connector_read_setting_py3 import ConnectorReadSetting +from .store_read_settings_py3 import StoreReadSettings -class SftpReadSetting(ConnectorReadSetting): +class SftpReadSettings(StoreReadSettings): """Sftp read settings. All required parameters must be populated in order to send to Azure. @@ -60,7 +60,7 @@ class SftpReadSetting(ConnectorReadSetting): } def __init__(self, *, type: str, additional_properties=None, max_concurrent_connections=None, recursive=None, wildcard_folder_path=None, wildcard_file_name=None, modified_datetime_start=None, modified_datetime_end=None, **kwargs) -> None: - super(SftpReadSetting, self).__init__(additional_properties=additional_properties, type=type, max_concurrent_connections=max_concurrent_connections, **kwargs) + super(SftpReadSettings, self).__init__(additional_properties=additional_properties, type=type, max_concurrent_connections=max_concurrent_connections, **kwargs) self.recursive = recursive self.wildcard_folder_path = wildcard_folder_path self.wildcard_file_name = wildcard_file_name diff --git a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/connector_read_setting.py b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/store_read_settings.py similarity index 94% rename from sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/connector_read_setting.py rename to sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/store_read_settings.py index e0af269aaafd..c12c0ce8860d 100644 --- a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/connector_read_setting.py +++ b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/store_read_settings.py @@ -12,7 +12,7 @@ from msrest.serialization import Model -class ConnectorReadSetting(Model): +class StoreReadSettings(Model): """Connector read setting. All required parameters must be populated in order to send to Azure. @@ -39,7 +39,7 @@ class ConnectorReadSetting(Model): } def __init__(self, **kwargs): - super(ConnectorReadSetting, self).__init__(**kwargs) + super(StoreReadSettings, self).__init__(**kwargs) self.additional_properties = kwargs.get('additional_properties', None) self.type = kwargs.get('type', None) self.max_concurrent_connections = kwargs.get('max_concurrent_connections', None) diff --git a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/connector_read_setting_py3.py b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/store_read_settings_py3.py similarity index 94% rename from sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/connector_read_setting_py3.py rename to sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/store_read_settings_py3.py index f6403ade8f71..e2026fd52b93 100644 --- a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/connector_read_setting_py3.py +++ b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/store_read_settings_py3.py @@ -12,7 +12,7 @@ from msrest.serialization import Model -class ConnectorReadSetting(Model): +class StoreReadSettings(Model): """Connector read setting. All required parameters must be populated in order to send to Azure. @@ -39,7 +39,7 @@ class ConnectorReadSetting(Model): } def __init__(self, *, type: str, additional_properties=None, max_concurrent_connections=None, **kwargs) -> None: - super(ConnectorReadSetting, self).__init__(**kwargs) + super(StoreReadSettings, self).__init__(**kwargs) self.additional_properties = additional_properties self.type = type self.max_concurrent_connections = max_concurrent_connections diff --git a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/connector_write_setting.py b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/store_write_settings.py similarity index 94% rename from sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/connector_write_setting.py rename to sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/store_write_settings.py index 65daf9f07794..aeaebc1a190e 100644 --- a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/connector_write_setting.py +++ b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/store_write_settings.py @@ -12,7 +12,7 @@ from msrest.serialization import Model -class ConnectorWriteSetting(Model): +class StoreWriteSettings(Model): """Connector write settings. All required parameters must be populated in order to send to Azure. @@ -42,7 +42,7 @@ class ConnectorWriteSetting(Model): } def __init__(self, **kwargs): - super(ConnectorWriteSetting, self).__init__(**kwargs) + super(StoreWriteSettings, self).__init__(**kwargs) self.additional_properties = kwargs.get('additional_properties', None) self.type = kwargs.get('type', None) self.max_concurrent_connections = kwargs.get('max_concurrent_connections', None) diff --git a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/connector_write_setting_py3.py b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/store_write_settings_py3.py similarity index 95% rename from sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/connector_write_setting_py3.py rename to sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/store_write_settings_py3.py index 7f4ea65c916d..d3dfa76faddc 100644 --- a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/connector_write_setting_py3.py +++ b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/store_write_settings_py3.py @@ -12,7 +12,7 @@ from msrest.serialization import Model -class ConnectorWriteSetting(Model): +class StoreWriteSettings(Model): """Connector write settings. All required parameters must be populated in order to send to Azure. @@ -42,7 +42,7 @@ class ConnectorWriteSetting(Model): } def __init__(self, *, type: str, additional_properties=None, max_concurrent_connections=None, copy_behavior=None, **kwargs) -> None: - super(ConnectorWriteSetting, self).__init__(**kwargs) + super(StoreWriteSettings, self).__init__(**kwargs) self.additional_properties = additional_properties self.type = type self.max_concurrent_connections = max_concurrent_connections