From 068c952ea1413289d68ccba8877ac4382ccb6670 Mon Sep 17 00:00:00 2001 From: SDKAuto Date: Thu, 19 Aug 2021 07:25:16 +0000 Subject: [PATCH] CodeGen from PR 15711 in Azure/azure-rest-api-specs Merge 9778e33994dbda11e3bd6fc8eea86ed88d7d6055 into 30d1537f0c9aa49e6e04401b1d16f612b31231e7 --- .../azure-mgmt-datafactory/_meta.json | 2 +- .../azure/mgmt/datafactory/_version.py | 2 +- .../_integration_runtimes_operations.py | 62 + .../azure/mgmt/datafactory/models/__init__.py | 40 + .../_data_factory_management_client_enums.py | 15 + .../azure/mgmt/datafactory/models/_models.py | 1099 +++++++++---- .../mgmt/datafactory/models/_models_py3.py | 1354 ++++++++++++----- .../_integration_runtimes_operations.py | 63 + 8 files changed, 2004 insertions(+), 633 deletions(-) diff --git a/sdk/datafactory/azure-mgmt-datafactory/_meta.json b/sdk/datafactory/azure-mgmt-datafactory/_meta.json index ca4e1b09611d..c73377593ffa 100644 --- a/sdk/datafactory/azure-mgmt-datafactory/_meta.json +++ b/sdk/datafactory/azure-mgmt-datafactory/_meta.json @@ -4,7 +4,7 @@ "@autorest/python@5.8.4", "@autorest/modelerfour@4.19.2" ], - "commit": "406474c3807f2dec010af72286f22aa7a9a54920", + "commit": "66612e5cd2c83acc5ceb912de0e8dac537934833", "repository_url": "https://github.com/Azure/azure-rest-api-specs", "autorest_command": "autorest specification/datafactory/resource-manager/readme.md --multiapi --python --python-mode=update --python-sdks-folder=/home/vsts/work/1/s/azure-sdk-for-python/sdk --track2 --use=@autorest/python@5.8.4 --use=@autorest/modelerfour@4.19.2 --version=3.4.5", "readme": "specification/datafactory/resource-manager/readme.md" diff --git a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/_version.py b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/_version.py index 48944bf3938a..c47f66669f1b 100644 --- a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/_version.py +++ b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/_version.py @@ -6,4 +6,4 @@ # Changes may cause incorrect behavior and will be lost if the code is regenerated. # -------------------------------------------------------------------------- -VERSION = "2.0.0" +VERSION = "1.0.0" diff --git a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/aio/operations/_integration_runtimes_operations.py b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/aio/operations/_integration_runtimes_operations.py index cfe891bfaed7..19a13bbaf567 100644 --- a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/aio/operations/_integration_runtimes_operations.py +++ b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/aio/operations/_integration_runtimes_operations.py @@ -455,6 +455,68 @@ async def get_status( return deserialized get_status.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.DataFactory/factories/{factoryName}/integrationRuntimes/{integrationRuntimeName}/getStatus'} # type: ignore + async def list_outbound_network_dependencies_endpoints( + self, + resource_group_name: str, + factory_name: str, + integration_runtime_name: str, + **kwargs: Any + ) -> "_models.IntegrationRuntimeOutboundNetworkDependenciesEndpointsResponse": + """Gets the list of outbound network dependencies for a given Azure-SSIS integration runtime. + + :param resource_group_name: The resource group name. + :type resource_group_name: str + :param factory_name: The factory name. + :type factory_name: str + :param integration_runtime_name: The integration runtime name. + :type integration_runtime_name: str + :keyword callable cls: A custom type or function that will be passed the direct response + :return: IntegrationRuntimeOutboundNetworkDependenciesEndpointsResponse, or the result of cls(response) + :rtype: ~azure.mgmt.datafactory.models.IntegrationRuntimeOutboundNetworkDependenciesEndpointsResponse + :raises: ~azure.core.exceptions.HttpResponseError + """ + cls = kwargs.pop('cls', None) # type: ClsType["_models.IntegrationRuntimeOutboundNetworkDependenciesEndpointsResponse"] + error_map = { + 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError + } + error_map.update(kwargs.pop('error_map', {})) + api_version = "2018-06-01" + accept = "application/json" + + # Construct URL + url = self.list_outbound_network_dependencies_endpoints.metadata['url'] # type: ignore + path_format_arguments = { + 'subscriptionId': self._serialize.url("self._config.subscription_id", self._config.subscription_id, 'str'), + 'resourceGroupName': self._serialize.url("resource_group_name", resource_group_name, 'str', max_length=90, min_length=1, pattern=r'^[-\w\._\(\)]+$'), + 'factoryName': self._serialize.url("factory_name", factory_name, 'str', max_length=63, min_length=3, pattern=r'^[A-Za-z0-9]+(?:-[A-Za-z0-9]+)*$'), + 'integrationRuntimeName': self._serialize.url("integration_runtime_name", integration_runtime_name, 'str', max_length=63, min_length=3, pattern=r'^[A-Za-z0-9]+(?:-[A-Za-z0-9]+)*$'), + } + url = self._client.format_url(url, **path_format_arguments) + + # Construct parameters + query_parameters = {} # type: Dict[str, Any] + query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') + + # Construct headers + header_parameters = {} # type: Dict[str, Any] + header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') + + request = self._client.get(url, query_parameters, header_parameters) + pipeline_response = await self._client._pipeline.run(request, stream=False, **kwargs) + response = pipeline_response.http_response + + if response.status_code not in [200]: + map_error(status_code=response.status_code, response=response, error_map=error_map) + raise HttpResponseError(response=response, error_format=ARMErrorFormat) + + deserialized = self._deserialize('IntegrationRuntimeOutboundNetworkDependenciesEndpointsResponse', pipeline_response) + + if cls: + return cls(pipeline_response, deserialized, {}) + + return deserialized + list_outbound_network_dependencies_endpoints.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.DataFactory/factories/{factoryName}/integrationRuntimes/{integrationRuntimeName}/outboundNetworkDependenciesEndpoints'} # type: ignore + async def get_connection_info( self, resource_group_name: str, diff --git a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py index ff60f03feff6..2c86779466ed 100644 --- a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py +++ b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py @@ -243,9 +243,12 @@ from ._models_py3 import ExcelDataset from ._models_py3 import ExcelSource from ._models_py3 import ExecuteDataFlowActivity + from ._models_py3 import ExecuteDataFlowActivityTypeProperties from ._models_py3 import ExecuteDataFlowActivityTypePropertiesCompute from ._models_py3 import ExecutePipelineActivity + from ._models_py3 import ExecutePowerQueryActivityTypeProperties from ._models_py3 import ExecuteSSISPackageActivity + from ._models_py3 import ExecuteWranglingDataflowActivity from ._models_py3 import ExecutionActivity from ._models_py3 import ExportSettings from ._models_py3 import ExposureControlBatchRequest @@ -340,6 +343,10 @@ from ._models_py3 import IntegrationRuntimeMonitoringData from ._models_py3 import IntegrationRuntimeNodeIpAddress from ._models_py3 import IntegrationRuntimeNodeMonitoringData + from ._models_py3 import IntegrationRuntimeOutboundNetworkDependenciesCategoryEndpoint + from ._models_py3 import IntegrationRuntimeOutboundNetworkDependenciesEndpoint + from ._models_py3 import IntegrationRuntimeOutboundNetworkDependenciesEndpointDetails + from ._models_py3 import IntegrationRuntimeOutboundNetworkDependenciesEndpointsResponse from ._models_py3 import IntegrationRuntimeReference from ._models_py3 import IntegrationRuntimeRegenerateKeyParameters from ._models_py3 import IntegrationRuntimeResource @@ -480,6 +487,8 @@ from ._models_py3 import PostgreSqlLinkedService from ._models_py3 import PostgreSqlSource from ._models_py3 import PostgreSqlTableDataset + from ._models_py3 import PowerQuerySink + from ._models_py3 import PowerQuerySource from ._models_py3 import PrestoLinkedService from ._models_py3 import PrestoObjectDataset from ._models_py3 import PrestoSource @@ -591,6 +600,7 @@ from ._models_py3 import SqlAlwaysEncryptedProperties from ._models_py3 import SqlDWSink from ._models_py3 import SqlDWSource + from ._models_py3 import SqlDWUpsertSettings from ._models_py3 import SqlMISink from ._models_py3 import SqlMISource from ._models_py3 import SqlPartitionSettings @@ -601,6 +611,7 @@ from ._models_py3 import SqlServerTableDataset from ._models_py3 import SqlSink from ._models_py3 import SqlSource + from ._models_py3 import SqlUpsertSettings from ._models_py3 import SquareLinkedService from ._models_py3 import SquareObjectDataset from ._models_py3 import SquareSource @@ -670,6 +681,7 @@ from ._models_py3 import WebLinkedServiceTypeProperties from ._models_py3 import WebSource from ._models_py3 import WebTableDataset + from ._models_py3 import WranglingDataFlow from ._models_py3 import XeroLinkedService from ._models_py3 import XeroObjectDataset from ._models_py3 import XeroSource @@ -917,9 +929,12 @@ from ._models import ExcelDataset # type: ignore from ._models import ExcelSource # type: ignore from ._models import ExecuteDataFlowActivity # type: ignore + from ._models import ExecuteDataFlowActivityTypeProperties # type: ignore from ._models import ExecuteDataFlowActivityTypePropertiesCompute # type: ignore from ._models import ExecutePipelineActivity # type: ignore + from ._models import ExecutePowerQueryActivityTypeProperties # type: ignore from ._models import ExecuteSSISPackageActivity # type: ignore + from ._models import ExecuteWranglingDataflowActivity # type: ignore from ._models import ExecutionActivity # type: ignore from ._models import ExportSettings # type: ignore from ._models import ExposureControlBatchRequest # type: ignore @@ -1014,6 +1029,10 @@ from ._models import IntegrationRuntimeMonitoringData # type: ignore from ._models import IntegrationRuntimeNodeIpAddress # type: ignore from ._models import IntegrationRuntimeNodeMonitoringData # type: ignore + from ._models import IntegrationRuntimeOutboundNetworkDependenciesCategoryEndpoint # type: ignore + from ._models import IntegrationRuntimeOutboundNetworkDependenciesEndpoint # type: ignore + from ._models import IntegrationRuntimeOutboundNetworkDependenciesEndpointDetails # type: ignore + from ._models import IntegrationRuntimeOutboundNetworkDependenciesEndpointsResponse # type: ignore from ._models import IntegrationRuntimeReference # type: ignore from ._models import IntegrationRuntimeRegenerateKeyParameters # type: ignore from ._models import IntegrationRuntimeResource # type: ignore @@ -1154,6 +1173,8 @@ from ._models import PostgreSqlLinkedService # type: ignore from ._models import PostgreSqlSource # type: ignore from ._models import PostgreSqlTableDataset # type: ignore + from ._models import PowerQuerySink # type: ignore + from ._models import PowerQuerySource # type: ignore from ._models import PrestoLinkedService # type: ignore from ._models import PrestoObjectDataset # type: ignore from ._models import PrestoSource # type: ignore @@ -1265,6 +1286,7 @@ from ._models import SqlAlwaysEncryptedProperties # type: ignore from ._models import SqlDWSink # type: ignore from ._models import SqlDWSource # type: ignore + from ._models import SqlDWUpsertSettings # type: ignore from ._models import SqlMISink # type: ignore from ._models import SqlMISource # type: ignore from ._models import SqlPartitionSettings # type: ignore @@ -1275,6 +1297,7 @@ from ._models import SqlServerTableDataset # type: ignore from ._models import SqlSink # type: ignore from ._models import SqlSource # type: ignore + from ._models import SqlUpsertSettings # type: ignore from ._models import SquareLinkedService # type: ignore from ._models import SquareObjectDataset # type: ignore from ._models import SquareSource # type: ignore @@ -1344,6 +1367,7 @@ from ._models import WebLinkedServiceTypeProperties # type: ignore from ._models import WebSource # type: ignore from ._models import WebTableDataset # type: ignore + from ._models import WranglingDataFlow # type: ignore from ._models import XeroLinkedService # type: ignore from ._models import XeroObjectDataset # type: ignore from ._models import XeroSource # type: ignore @@ -1433,7 +1457,9 @@ SparkServerType, SparkThriftTransportProtocol, SqlAlwaysEncryptedAkvAuthType, + SqlDWWriteBehaviorEnum, SqlPartitionOption, + SqlWriteBehaviorEnum, SsisLogLocationType, SsisObjectMetadataType, SsisPackageLocationType, @@ -1687,9 +1713,12 @@ 'ExcelDataset', 'ExcelSource', 'ExecuteDataFlowActivity', + 'ExecuteDataFlowActivityTypeProperties', 'ExecuteDataFlowActivityTypePropertiesCompute', 'ExecutePipelineActivity', + 'ExecutePowerQueryActivityTypeProperties', 'ExecuteSSISPackageActivity', + 'ExecuteWranglingDataflowActivity', 'ExecutionActivity', 'ExportSettings', 'ExposureControlBatchRequest', @@ -1784,6 +1813,10 @@ 'IntegrationRuntimeMonitoringData', 'IntegrationRuntimeNodeIpAddress', 'IntegrationRuntimeNodeMonitoringData', + 'IntegrationRuntimeOutboundNetworkDependenciesCategoryEndpoint', + 'IntegrationRuntimeOutboundNetworkDependenciesEndpoint', + 'IntegrationRuntimeOutboundNetworkDependenciesEndpointDetails', + 'IntegrationRuntimeOutboundNetworkDependenciesEndpointsResponse', 'IntegrationRuntimeReference', 'IntegrationRuntimeRegenerateKeyParameters', 'IntegrationRuntimeResource', @@ -1924,6 +1957,8 @@ 'PostgreSqlLinkedService', 'PostgreSqlSource', 'PostgreSqlTableDataset', + 'PowerQuerySink', + 'PowerQuerySource', 'PrestoLinkedService', 'PrestoObjectDataset', 'PrestoSource', @@ -2035,6 +2070,7 @@ 'SqlAlwaysEncryptedProperties', 'SqlDWSink', 'SqlDWSource', + 'SqlDWUpsertSettings', 'SqlMISink', 'SqlMISource', 'SqlPartitionSettings', @@ -2045,6 +2081,7 @@ 'SqlServerTableDataset', 'SqlSink', 'SqlSource', + 'SqlUpsertSettings', 'SquareLinkedService', 'SquareObjectDataset', 'SquareSource', @@ -2114,6 +2151,7 @@ 'WebLinkedServiceTypeProperties', 'WebSource', 'WebTableDataset', + 'WranglingDataFlow', 'XeroLinkedService', 'XeroObjectDataset', 'XeroSource', @@ -2201,7 +2239,9 @@ 'SparkServerType', 'SparkThriftTransportProtocol', 'SqlAlwaysEncryptedAkvAuthType', + 'SqlDWWriteBehaviorEnum', 'SqlPartitionOption', + 'SqlWriteBehaviorEnum', 'SsisLogLocationType', 'SsisObjectMetadataType', 'SsisPackageLocationType', diff --git a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/_data_factory_management_client_enums.py b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/_data_factory_management_client_enums.py index 466aea7a8f74..a8e3dd30f845 100644 --- a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/_data_factory_management_client_enums.py +++ b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/_data_factory_management_client_enums.py @@ -697,6 +697,13 @@ class SqlAlwaysEncryptedAkvAuthType(with_metaclass(_CaseInsensitiveEnumMeta, str SERVICE_PRINCIPAL = "ServicePrincipal" MANAGED_IDENTITY = "ManagedIdentity" +class SqlDWWriteBehaviorEnum(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): + """Specify the write behavior when copying data into sql dw. + """ + + INSERT = "Insert" + UPSERT = "Upsert" + class SqlPartitionOption(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): """The partition mechanism that will be used for Sql read in parallel. """ @@ -705,6 +712,14 @@ class SqlPartitionOption(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): PHYSICAL_PARTITIONS_OF_TABLE = "PhysicalPartitionsOfTable" DYNAMIC_RANGE = "DynamicRange" +class SqlWriteBehaviorEnum(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): + """Specify the write behavior when copying data into sql. + """ + + INSERT = "Insert" + UPSERT = "Upsert" + STORED_PROCEDURE = "StoredProcedure" + class SsisLogLocationType(with_metaclass(_CaseInsensitiveEnumMeta, str, Enum)): """The type of SSIS log location. """ diff --git a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/_models.py b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/_models.py index 98d8bc411749..b93675ab7fd8 100644 --- a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/_models.py +++ b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/_models.py @@ -40,7 +40,7 @@ class Activity(msrest.serialization.Model): """A pipeline activity. You probably want to use the sub-classes and not this class directly. Known - sub-classes are: ControlActivity, ExecutionActivity. + sub-classes are: ControlActivity, ExecuteWranglingDataflowActivity, ExecutionActivity. All required parameters must be populated in order to send to Azure. @@ -74,7 +74,7 @@ class Activity(msrest.serialization.Model): } _subtype_map = { - 'type': {'Container': 'ControlActivity', 'Execution': 'ExecutionActivity'} + 'type': {'Container': 'ControlActivity', 'ExecuteWranglingDataflow': 'ExecuteWranglingDataflowActivity', 'Execution': 'ExecutionActivity'} } def __init__( @@ -697,8 +697,8 @@ class TabularSource(CopySource): pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type query_timeout: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any """ _validation = { @@ -713,7 +713,7 @@ class TabularSource(CopySource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, } _subtype_map = { @@ -756,8 +756,8 @@ class AmazonMWSSource(TabularSource): pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type query_timeout: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any :param query: A query to retrieve data from source. Type: string (or Expression with resultType string). :type query: any @@ -775,7 +775,7 @@ class AmazonMWSSource(TabularSource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, 'query': {'key': 'query', 'type': 'object'}, } @@ -887,8 +887,8 @@ class AmazonRedshiftSource(TabularSource): pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type query_timeout: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any :param query: Database query. Type: string (or Expression with resultType string). :type query: any :param redshift_unload_settings: The Amazon S3 settings needed for the interim Amazon S3 when @@ -909,7 +909,7 @@ class AmazonRedshiftSource(TabularSource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, 'query': {'key': 'query', 'type': 'object'}, 'redshift_unload_settings': {'key': 'redshiftUnloadSettings', 'type': 'RedshiftUnloadSettings'}, } @@ -2014,8 +2014,8 @@ class AvroSource(CopySource): :param store_settings: Avro store settings. :type store_settings: ~azure.mgmt.datafactory.models.StoreReadSettings :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any """ _validation = { @@ -2030,7 +2030,7 @@ class AvroSource(CopySource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'store_settings': {'key': 'storeSettings', 'type': 'StoreReadSettings'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, } def __init__( @@ -3959,8 +3959,8 @@ class AzureDataExplorerSource(CopySource): pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])).. :type query_timeout: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any """ _validation = { @@ -3978,7 +3978,7 @@ class AzureDataExplorerSource(CopySource): 'query': {'key': 'query', 'type': 'object'}, 'no_truncation': {'key': 'noTruncation', 'type': 'object'}, 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, } def __init__( @@ -5189,8 +5189,8 @@ class AzureMariaDBSource(TabularSource): pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type query_timeout: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any :param query: A query to retrieve data from source. Type: string (or Expression with resultType string). :type query: any @@ -5208,7 +5208,7 @@ class AzureMariaDBSource(TabularSource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, 'query': {'key': 'query', 'type': 'object'}, } @@ -5848,8 +5848,8 @@ class AzureMySqlSource(TabularSource): pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type query_timeout: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any :param query: Database query. Type: string (or Expression with resultType string). :type query: any """ @@ -5866,7 +5866,7 @@ class AzureMySqlSource(TabularSource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, 'query': {'key': 'query', 'type': 'object'}, } @@ -6083,8 +6083,8 @@ class AzurePostgreSqlSource(TabularSource): pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type query_timeout: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any :param query: A query to retrieve data from source. Type: string (or Expression with resultType string). :type query: any @@ -6102,7 +6102,7 @@ class AzurePostgreSqlSource(TabularSource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, 'query': {'key': 'query', 'type': 'object'}, } @@ -6850,6 +6850,14 @@ class AzureSqlSink(CopySink): :param table_option: The option to handle sink table, such as autoCreate. For now only 'autoCreate' value is supported. Type: string (or Expression with resultType string). :type table_option: any + :param sql_writer_use_table_lock: Whether to use table lock during bulk copy. Type: boolean (or + Expression with resultType boolean). + :type sql_writer_use_table_lock: any + :param write_behavior: Write behavior when copying data into Azure SQL. Type: + SqlWriteBehaviorEnum (or Expression with resultType SqlWriteBehaviorEnum). + :type write_behavior: any + :param upsert_settings: SQL upsert settings. + :type upsert_settings: ~azure.mgmt.datafactory.models.SqlUpsertSettings """ _validation = { @@ -6871,6 +6879,9 @@ class AzureSqlSink(CopySink): 'stored_procedure_parameters': {'key': 'storedProcedureParameters', 'type': '{StoredProcedureParameter}'}, 'stored_procedure_table_type_parameter_name': {'key': 'storedProcedureTableTypeParameterName', 'type': 'object'}, 'table_option': {'key': 'tableOption', 'type': 'object'}, + 'sql_writer_use_table_lock': {'key': 'sqlWriterUseTableLock', 'type': 'object'}, + 'write_behavior': {'key': 'writeBehavior', 'type': 'object'}, + 'upsert_settings': {'key': 'upsertSettings', 'type': 'SqlUpsertSettings'}, } def __init__( @@ -6885,6 +6896,9 @@ def __init__( self.stored_procedure_parameters = kwargs.get('stored_procedure_parameters', None) self.stored_procedure_table_type_parameter_name = kwargs.get('stored_procedure_table_type_parameter_name', None) self.table_option = kwargs.get('table_option', None) + self.sql_writer_use_table_lock = kwargs.get('sql_writer_use_table_lock', None) + self.write_behavior = kwargs.get('write_behavior', None) + self.upsert_settings = kwargs.get('upsert_settings', None) class AzureSqlSource(TabularSource): @@ -6913,8 +6927,8 @@ class AzureSqlSource(TabularSource): pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type query_timeout: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any :param sql_reader_query: SQL reader query. Type: string (or Expression with resultType string). :type sql_reader_query: any :param sql_reader_stored_procedure_name: Name of the stored procedure for a SQL Database @@ -6946,7 +6960,7 @@ class AzureSqlSource(TabularSource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, 'sql_reader_query': {'key': 'sqlReaderQuery', 'type': 'object'}, 'sql_reader_stored_procedure_name': {'key': 'sqlReaderStoredProcedureName', 'type': 'object'}, 'stored_procedure_parameters': {'key': 'storedProcedureParameters', 'type': '{StoredProcedureParameter}'}, @@ -7262,8 +7276,8 @@ class AzureTableSource(TabularSource): pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type query_timeout: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any :param azure_table_source_query: Azure Table source query. Type: string (or Expression with resultType string). :type azure_table_source_query: any @@ -7284,7 +7298,7 @@ class AzureTableSource(TabularSource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, 'azure_table_source_query': {'key': 'azureTableSourceQuery', 'type': 'object'}, 'azure_table_source_ignore_table_not_found': {'key': 'azureTableSourceIgnoreTableNotFound', 'type': 'object'}, } @@ -8080,8 +8094,8 @@ class CassandraSource(TabularSource): pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type query_timeout: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any :param query: Database query. Should be a SQL-92 query expression or Cassandra Query Language (CQL) command. Type: string (or Expression with resultType string). :type query: any @@ -8107,7 +8121,7 @@ class CassandraSource(TabularSource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, 'query': {'key': 'query', 'type': 'object'}, 'consistency_level': {'key': 'consistencyLevel', 'type': 'str'}, } @@ -8614,8 +8628,8 @@ class CommonDataServiceForAppsSource(CopySource): Service for Apps (online & on-premises). Type: string (or Expression with resultType string). :type query: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any """ _validation = { @@ -8630,7 +8644,7 @@ class CommonDataServiceForAppsSource(CopySource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'query': {'key': 'query', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, } def __init__( @@ -8881,8 +8895,8 @@ class ConcurSource(TabularSource): pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type query_timeout: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any :param query: A query to retrieve data from source. Type: string (or Expression with resultType string). :type query: any @@ -8900,7 +8914,7 @@ class ConcurSource(TabularSource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, 'query': {'key': 'query', 'type': 'object'}, } @@ -9318,6 +9332,9 @@ class CosmosDbMongoDbApiLinkedService(LinkedService): :type parameters: dict[str, ~azure.mgmt.datafactory.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] + :param is_server_version_above32: Whether the CosmosDB (MongoDB API) server version is higher + than 3.2. The default value is false. Type: boolean (or Expression with resultType boolean). + :type is_server_version_above32: any :param connection_string: Required. The CosmosDB (MongoDB API) connection string. Type: string, SecureString or AzureKeyVaultSecretReference. Type: string, SecureString or AzureKeyVaultSecretReference. @@ -9340,6 +9357,7 @@ class CosmosDbMongoDbApiLinkedService(LinkedService): 'description': {'key': 'description', 'type': 'str'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'is_server_version_above32': {'key': 'typeProperties.isServerVersionAbove32', 'type': 'object'}, 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, 'database': {'key': 'typeProperties.database', 'type': 'object'}, } @@ -9350,6 +9368,7 @@ def __init__( ): super(CosmosDbMongoDbApiLinkedService, self).__init__(**kwargs) self.type = 'CosmosDbMongoDbApi' # type: str + self.is_server_version_above32 = kwargs.get('is_server_version_above32', None) self.connection_string = kwargs['connection_string'] self.database = kwargs['database'] @@ -9450,8 +9469,8 @@ class CosmosDbMongoDbApiSource(CopySource): pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type query_timeout: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any """ _validation = { @@ -9469,7 +9488,7 @@ class CosmosDbMongoDbApiSource(CopySource): 'cursor_methods': {'key': 'cursorMethods', 'type': 'MongoDbCursorMethodsProperties'}, 'batch_size': {'key': 'batchSize', 'type': 'object'}, 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, } def __init__( @@ -9637,8 +9656,8 @@ class CosmosDbSqlApiSource(CopySource): Expression with resultType boolean). :type detect_datetime: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any """ _validation = { @@ -9656,7 +9675,7 @@ class CosmosDbSqlApiSource(CopySource): 'page_size': {'key': 'pageSize', 'type': 'object'}, 'preferred_regions': {'key': 'preferredRegions', 'type': 'object'}, 'detect_datetime': {'key': 'detectDatetime', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, } def __init__( @@ -9754,8 +9773,8 @@ class CouchbaseSource(TabularSource): pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type query_timeout: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any :param query: A query to retrieve data from source. Type: string (or Expression with resultType string). :type query: any @@ -9773,7 +9792,7 @@ class CouchbaseSource(TabularSource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, 'query': {'key': 'query', 'type': 'object'}, } @@ -10593,7 +10612,7 @@ class DataFlow(msrest.serialization.Model): """Azure Data Factory nested object which contains a flow with data movements and transformations. You probably want to use the sub-classes and not this class directly. Known - sub-classes are: MappingDataFlow. + sub-classes are: MappingDataFlow, WranglingDataFlow. All required parameters must be populated in order to send to Azure. @@ -10620,7 +10639,7 @@ class DataFlow(msrest.serialization.Model): } _subtype_map = { - 'type': {'MappingDataFlow': 'MappingDataFlow'} + 'type': {'MappingDataFlow': 'MappingDataFlow', 'WranglingDataFlow': 'WranglingDataFlow'} } def __init__( @@ -11283,8 +11302,9 @@ class DatasetCompression(msrest.serialization.Model): :param additional_properties: Unmatched properties from the message are deserialized to this collection. :type additional_properties: dict[str, any] - :param type: Required. Type of dataset compression.Constant filled by server. - :type type: str + :param type: Required. Type of dataset compression. Type: string (or Expression with resultType + string).Constant filled by server. + :type type: any """ _validation = { @@ -11293,7 +11313,7 @@ class DatasetCompression(msrest.serialization.Model): _attribute_map = { 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, + 'type': {'key': 'type', 'type': 'object'}, } _subtype_map = { @@ -11317,8 +11337,9 @@ class DatasetBZip2Compression(DatasetCompression): :param additional_properties: Unmatched properties from the message are deserialized to this collection. :type additional_properties: dict[str, any] - :param type: Required. Type of dataset compression.Constant filled by server. - :type type: str + :param type: Required. Type of dataset compression. Type: string (or Expression with resultType + string).Constant filled by server. + :type type: any """ _validation = { @@ -11327,7 +11348,7 @@ class DatasetBZip2Compression(DatasetCompression): _attribute_map = { 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, + 'type': {'key': 'type', 'type': 'object'}, } def __init__( @@ -11397,8 +11418,9 @@ class DatasetDeflateCompression(DatasetCompression): :param additional_properties: Unmatched properties from the message are deserialized to this collection. :type additional_properties: dict[str, any] - :param type: Required. Type of dataset compression.Constant filled by server. - :type type: str + :param type: Required. Type of dataset compression. Type: string (or Expression with resultType + string).Constant filled by server. + :type type: any :param level: The Deflate compression level. :type level: any """ @@ -11409,7 +11431,7 @@ class DatasetDeflateCompression(DatasetCompression): _attribute_map = { 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, + 'type': {'key': 'type', 'type': 'object'}, 'level': {'key': 'level', 'type': 'object'}, } @@ -11449,8 +11471,9 @@ class DatasetGZipCompression(DatasetCompression): :param additional_properties: Unmatched properties from the message are deserialized to this collection. :type additional_properties: dict[str, any] - :param type: Required. Type of dataset compression.Constant filled by server. - :type type: str + :param type: Required. Type of dataset compression. Type: string (or Expression with resultType + string).Constant filled by server. + :type type: any :param level: The GZip compression level. :type level: any """ @@ -11461,7 +11484,7 @@ class DatasetGZipCompression(DatasetCompression): _attribute_map = { 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, + 'type': {'key': 'type', 'type': 'object'}, 'level': {'key': 'level', 'type': 'object'}, } @@ -11619,8 +11642,9 @@ class DatasetTarCompression(DatasetCompression): :param additional_properties: Unmatched properties from the message are deserialized to this collection. :type additional_properties: dict[str, any] - :param type: Required. Type of dataset compression.Constant filled by server. - :type type: str + :param type: Required. Type of dataset compression. Type: string (or Expression with resultType + string).Constant filled by server. + :type type: any """ _validation = { @@ -11629,7 +11653,7 @@ class DatasetTarCompression(DatasetCompression): _attribute_map = { 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, + 'type': {'key': 'type', 'type': 'object'}, } def __init__( @@ -11648,8 +11672,9 @@ class DatasetTarGZipCompression(DatasetCompression): :param additional_properties: Unmatched properties from the message are deserialized to this collection. :type additional_properties: dict[str, any] - :param type: Required. Type of dataset compression.Constant filled by server. - :type type: str + :param type: Required. Type of dataset compression. Type: string (or Expression with resultType + string).Constant filled by server. + :type type: any :param level: The TarGZip compression level. :type level: any """ @@ -11660,7 +11685,7 @@ class DatasetTarGZipCompression(DatasetCompression): _attribute_map = { 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, + 'type': {'key': 'type', 'type': 'object'}, 'level': {'key': 'level', 'type': 'object'}, } @@ -11681,8 +11706,9 @@ class DatasetZipDeflateCompression(DatasetCompression): :param additional_properties: Unmatched properties from the message are deserialized to this collection. :type additional_properties: dict[str, any] - :param type: Required. Type of dataset compression.Constant filled by server. - :type type: str + :param type: Required. Type of dataset compression. Type: string (or Expression with resultType + string).Constant filled by server. + :type type: any :param level: The ZipDeflate compression level. :type level: any """ @@ -11693,7 +11719,7 @@ class DatasetZipDeflateCompression(DatasetCompression): _attribute_map = { 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, + 'type': {'key': 'type', 'type': 'object'}, 'level': {'key': 'level', 'type': 'object'}, } @@ -11820,8 +11846,8 @@ class Db2Source(TabularSource): pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type query_timeout: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any :param query: Database query. Type: string (or Expression with resultType string). :type query: any """ @@ -11838,7 +11864,7 @@ class Db2Source(TabularSource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, 'query': {'key': 'query', 'type': 'object'}, } @@ -12244,8 +12270,8 @@ class DelimitedTextSource(CopySource): :param format_settings: DelimitedText format settings. :type format_settings: ~azure.mgmt.datafactory.models.DelimitedTextReadSettings :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any """ _validation = { @@ -12261,7 +12287,7 @@ class DelimitedTextSource(CopySource): 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'store_settings': {'key': 'storeSettings', 'type': 'StoreReadSettings'}, 'format_settings': {'key': 'formatSettings', 'type': 'DelimitedTextReadSettings'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, } def __init__( @@ -12550,8 +12576,8 @@ class DocumentDbCollectionSource(CopySource): pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type query_timeout: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any """ _validation = { @@ -12568,7 +12594,7 @@ class DocumentDbCollectionSource(CopySource): 'query': {'key': 'query', 'type': 'object'}, 'nesting_separator': {'key': 'nestingSeparator', 'type': 'object'}, 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, } def __init__( @@ -12665,8 +12691,8 @@ class DrillSource(TabularSource): pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type query_timeout: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any :param query: A query to retrieve data from source. Type: string (or Expression with resultType string). :type query: any @@ -12684,7 +12710,7 @@ class DrillSource(TabularSource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, 'query': {'key': 'query', 'type': 'object'}, } @@ -12982,8 +13008,8 @@ class DynamicsAXSource(TabularSource): pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type query_timeout: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any :param query: A query to retrieve data from source. Type: string (or Expression with resultType string). :type query: any @@ -13006,7 +13032,7 @@ class DynamicsAXSource(TabularSource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, 'query': {'key': 'query', 'type': 'object'}, 'http_request_timeout': {'key': 'httpRequestTimeout', 'type': 'object'}, } @@ -13289,8 +13315,8 @@ class DynamicsCrmSource(CopySource): (online & on-premises). Type: string (or Expression with resultType string). :type query: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any """ _validation = { @@ -13305,7 +13331,7 @@ class DynamicsCrmSource(CopySource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'query': {'key': 'query', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, } def __init__( @@ -13585,8 +13611,8 @@ class DynamicsSource(CopySource): (online & on-premises). Type: string (or Expression with resultType string). :type query: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any """ _validation = { @@ -13601,7 +13627,7 @@ class DynamicsSource(CopySource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'query': {'key': 'query', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, } def __init__( @@ -13776,8 +13802,8 @@ class EloquaSource(TabularSource): pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type query_timeout: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any :param query: A query to retrieve data from source. Type: string (or Expression with resultType string). :type query: any @@ -13795,7 +13821,7 @@ class EloquaSource(TabularSource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, 'query': {'key': 'query', 'type': 'object'}, } @@ -14020,8 +14046,8 @@ class ExcelSource(CopySource): :param store_settings: Excel store settings. :type store_settings: ~azure.mgmt.datafactory.models.StoreReadSettings :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any """ _validation = { @@ -14036,7 +14062,7 @@ class ExcelSource(CopySource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'store_settings': {'key': 'storeSettings', 'type': 'StoreReadSettings'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, } def __init__( @@ -14130,6 +14156,59 @@ def __init__( self.run_concurrently = kwargs.get('run_concurrently', None) +class ExecuteDataFlowActivityTypeProperties(msrest.serialization.Model): + """Execute data flow activity properties. + + All required parameters must be populated in order to send to Azure. + + :param data_flow: Required. Data flow reference. + :type data_flow: ~azure.mgmt.datafactory.models.DataFlowReference + :param staging: Staging info for execute data flow activity. + :type staging: ~azure.mgmt.datafactory.models.DataFlowStagingInfo + :param integration_runtime: The integration runtime reference. + :type integration_runtime: ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :param compute: Compute properties for data flow activity. + :type compute: ~azure.mgmt.datafactory.models.ExecuteDataFlowActivityTypePropertiesCompute + :param trace_level: Trace level setting used for data flow monitoring output. Supported values + are: 'coarse', 'fine', and 'none'. Type: string (or Expression with resultType string). + :type trace_level: any + :param continue_on_error: Continue on error setting used for data flow execution. Enables + processing to continue if a sink fails. Type: boolean (or Expression with resultType boolean). + :type continue_on_error: any + :param run_concurrently: Concurrent run setting used for data flow execution. Allows sinks with + the same save order to be processed concurrently. Type: boolean (or Expression with resultType + boolean). + :type run_concurrently: any + """ + + _validation = { + 'data_flow': {'required': True}, + } + + _attribute_map = { + 'data_flow': {'key': 'dataFlow', 'type': 'DataFlowReference'}, + 'staging': {'key': 'staging', 'type': 'DataFlowStagingInfo'}, + 'integration_runtime': {'key': 'integrationRuntime', 'type': 'IntegrationRuntimeReference'}, + 'compute': {'key': 'compute', 'type': 'ExecuteDataFlowActivityTypePropertiesCompute'}, + 'trace_level': {'key': 'traceLevel', 'type': 'object'}, + 'continue_on_error': {'key': 'continueOnError', 'type': 'object'}, + 'run_concurrently': {'key': 'runConcurrently', 'type': 'object'}, + } + + def __init__( + self, + **kwargs + ): + super(ExecuteDataFlowActivityTypeProperties, self).__init__(**kwargs) + self.data_flow = kwargs['data_flow'] + self.staging = kwargs.get('staging', None) + self.integration_runtime = kwargs.get('integration_runtime', None) + self.compute = kwargs.get('compute', None) + self.trace_level = kwargs.get('trace_level', None) + self.continue_on_error = kwargs.get('continue_on_error', None) + self.run_concurrently = kwargs.get('run_concurrently', None) + + class ExecuteDataFlowActivityTypePropertiesCompute(msrest.serialization.Model): """Compute properties for data flow activity. @@ -14212,6 +14291,56 @@ def __init__( self.wait_on_completion = kwargs.get('wait_on_completion', None) +class ExecutePowerQueryActivityTypeProperties(ExecuteDataFlowActivityTypeProperties): + """Execute power query data flow activity properties. + + All required parameters must be populated in order to send to Azure. + + :param data_flow: Required. Data flow reference. + :type data_flow: ~azure.mgmt.datafactory.models.DataFlowReference + :param staging: Staging info for execute data flow activity. + :type staging: ~azure.mgmt.datafactory.models.DataFlowStagingInfo + :param integration_runtime: The integration runtime reference. + :type integration_runtime: ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :param compute: Compute properties for data flow activity. + :type compute: ~azure.mgmt.datafactory.models.ExecuteDataFlowActivityTypePropertiesCompute + :param trace_level: Trace level setting used for data flow monitoring output. Supported values + are: 'coarse', 'fine', and 'none'. Type: string (or Expression with resultType string). + :type trace_level: any + :param continue_on_error: Continue on error setting used for data flow execution. Enables + processing to continue if a sink fails. Type: boolean (or Expression with resultType boolean). + :type continue_on_error: any + :param run_concurrently: Concurrent run setting used for data flow execution. Allows sinks with + the same save order to be processed concurrently. Type: boolean (or Expression with resultType + boolean). + :type run_concurrently: any + :param sinks: List of Power Query activity sinks mapped to a queryName. + :type sinks: dict[str, ~azure.mgmt.datafactory.models.PowerQuerySink] + """ + + _validation = { + 'data_flow': {'required': True}, + } + + _attribute_map = { + 'data_flow': {'key': 'dataFlow', 'type': 'DataFlowReference'}, + 'staging': {'key': 'staging', 'type': 'DataFlowStagingInfo'}, + 'integration_runtime': {'key': 'integrationRuntime', 'type': 'IntegrationRuntimeReference'}, + 'compute': {'key': 'compute', 'type': 'ExecuteDataFlowActivityTypePropertiesCompute'}, + 'trace_level': {'key': 'traceLevel', 'type': 'object'}, + 'continue_on_error': {'key': 'continueOnError', 'type': 'object'}, + 'run_concurrently': {'key': 'runConcurrently', 'type': 'object'}, + 'sinks': {'key': 'sinks', 'type': '{PowerQuerySink}'}, + } + + def __init__( + self, + **kwargs + ): + super(ExecutePowerQueryActivityTypeProperties, self).__init__(**kwargs) + self.sinks = kwargs.get('sinks', None) + + class ExecuteSSISPackageActivity(ExecutionActivity): """Execute SSIS package activity. @@ -14317,6 +14446,89 @@ def __init__( self.log_location = kwargs.get('log_location', None) +class ExecuteWranglingDataflowActivity(Activity): + """Execute power query activity. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are deserialized to this + collection. + :type additional_properties: dict[str, any] + :param name: Required. Activity name. + :type name: str + :param type: Required. Type of activity.Constant filled by server. + :type type: str + :param description: Activity description. + :type description: str + :param depends_on: Activity depends on condition. + :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] + :param user_properties: Activity user properties. + :type user_properties: list[~azure.mgmt.datafactory.models.UserProperty] + :param policy: Activity policy. + :type policy: ~azure.mgmt.datafactory.models.ActivityPolicy + :param data_flow: Required. Data flow reference. + :type data_flow: ~azure.mgmt.datafactory.models.DataFlowReference + :param staging: Staging info for execute data flow activity. + :type staging: ~azure.mgmt.datafactory.models.DataFlowStagingInfo + :param integration_runtime: The integration runtime reference. + :type integration_runtime: ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :param compute: Compute properties for data flow activity. + :type compute: ~azure.mgmt.datafactory.models.ExecuteDataFlowActivityTypePropertiesCompute + :param trace_level: Trace level setting used for data flow monitoring output. Supported values + are: 'coarse', 'fine', and 'none'. Type: string (or Expression with resultType string). + :type trace_level: any + :param continue_on_error: Continue on error setting used for data flow execution. Enables + processing to continue if a sink fails. Type: boolean (or Expression with resultType boolean). + :type continue_on_error: any + :param run_concurrently: Concurrent run setting used for data flow execution. Allows sinks with + the same save order to be processed concurrently. Type: boolean (or Expression with resultType + boolean). + :type run_concurrently: any + :param sinks: List of Power Query activity sinks mapped to a queryName. + :type sinks: dict[str, ~azure.mgmt.datafactory.models.PowerQuerySink] + """ + + _validation = { + 'name': {'required': True}, + 'type': {'required': True}, + 'data_flow': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'name': {'key': 'name', 'type': 'str'}, + 'type': {'key': 'type', 'type': 'str'}, + 'description': {'key': 'description', 'type': 'str'}, + 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, + 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, + 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, + 'data_flow': {'key': 'typeProperties.dataFlow', 'type': 'DataFlowReference'}, + 'staging': {'key': 'typeProperties.staging', 'type': 'DataFlowStagingInfo'}, + 'integration_runtime': {'key': 'typeProperties.integrationRuntime', 'type': 'IntegrationRuntimeReference'}, + 'compute': {'key': 'typeProperties.compute', 'type': 'ExecuteDataFlowActivityTypePropertiesCompute'}, + 'trace_level': {'key': 'typeProperties.traceLevel', 'type': 'object'}, + 'continue_on_error': {'key': 'typeProperties.continueOnError', 'type': 'object'}, + 'run_concurrently': {'key': 'typeProperties.runConcurrently', 'type': 'object'}, + 'sinks': {'key': 'typeProperties.sinks', 'type': '{PowerQuerySink}'}, + } + + def __init__( + self, + **kwargs + ): + super(ExecuteWranglingDataflowActivity, self).__init__(**kwargs) + self.type = 'ExecuteWranglingDataflow' # type: str + self.policy = kwargs.get('policy', None) + self.data_flow = kwargs['data_flow'] + self.staging = kwargs.get('staging', None) + self.integration_runtime = kwargs.get('integration_runtime', None) + self.compute = kwargs.get('compute', None) + self.trace_level = kwargs.get('trace_level', None) + self.continue_on_error = kwargs.get('continue_on_error', None) + self.run_concurrently = kwargs.get('run_concurrently', None) + self.sinks = kwargs.get('sinks', None) + + class ExposureControlBatchRequest(msrest.serialization.Model): """A list of exposure control features. @@ -14659,6 +14871,10 @@ class FactoryGitHubConfiguration(FactoryRepoConfiguration): :type last_commit_id: str :param host_name: GitHub Enterprise host name. For example: https://github.mydomain.com. :type host_name: str + :param client_id: GitHub bring your own app client id. + :type client_id: str + :param client_secret: GitHub bring your own app client secret information. + :type client_secret: ~azure.mgmt.datafactory.models.GitHubClientSecret """ _validation = { @@ -14677,6 +14893,8 @@ class FactoryGitHubConfiguration(FactoryRepoConfiguration): 'root_folder': {'key': 'rootFolder', 'type': 'str'}, 'last_commit_id': {'key': 'lastCommitId', 'type': 'str'}, 'host_name': {'key': 'hostName', 'type': 'str'}, + 'client_id': {'key': 'clientId', 'type': 'str'}, + 'client_secret': {'key': 'clientSecret', 'type': 'GitHubClientSecret'}, } def __init__( @@ -14686,6 +14904,8 @@ def __init__( super(FactoryGitHubConfiguration, self).__init__(**kwargs) self.type = 'FactoryGitHubConfiguration' # type: str self.host_name = kwargs.get('host_name', None) + self.client_id = kwargs.get('client_id', None) + self.client_secret = kwargs.get('client_secret', None) class FactoryIdentity(msrest.serialization.Model): @@ -15254,8 +15474,8 @@ class FileSystemSource(CopySource): true. Type: boolean (or Expression with resultType boolean). :type recursive: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any """ _validation = { @@ -15270,7 +15490,7 @@ class FileSystemSource(CopySource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'recursive': {'key': 'recursive', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, } def __init__( @@ -15999,8 +16219,8 @@ class GoogleAdWordsSource(TabularSource): pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type query_timeout: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any :param query: A query to retrieve data from source. Type: string (or Expression with resultType string). :type query: any @@ -16018,7 +16238,7 @@ class GoogleAdWordsSource(TabularSource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, 'query': {'key': 'query', 'type': 'object'}, } @@ -16232,8 +16452,8 @@ class GoogleBigQuerySource(TabularSource): pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type query_timeout: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any :param query: A query to retrieve data from source. Type: string (or Expression with resultType string). :type query: any @@ -16251,7 +16471,7 @@ class GoogleBigQuerySource(TabularSource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, 'query': {'key': 'query', 'type': 'object'}, } @@ -16544,8 +16764,8 @@ class GreenplumSource(TabularSource): pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type query_timeout: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any :param query: A query to retrieve data from source. Type: string (or Expression with resultType string). :type query: any @@ -16563,7 +16783,7 @@ class GreenplumSource(TabularSource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, 'query': {'key': 'query', 'type': 'object'}, } @@ -16825,8 +17045,8 @@ class HBaseSource(TabularSource): pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type query_timeout: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any :param query: A query to retrieve data from source. Type: string (or Expression with resultType string). :type query: any @@ -16844,7 +17064,7 @@ class HBaseSource(TabularSource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, 'query': {'key': 'query', 'type': 'object'}, } @@ -18046,8 +18266,8 @@ class HiveSource(TabularSource): pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type query_timeout: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any :param query: A query to retrieve data from source. Type: string (or Expression with resultType string). :type query: any @@ -18065,7 +18285,7 @@ class HiveSource(TabularSource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, 'query': {'key': 'query', 'type': 'object'}, } @@ -18581,8 +18801,8 @@ class HubspotSource(TabularSource): pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type query_timeout: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any :param query: A query to retrieve data from source. Type: string (or Expression with resultType string). :type query: any @@ -18600,7 +18820,7 @@ class HubspotSource(TabularSource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, 'query': {'key': 'query', 'type': 'object'}, } @@ -18866,8 +19086,8 @@ class ImpalaSource(TabularSource): pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type query_timeout: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any :param query: A query to retrieve data from source. Type: string (or Expression with resultType string). :type query: any @@ -18885,7 +19105,7 @@ class ImpalaSource(TabularSource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, 'query': {'key': 'query', 'type': 'object'}, } @@ -19056,8 +19276,8 @@ class InformixSource(TabularSource): pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type query_timeout: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any :param query: Database query. Type: string (or Expression with resultType string). :type query: any """ @@ -19074,7 +19294,7 @@ class InformixSource(TabularSource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, 'query': {'key': 'query', 'type': 'object'}, } @@ -19360,6 +19580,9 @@ class IntegrationRuntimeDataFlowProperties(msrest.serialization.Model): :param time_to_live: Time to live (in minutes) setting of the cluster which will execute data flow job. :type time_to_live: int + :param cleanup: Cluster will not be recycled and it will be used in next data flow activity run + until TTL (time to live) is reached if this is set as false. Default is true. + :type cleanup: bool """ _validation = { @@ -19371,6 +19594,7 @@ class IntegrationRuntimeDataFlowProperties(msrest.serialization.Model): 'compute_type': {'key': 'computeType', 'type': 'str'}, 'core_count': {'key': 'coreCount', 'type': 'int'}, 'time_to_live': {'key': 'timeToLive', 'type': 'int'}, + 'cleanup': {'key': 'cleanup', 'type': 'bool'}, } def __init__( @@ -19382,6 +19606,7 @@ def __init__( self.compute_type = kwargs.get('compute_type', None) self.core_count = kwargs.get('core_count', None) self.time_to_live = kwargs.get('time_to_live', None) + self.cleanup = kwargs.get('cleanup', None) class IntegrationRuntimeDataProxyProperties(msrest.serialization.Model): @@ -19582,6 +19807,93 @@ def __init__( self.received_bytes = None +class IntegrationRuntimeOutboundNetworkDependenciesCategoryEndpoint(msrest.serialization.Model): + """Azure-SSIS integration runtime outbound network dependency endpoints for one category. + + :param category: The category of outbound network dependency. + :type category: str + :param endpoints: The endpoints for outbound network dependency. + :type endpoints: + list[~azure.mgmt.datafactory.models.IntegrationRuntimeOutboundNetworkDependenciesEndpoint] + """ + + _attribute_map = { + 'category': {'key': 'category', 'type': 'str'}, + 'endpoints': {'key': 'endpoints', 'type': '[IntegrationRuntimeOutboundNetworkDependenciesEndpoint]'}, + } + + def __init__( + self, + **kwargs + ): + super(IntegrationRuntimeOutboundNetworkDependenciesCategoryEndpoint, self).__init__(**kwargs) + self.category = kwargs.get('category', None) + self.endpoints = kwargs.get('endpoints', None) + + +class IntegrationRuntimeOutboundNetworkDependenciesEndpoint(msrest.serialization.Model): + """The endpoint for Azure-SSIS integration runtime outbound network dependency. + + :param domain_name: The domain name of endpoint. + :type domain_name: str + :param endpoint_details: The details of endpoint. + :type endpoint_details: + list[~azure.mgmt.datafactory.models.IntegrationRuntimeOutboundNetworkDependenciesEndpointDetails] + """ + + _attribute_map = { + 'domain_name': {'key': 'domainName', 'type': 'str'}, + 'endpoint_details': {'key': 'endpointDetails', 'type': '[IntegrationRuntimeOutboundNetworkDependenciesEndpointDetails]'}, + } + + def __init__( + self, + **kwargs + ): + super(IntegrationRuntimeOutboundNetworkDependenciesEndpoint, self).__init__(**kwargs) + self.domain_name = kwargs.get('domain_name', None) + self.endpoint_details = kwargs.get('endpoint_details', None) + + +class IntegrationRuntimeOutboundNetworkDependenciesEndpointDetails(msrest.serialization.Model): + """The details of Azure-SSIS integration runtime outbound network dependency endpoint. + + :param port: The port of endpoint. + :type port: int + """ + + _attribute_map = { + 'port': {'key': 'port', 'type': 'int'}, + } + + def __init__( + self, + **kwargs + ): + super(IntegrationRuntimeOutboundNetworkDependenciesEndpointDetails, self).__init__(**kwargs) + self.port = kwargs.get('port', None) + + +class IntegrationRuntimeOutboundNetworkDependenciesEndpointsResponse(msrest.serialization.Model): + """Azure-SSIS integration runtime outbound network dependency endpoints. + + :param value: The list of outbound network dependency endpoints. + :type value: + list[~azure.mgmt.datafactory.models.IntegrationRuntimeOutboundNetworkDependenciesCategoryEndpoint] + """ + + _attribute_map = { + 'value': {'key': 'value', 'type': '[IntegrationRuntimeOutboundNetworkDependenciesCategoryEndpoint]'}, + } + + def __init__( + self, + **kwargs + ): + super(IntegrationRuntimeOutboundNetworkDependenciesEndpointsResponse, self).__init__(**kwargs) + self.value = kwargs.get('value', None) + + class IntegrationRuntimeReference(msrest.serialization.Model): """Integration runtime reference type. @@ -19916,6 +20228,9 @@ class IntegrationRuntimeVNetProperties(msrest.serialization.Model): :param public_i_ps: Resource IDs of the public IP addresses that this integration runtime will use. :type public_i_ps: list[str] + :param subnet_id: The ID of subnet, to which this Azure-SSIS integration runtime will be + joined. + :type subnet_id: str """ _attribute_map = { @@ -19923,6 +20238,7 @@ class IntegrationRuntimeVNetProperties(msrest.serialization.Model): 'v_net_id': {'key': 'vNetId', 'type': 'str'}, 'subnet': {'key': 'subnet', 'type': 'str'}, 'public_i_ps': {'key': 'publicIPs', 'type': '[str]'}, + 'subnet_id': {'key': 'subnetId', 'type': 'str'}, } def __init__( @@ -19934,6 +20250,7 @@ def __init__( self.v_net_id = kwargs.get('v_net_id', None) self.subnet = kwargs.get('subnet', None) self.public_i_ps = kwargs.get('public_i_ps', None) + self.subnet_id = kwargs.get('subnet_id', None) class JiraLinkedService(LinkedService): @@ -20104,8 +20421,8 @@ class JiraSource(TabularSource): pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type query_timeout: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any :param query: A query to retrieve data from source. Type: string (or Expression with resultType string). :type query: any @@ -20123,7 +20440,7 @@ class JiraSource(TabularSource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, 'query': {'key': 'query', 'type': 'object'}, } @@ -20393,8 +20710,8 @@ class JsonSource(CopySource): :param format_settings: Json format settings. :type format_settings: ~azure.mgmt.datafactory.models.JsonReadSettings :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any """ _validation = { @@ -20410,7 +20727,7 @@ class JsonSource(CopySource): 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'store_settings': {'key': 'storeSettings', 'type': 'StoreReadSettings'}, 'format_settings': {'key': 'formatSettings', 'type': 'JsonReadSettings'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, } def __init__( @@ -21091,8 +21408,8 @@ class MagentoSource(TabularSource): pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type query_timeout: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any :param query: A query to retrieve data from source. Type: string (or Expression with resultType string). :type query: any @@ -21110,7 +21427,7 @@ class MagentoSource(TabularSource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, 'query': {'key': 'query', 'type': 'object'}, } @@ -21817,8 +22134,8 @@ class MariaDBSource(TabularSource): pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type query_timeout: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any :param query: A query to retrieve data from source. Type: string (or Expression with resultType string). :type query: any @@ -21836,7 +22153,7 @@ class MariaDBSource(TabularSource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, 'query': {'key': 'query', 'type': 'object'}, } @@ -22068,8 +22385,8 @@ class MarketoSource(TabularSource): pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type query_timeout: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any :param query: A query to retrieve data from source. Type: string (or Expression with resultType string). :type query: any @@ -22087,7 +22404,7 @@ class MarketoSource(TabularSource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, 'query': {'key': 'query', 'type': 'object'}, } @@ -22280,8 +22597,8 @@ class MicrosoftAccessSource(CopySource): :param query: Database query. Type: string (or Expression with resultType string). :type query: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any """ _validation = { @@ -22296,7 +22613,7 @@ class MicrosoftAccessSource(CopySource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'query': {'key': 'query', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, } def __init__( @@ -22578,8 +22895,8 @@ class MongoDbAtlasSource(CopySource): pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type query_timeout: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any """ _validation = { @@ -22597,7 +22914,7 @@ class MongoDbAtlasSource(CopySource): 'cursor_methods': {'key': 'cursorMethods', 'type': 'MongoDbCursorMethodsProperties'}, 'batch_size': {'key': 'batchSize', 'type': 'object'}, 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, } def __init__( @@ -22835,8 +23152,8 @@ class MongoDbSource(CopySource): with resultType string). :type query: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any """ _validation = { @@ -22851,7 +23168,7 @@ class MongoDbSource(CopySource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'query': {'key': 'query', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, } def __init__( @@ -23073,8 +23390,8 @@ class MongoDbV2Source(CopySource): pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type query_timeout: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any """ _validation = { @@ -23092,7 +23409,7 @@ class MongoDbV2Source(CopySource): 'cursor_methods': {'key': 'cursorMethods', 'type': 'MongoDbCursorMethodsProperties'}, 'batch_size': {'key': 'batchSize', 'type': 'object'}, 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, } def __init__( @@ -23190,8 +23507,8 @@ class MySqlSource(TabularSource): pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type query_timeout: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any :param query: Database query. Type: string (or Expression with resultType string). :type query: any """ @@ -23208,7 +23525,7 @@ class MySqlSource(TabularSource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, 'query': {'key': 'query', 'type': 'object'}, } @@ -23393,8 +23710,8 @@ class NetezzaSource(TabularSource): pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type query_timeout: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any :param query: A query to retrieve data from source. Type: string (or Expression with resultType string). :type query: any @@ -23417,7 +23734,7 @@ class NetezzaSource(TabularSource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, 'query': {'key': 'query', 'type': 'object'}, 'partition_option': {'key': 'partitionOption', 'type': 'object'}, 'partition_settings': {'key': 'partitionSettings', 'type': 'NetezzaPartitionSettings'}, @@ -23709,8 +24026,8 @@ class ODataSource(CopySource): ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type http_request_timeout: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any """ _validation = { @@ -23726,7 +24043,7 @@ class ODataSource(CopySource): 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'query': {'key': 'query', 'type': 'object'}, 'http_request_timeout': {'key': 'httpRequestTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, } def __init__( @@ -23897,8 +24214,8 @@ class OdbcSource(TabularSource): pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type query_timeout: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any :param query: Database query. Type: string (or Expression with resultType string). :type query: any """ @@ -23915,7 +24232,7 @@ class OdbcSource(TabularSource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, 'query': {'key': 'query', 'type': 'object'}, } @@ -24892,8 +25209,8 @@ class OracleServiceCloudSource(TabularSource): pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type query_timeout: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any :param query: A query to retrieve data from source. Type: string (or Expression with resultType string). :type query: any @@ -24911,7 +25228,7 @@ class OracleServiceCloudSource(TabularSource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, 'query': {'key': 'query', 'type': 'object'}, } @@ -25016,8 +25333,8 @@ class OracleSource(CopySource): :param partition_settings: The settings that will be leveraged for Oracle source partitioning. :type partition_settings: ~azure.mgmt.datafactory.models.OraclePartitionSettings :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any """ _validation = { @@ -25035,7 +25352,7 @@ class OracleSource(CopySource): 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, 'partition_option': {'key': 'partitionOption', 'type': 'object'}, 'partition_settings': {'key': 'partitionSettings', 'type': 'OraclePartitionSettings'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, } def __init__( @@ -25304,8 +25621,8 @@ class OrcSource(CopySource): :param store_settings: ORC store settings. :type store_settings: ~azure.mgmt.datafactory.models.StoreReadSettings :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any """ _validation = { @@ -25320,7 +25637,7 @@ class OrcSource(CopySource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'store_settings': {'key': 'storeSettings', 'type': 'StoreReadSettings'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, } def __init__( @@ -25617,8 +25934,8 @@ class ParquetSource(CopySource): :param store_settings: Parquet store settings. :type store_settings: ~azure.mgmt.datafactory.models.StoreReadSettings :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any """ _validation = { @@ -25633,7 +25950,7 @@ class ParquetSource(CopySource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'store_settings': {'key': 'storeSettings', 'type': 'StoreReadSettings'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, } def __init__( @@ -25847,8 +26164,8 @@ class PaypalSource(TabularSource): pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type query_timeout: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any :param query: A query to retrieve data from source. Type: string (or Expression with resultType string). :type query: any @@ -25866,7 +26183,7 @@ class PaypalSource(TabularSource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, 'query': {'key': 'query', 'type': 'object'}, } @@ -26079,8 +26396,8 @@ class PhoenixSource(TabularSource): pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type query_timeout: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any :param query: A query to retrieve data from source. Type: string (or Expression with resultType string). :type query: any @@ -26098,7 +26415,7 @@ class PhoenixSource(TabularSource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, 'query': {'key': 'query', 'type': 'object'}, } @@ -26602,8 +26919,8 @@ class PostgreSqlSource(TabularSource): pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type query_timeout: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any :param query: Database query. Type: string (or Expression with resultType string). :type query: any """ @@ -26620,7 +26937,7 @@ class PostgreSqlSource(TabularSource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, 'query': {'key': 'query', 'type': 'object'}, } @@ -26701,6 +27018,86 @@ def __init__( self.schema_type_properties_schema = kwargs.get('schema_type_properties_schema', None) +class PowerQuerySink(DataFlowSink): + """Power query sink. + + All required parameters must be populated in order to send to Azure. + + :param name: Required. Transformation name. + :type name: str + :param description: Transformation description. + :type description: str + :param dataset: Dataset reference. + :type dataset: ~azure.mgmt.datafactory.models.DatasetReference + :param linked_service: Linked service reference. + :type linked_service: ~azure.mgmt.datafactory.models.LinkedServiceReference + :param schema_linked_service: Schema linked service reference. + :type schema_linked_service: ~azure.mgmt.datafactory.models.LinkedServiceReference + :param script: sink script. + :type script: str + """ + + _validation = { + 'name': {'required': True}, + } + + _attribute_map = { + 'name': {'key': 'name', 'type': 'str'}, + 'description': {'key': 'description', 'type': 'str'}, + 'dataset': {'key': 'dataset', 'type': 'DatasetReference'}, + 'linked_service': {'key': 'linkedService', 'type': 'LinkedServiceReference'}, + 'schema_linked_service': {'key': 'schemaLinkedService', 'type': 'LinkedServiceReference'}, + 'script': {'key': 'script', 'type': 'str'}, + } + + def __init__( + self, + **kwargs + ): + super(PowerQuerySink, self).__init__(**kwargs) + self.script = kwargs.get('script', None) + + +class PowerQuerySource(DataFlowSource): + """Power query source. + + All required parameters must be populated in order to send to Azure. + + :param name: Required. Transformation name. + :type name: str + :param description: Transformation description. + :type description: str + :param dataset: Dataset reference. + :type dataset: ~azure.mgmt.datafactory.models.DatasetReference + :param linked_service: Linked service reference. + :type linked_service: ~azure.mgmt.datafactory.models.LinkedServiceReference + :param schema_linked_service: Schema linked service reference. + :type schema_linked_service: ~azure.mgmt.datafactory.models.LinkedServiceReference + :param script: source script. + :type script: str + """ + + _validation = { + 'name': {'required': True}, + } + + _attribute_map = { + 'name': {'key': 'name', 'type': 'str'}, + 'description': {'key': 'description', 'type': 'str'}, + 'dataset': {'key': 'dataset', 'type': 'DatasetReference'}, + 'linked_service': {'key': 'linkedService', 'type': 'LinkedServiceReference'}, + 'schema_linked_service': {'key': 'schemaLinkedService', 'type': 'LinkedServiceReference'}, + 'script': {'key': 'script', 'type': 'str'}, + } + + def __init__( + self, + **kwargs + ): + super(PowerQuerySource, self).__init__(**kwargs) + self.script = kwargs.get('script', None) + + class PrestoLinkedService(LinkedService): """Presto server linked service. @@ -26909,8 +27306,8 @@ class PrestoSource(TabularSource): pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type query_timeout: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any :param query: A query to retrieve data from source. Type: string (or Expression with resultType string). :type query: any @@ -26928,7 +27325,7 @@ class PrestoSource(TabularSource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, 'query': {'key': 'query', 'type': 'object'}, } @@ -27385,8 +27782,8 @@ class QuickBooksSource(TabularSource): pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type query_timeout: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any :param query: A query to retrieve data from source. Type: string (or Expression with resultType string). :type query: any @@ -27404,7 +27801,7 @@ class QuickBooksSource(TabularSource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, 'query': {'key': 'query', 'type': 'object'}, } @@ -27582,8 +27979,8 @@ class RelationalSource(CopySource): :param query: Database query. Type: string (or Expression with resultType string). :type query: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any """ _validation = { @@ -27598,7 +27995,7 @@ class RelationalSource(CopySource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'query': {'key': 'query', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, } def __init__( @@ -27933,8 +28330,8 @@ class ResponsysSource(TabularSource): pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type query_timeout: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any :param query: A query to retrieve data from source. Type: string (or Expression with resultType string). :type query: any @@ -27952,7 +28349,7 @@ class ResponsysSource(TabularSource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, 'query': {'key': 'query', 'type': 'object'}, } @@ -28272,8 +28669,8 @@ class RestSource(CopySource): :param request_interval: The time to await before sending next page request. :type request_interval: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any """ _validation = { @@ -28293,7 +28690,7 @@ class RestSource(CopySource): 'pagination_rules': {'key': 'paginationRules', 'type': 'object'}, 'http_request_timeout': {'key': 'httpRequestTimeout', 'type': 'object'}, 'request_interval': {'key': 'requestInterval', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, } def __init__( @@ -28695,8 +29092,8 @@ class SalesforceMarketingCloudSource(TabularSource): pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type query_timeout: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any :param query: A query to retrieve data from source. Type: string (or Expression with resultType string). :type query: any @@ -28714,7 +29111,7 @@ class SalesforceMarketingCloudSource(TabularSource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, 'query': {'key': 'query', 'type': 'object'}, } @@ -29022,8 +29419,8 @@ class SalesforceServiceCloudSource(CopySource): include: "Query", "QueryAll". :type read_behavior: str or ~azure.mgmt.datafactory.models.SalesforceSourceReadBehavior :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any """ _validation = { @@ -29039,7 +29436,7 @@ class SalesforceServiceCloudSource(CopySource): 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'query': {'key': 'query', 'type': 'object'}, 'read_behavior': {'key': 'readBehavior', 'type': 'str'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, } def __init__( @@ -29151,8 +29548,8 @@ class SalesforceSource(TabularSource): pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type query_timeout: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any :param query: Database query. Type: string (or Expression with resultType string). :type query: any :param read_behavior: The read behavior for the operation. Default is Query. Possible values @@ -29172,7 +29569,7 @@ class SalesforceSource(TabularSource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, 'query': {'key': 'query', 'type': 'object'}, 'read_behavior': {'key': 'readBehavior', 'type': 'str'}, } @@ -29341,8 +29738,8 @@ class SapBwSource(TabularSource): pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type query_timeout: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any :param query: MDX query. Type: string (or Expression with resultType string). :type query: any """ @@ -29359,7 +29756,7 @@ class SapBwSource(TabularSource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, 'query': {'key': 'query', 'type': 'object'}, } @@ -29587,8 +29984,8 @@ class SapCloudForCustomerSource(TabularSource): pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type query_timeout: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any :param query: SAP Cloud for Customer OData query. For example, "$top=1". Type: string (or Expression with resultType string). :type query: any @@ -29611,7 +30008,7 @@ class SapCloudForCustomerSource(TabularSource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, 'query': {'key': 'query', 'type': 'object'}, 'http_request_timeout': {'key': 'httpRequestTimeout', 'type': 'object'}, } @@ -29775,8 +30172,8 @@ class SapEccSource(TabularSource): pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type query_timeout: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any :param query: SAP ECC OData query. For example, "$top=1". Type: string (or Expression with resultType string). :type query: any @@ -29799,7 +30196,7 @@ class SapEccSource(TabularSource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, 'query': {'key': 'query', 'type': 'object'}, 'http_request_timeout': {'key': 'httpRequestTimeout', 'type': 'object'}, } @@ -29931,8 +30328,8 @@ class SapHanaSource(TabularSource): pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type query_timeout: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any :param query: SAP HANA Sql query. Type: string (or Expression with resultType string). :type query: any :param packet_size: The packet size of data read from SAP HANA. Type: integer(or Expression @@ -29958,7 +30355,7 @@ class SapHanaSource(TabularSource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, 'query': {'key': 'query', 'type': 'object'}, 'packet_size': {'key': 'packetSize', 'type': 'object'}, 'partition_option': {'key': 'partitionOption', 'type': 'object'}, @@ -30165,8 +30562,8 @@ class SapOpenHubSource(TabularSource): pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type query_timeout: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any :param exclude_last_request: Whether to exclude the records of the last request. The default value is true. Type: boolean (or Expression with resultType boolean). :type exclude_last_request: any @@ -30195,7 +30592,7 @@ class SapOpenHubSource(TabularSource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, 'exclude_last_request': {'key': 'excludeLastRequest', 'type': 'object'}, 'base_request_id': {'key': 'baseRequestId', 'type': 'object'}, 'custom_rfc_read_table_function_module': {'key': 'customRfcReadTableFunctionModule', 'type': 'object'}, @@ -30531,8 +30928,8 @@ class SapTableSource(TabularSource): pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type query_timeout: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any :param row_count: The number of rows to be retrieved. Type: integer(or Expression with resultType integer). :type row_count: any @@ -30576,7 +30973,7 @@ class SapTableSource(TabularSource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, 'row_count': {'key': 'rowCount', 'type': 'object'}, 'row_skips': {'key': 'rowSkips', 'type': 'object'}, 'rfc_table_fields': {'key': 'rfcTableFields', 'type': 'object'}, @@ -31270,8 +31667,8 @@ class ServiceNowSource(TabularSource): pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type query_timeout: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any :param query: A query to retrieve data from source. Type: string (or Expression with resultType string). :type query: any @@ -31289,7 +31686,7 @@ class ServiceNowSource(TabularSource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, 'query': {'key': 'query', 'type': 'object'}, } @@ -32016,8 +32413,8 @@ class ShopifySource(TabularSource): pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type query_timeout: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any :param query: A query to retrieve data from source. Type: string (or Expression with resultType string). :type query: any @@ -32035,7 +32432,7 @@ class ShopifySource(TabularSource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, 'query': {'key': 'query', 'type': 'object'}, } @@ -32604,8 +33001,8 @@ class SparkSource(TabularSource): pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type query_timeout: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any :param query: A query to retrieve data from source. Type: string (or Expression with resultType string). :type query: any @@ -32623,7 +33020,7 @@ class SparkSource(TabularSource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, 'query': {'key': 'query', 'type': 'object'}, } @@ -32719,6 +33116,14 @@ class SqlDWSink(CopySink): :param table_option: The option to handle sink table, such as autoCreate. For now only 'autoCreate' value is supported. Type: string (or Expression with resultType string). :type table_option: any + :param sql_writer_use_table_lock: Whether to use table lock during bulk copy. Type: boolean (or + Expression with resultType boolean). + :type sql_writer_use_table_lock: any + :param write_behavior: Write behavior when copying data into azure SQL DW. Type: + SqlDWWriteBehaviorEnum (or Expression with resultType SqlDWWriteBehaviorEnum). + :type write_behavior: any + :param upsert_settings: SQL DW upsert settings. + :type upsert_settings: ~azure.mgmt.datafactory.models.SqlDWUpsertSettings """ _validation = { @@ -32740,6 +33145,9 @@ class SqlDWSink(CopySink): 'allow_copy_command': {'key': 'allowCopyCommand', 'type': 'object'}, 'copy_command_settings': {'key': 'copyCommandSettings', 'type': 'DWCopyCommandSettings'}, 'table_option': {'key': 'tableOption', 'type': 'object'}, + 'sql_writer_use_table_lock': {'key': 'sqlWriterUseTableLock', 'type': 'object'}, + 'write_behavior': {'key': 'writeBehavior', 'type': 'object'}, + 'upsert_settings': {'key': 'upsertSettings', 'type': 'SqlDWUpsertSettings'}, } def __init__( @@ -32754,6 +33162,9 @@ def __init__( self.allow_copy_command = kwargs.get('allow_copy_command', None) self.copy_command_settings = kwargs.get('copy_command_settings', None) self.table_option = kwargs.get('table_option', None) + self.sql_writer_use_table_lock = kwargs.get('sql_writer_use_table_lock', None) + self.write_behavior = kwargs.get('write_behavior', None) + self.upsert_settings = kwargs.get('upsert_settings', None) class SqlDWSource(TabularSource): @@ -32782,8 +33193,8 @@ class SqlDWSource(TabularSource): pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type query_timeout: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any :param sql_reader_query: SQL Data Warehouse reader query. Type: string (or Expression with resultType string). :type sql_reader_query: any @@ -32814,7 +33225,7 @@ class SqlDWSource(TabularSource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, 'sql_reader_query': {'key': 'sqlReaderQuery', 'type': 'object'}, 'sql_reader_stored_procedure_name': {'key': 'sqlReaderStoredProcedureName', 'type': 'object'}, 'stored_procedure_parameters': {'key': 'storedProcedureParameters', 'type': 'object'}, @@ -32835,6 +33246,31 @@ def __init__( self.partition_settings = kwargs.get('partition_settings', None) +class SqlDWUpsertSettings(msrest.serialization.Model): + """Sql DW upsert option settings. + + :param interim_schema_name: Schema name for interim table. Type: string (or Expression with + resultType string). + :type interim_schema_name: any + :param keys: Key column names for unique row identification. Type: array of strings (or + Expression with resultType array of strings). + :type keys: any + """ + + _attribute_map = { + 'interim_schema_name': {'key': 'interimSchemaName', 'type': 'object'}, + 'keys': {'key': 'keys', 'type': 'object'}, + } + + def __init__( + self, + **kwargs + ): + super(SqlDWUpsertSettings, self).__init__(**kwargs) + self.interim_schema_name = kwargs.get('interim_schema_name', None) + self.keys = kwargs.get('keys', None) + + class SqlMISink(CopySink): """A copy activity Azure SQL Managed Instance sink. @@ -32881,6 +33317,14 @@ class SqlMISink(CopySink): :param table_option: The option to handle sink table, such as autoCreate. For now only 'autoCreate' value is supported. Type: string (or Expression with resultType string). :type table_option: any + :param sql_writer_use_table_lock: Whether to use table lock during bulk copy. Type: boolean (or + Expression with resultType boolean). + :type sql_writer_use_table_lock: any + :param write_behavior: White behavior when copying data into azure SQL MI. Type: + SqlWriteBehaviorEnum (or Expression with resultType SqlWriteBehaviorEnum). + :type write_behavior: any + :param upsert_settings: SQL upsert settings. + :type upsert_settings: ~azure.mgmt.datafactory.models.SqlUpsertSettings """ _validation = { @@ -32902,6 +33346,9 @@ class SqlMISink(CopySink): 'stored_procedure_parameters': {'key': 'storedProcedureParameters', 'type': '{StoredProcedureParameter}'}, 'stored_procedure_table_type_parameter_name': {'key': 'storedProcedureTableTypeParameterName', 'type': 'object'}, 'table_option': {'key': 'tableOption', 'type': 'object'}, + 'sql_writer_use_table_lock': {'key': 'sqlWriterUseTableLock', 'type': 'object'}, + 'write_behavior': {'key': 'writeBehavior', 'type': 'object'}, + 'upsert_settings': {'key': 'upsertSettings', 'type': 'SqlUpsertSettings'}, } def __init__( @@ -32916,6 +33363,9 @@ def __init__( self.stored_procedure_parameters = kwargs.get('stored_procedure_parameters', None) self.stored_procedure_table_type_parameter_name = kwargs.get('stored_procedure_table_type_parameter_name', None) self.table_option = kwargs.get('table_option', None) + self.sql_writer_use_table_lock = kwargs.get('sql_writer_use_table_lock', None) + self.write_behavior = kwargs.get('write_behavior', None) + self.upsert_settings = kwargs.get('upsert_settings', None) class SqlMISource(TabularSource): @@ -32944,8 +33394,8 @@ class SqlMISource(TabularSource): pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type query_timeout: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any :param sql_reader_query: SQL reader query. Type: string (or Expression with resultType string). :type sql_reader_query: any :param sql_reader_stored_procedure_name: Name of the stored procedure for a Azure SQL Managed @@ -32977,7 +33427,7 @@ class SqlMISource(TabularSource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, 'sql_reader_query': {'key': 'sqlReaderQuery', 'type': 'object'}, 'sql_reader_stored_procedure_name': {'key': 'sqlReaderStoredProcedureName', 'type': 'object'}, 'stored_procedure_parameters': {'key': 'storedProcedureParameters', 'type': '{StoredProcedureParameter}'}, @@ -33148,6 +33598,14 @@ class SqlServerSink(CopySink): :param table_option: The option to handle sink table, such as autoCreate. For now only 'autoCreate' value is supported. Type: string (or Expression with resultType string). :type table_option: any + :param sql_writer_use_table_lock: Whether to use table lock during bulk copy. Type: boolean (or + Expression with resultType boolean). + :type sql_writer_use_table_lock: any + :param write_behavior: Write behavior when copying data into sql server. Type: + SqlWriteBehaviorEnum (or Expression with resultType SqlWriteBehaviorEnum). + :type write_behavior: any + :param upsert_settings: SQL upsert settings. + :type upsert_settings: ~azure.mgmt.datafactory.models.SqlUpsertSettings """ _validation = { @@ -33169,6 +33627,9 @@ class SqlServerSink(CopySink): 'stored_procedure_parameters': {'key': 'storedProcedureParameters', 'type': '{StoredProcedureParameter}'}, 'stored_procedure_table_type_parameter_name': {'key': 'storedProcedureTableTypeParameterName', 'type': 'object'}, 'table_option': {'key': 'tableOption', 'type': 'object'}, + 'sql_writer_use_table_lock': {'key': 'sqlWriterUseTableLock', 'type': 'object'}, + 'write_behavior': {'key': 'writeBehavior', 'type': 'object'}, + 'upsert_settings': {'key': 'upsertSettings', 'type': 'SqlUpsertSettings'}, } def __init__( @@ -33183,6 +33644,9 @@ def __init__( self.stored_procedure_parameters = kwargs.get('stored_procedure_parameters', None) self.stored_procedure_table_type_parameter_name = kwargs.get('stored_procedure_table_type_parameter_name', None) self.table_option = kwargs.get('table_option', None) + self.sql_writer_use_table_lock = kwargs.get('sql_writer_use_table_lock', None) + self.write_behavior = kwargs.get('write_behavior', None) + self.upsert_settings = kwargs.get('upsert_settings', None) class SqlServerSource(TabularSource): @@ -33211,8 +33675,8 @@ class SqlServerSource(TabularSource): pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type query_timeout: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any :param sql_reader_query: SQL reader query. Type: string (or Expression with resultType string). :type sql_reader_query: any :param sql_reader_stored_procedure_name: Name of the stored procedure for a SQL Database @@ -33244,7 +33708,7 @@ class SqlServerSource(TabularSource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, 'sql_reader_query': {'key': 'sqlReaderQuery', 'type': 'object'}, 'sql_reader_stored_procedure_name': {'key': 'sqlReaderStoredProcedureName', 'type': 'object'}, 'stored_procedure_parameters': {'key': 'storedProcedureParameters', 'type': '{StoredProcedureParameter}'}, @@ -33442,6 +33906,14 @@ class SqlSink(CopySink): :param table_option: The option to handle sink table, such as autoCreate. For now only 'autoCreate' value is supported. Type: string (or Expression with resultType string). :type table_option: any + :param sql_writer_use_table_lock: Whether to use table lock during bulk copy. Type: boolean (or + Expression with resultType boolean). + :type sql_writer_use_table_lock: any + :param write_behavior: Write behavior when copying data into sql. Type: SqlWriteBehaviorEnum + (or Expression with resultType SqlWriteBehaviorEnum). + :type write_behavior: any + :param upsert_settings: SQL upsert settings. + :type upsert_settings: ~azure.mgmt.datafactory.models.SqlUpsertSettings """ _validation = { @@ -33463,6 +33935,9 @@ class SqlSink(CopySink): 'stored_procedure_parameters': {'key': 'storedProcedureParameters', 'type': '{StoredProcedureParameter}'}, 'stored_procedure_table_type_parameter_name': {'key': 'storedProcedureTableTypeParameterName', 'type': 'object'}, 'table_option': {'key': 'tableOption', 'type': 'object'}, + 'sql_writer_use_table_lock': {'key': 'sqlWriterUseTableLock', 'type': 'object'}, + 'write_behavior': {'key': 'writeBehavior', 'type': 'object'}, + 'upsert_settings': {'key': 'upsertSettings', 'type': 'SqlUpsertSettings'}, } def __init__( @@ -33477,6 +33952,9 @@ def __init__( self.stored_procedure_parameters = kwargs.get('stored_procedure_parameters', None) self.stored_procedure_table_type_parameter_name = kwargs.get('stored_procedure_table_type_parameter_name', None) self.table_option = kwargs.get('table_option', None) + self.sql_writer_use_table_lock = kwargs.get('sql_writer_use_table_lock', None) + self.write_behavior = kwargs.get('write_behavior', None) + self.upsert_settings = kwargs.get('upsert_settings', None) class SqlSource(TabularSource): @@ -33505,8 +33983,8 @@ class SqlSource(TabularSource): pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type query_timeout: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any :param sql_reader_query: SQL reader query. Type: string (or Expression with resultType string). :type sql_reader_query: any :param sql_reader_stored_procedure_name: Name of the stored procedure for a SQL Database @@ -33540,7 +34018,7 @@ class SqlSource(TabularSource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, 'sql_reader_query': {'key': 'sqlReaderQuery', 'type': 'object'}, 'sql_reader_stored_procedure_name': {'key': 'sqlReaderStoredProcedureName', 'type': 'object'}, 'stored_procedure_parameters': {'key': 'storedProcedureParameters', 'type': '{StoredProcedureParameter}'}, @@ -33563,6 +34041,36 @@ def __init__( self.partition_settings = kwargs.get('partition_settings', None) +class SqlUpsertSettings(msrest.serialization.Model): + """Sql upsert option settings. + + :param use_temp_db: Specifies whether to use temp db for upsert interim table. Type: boolean + (or Expression with resultType boolean). + :type use_temp_db: any + :param interim_schema_name: Schema name for interim table. Type: string (or Expression with + resultType string). + :type interim_schema_name: any + :param keys: Key column names for unique row identification. Type: array of strings (or + Expression with resultType array of strings). + :type keys: any + """ + + _attribute_map = { + 'use_temp_db': {'key': 'useTempDB', 'type': 'object'}, + 'interim_schema_name': {'key': 'interimSchemaName', 'type': 'object'}, + 'keys': {'key': 'keys', 'type': 'object'}, + } + + def __init__( + self, + **kwargs + ): + super(SqlUpsertSettings, self).__init__(**kwargs) + self.use_temp_db = kwargs.get('use_temp_db', None) + self.interim_schema_name = kwargs.get('interim_schema_name', None) + self.keys = kwargs.get('keys', None) + + class SquareLinkedService(LinkedService): """Square Service linked service. @@ -33732,8 +34240,8 @@ class SquareSource(TabularSource): pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type query_timeout: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any :param query: A query to retrieve data from source. Type: string (or Expression with resultType string). :type query: any @@ -33751,7 +34259,7 @@ class SquareSource(TabularSource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, 'query': {'key': 'query', 'type': 'object'}, } @@ -34697,8 +35205,8 @@ class SybaseSource(TabularSource): pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type query_timeout: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any :param query: Database query. Type: string (or Expression with resultType string). :type query: any """ @@ -34715,7 +35223,7 @@ class SybaseSource(TabularSource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, 'query': {'key': 'query', 'type': 'object'}, } @@ -35051,8 +35559,8 @@ class TeradataSource(TabularSource): pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type query_timeout: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any :param query: Teradata query. Type: string (or Expression with resultType string). :type query: any :param partition_option: The partition mechanism that will be used for teradata read in @@ -35075,7 +35583,7 @@ class TeradataSource(TabularSource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, 'query': {'key': 'query', 'type': 'object'}, 'partition_option': {'key': 'partitionOption', 'type': 'object'}, 'partition_settings': {'key': 'partitionSettings', 'type': 'TeradataPartitionSettings'}, @@ -36132,8 +36640,8 @@ class VerticaSource(TabularSource): pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type query_timeout: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any :param query: A query to retrieve data from source. Type: string (or Expression with resultType string). :type query: any @@ -36151,7 +36659,7 @@ class VerticaSource(TabularSource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, 'query': {'key': 'query', 'type': 'object'}, } @@ -36722,8 +37230,8 @@ class WebSource(CopySource): false. Type: boolean (or Expression with resultType boolean). :type disable_metrics_collection: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any """ _validation = { @@ -36737,7 +37245,7 @@ class WebSource(CopySource): 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, } def __init__( @@ -36814,6 +37322,49 @@ def __init__( self.path = kwargs.get('path', None) +class WranglingDataFlow(DataFlow): + """Power Query data flow. + + All required parameters must be populated in order to send to Azure. + + :param type: Required. Type of data flow.Constant filled by server. + :type type: str + :param description: The description of the data flow. + :type description: str + :param annotations: List of tags that can be used for describing the data flow. + :type annotations: list[any] + :param folder: The folder that this data flow is in. If not specified, Data flow will appear at + the root level. + :type folder: ~azure.mgmt.datafactory.models.DataFlowFolder + :param sources: List of sources in Power Query. + :type sources: list[~azure.mgmt.datafactory.models.PowerQuerySource] + :param script: Power query mashup script. + :type script: str + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'type': {'key': 'type', 'type': 'str'}, + 'description': {'key': 'description', 'type': 'str'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'folder': {'key': 'folder', 'type': 'DataFlowFolder'}, + 'sources': {'key': 'typeProperties.sources', 'type': '[PowerQuerySource]'}, + 'script': {'key': 'typeProperties.script', 'type': 'str'}, + } + + def __init__( + self, + **kwargs + ): + super(WranglingDataFlow, self).__init__(**kwargs) + self.type = 'WranglingDataFlow' # type: str + self.sources = kwargs.get('sources', None) + self.script = kwargs.get('script', None) + + class XeroLinkedService(LinkedService): """Xero Service linked service. @@ -36980,8 +37531,8 @@ class XeroSource(TabularSource): pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type query_timeout: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any :param query: A query to retrieve data from source. Type: string (or Expression with resultType string). :type query: any @@ -36999,7 +37550,7 @@ class XeroSource(TabularSource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, 'query': {'key': 'query', 'type': 'object'}, } @@ -37168,8 +37719,8 @@ class XmlSource(CopySource): :param format_settings: Xml format settings. :type format_settings: ~azure.mgmt.datafactory.models.XmlReadSettings :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any """ _validation = { @@ -37185,7 +37736,7 @@ class XmlSource(CopySource): 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'store_settings': {'key': 'storeSettings', 'type': 'StoreReadSettings'}, 'format_settings': {'key': 'formatSettings', 'type': 'XmlReadSettings'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, } def __init__( @@ -37393,8 +37944,8 @@ class ZohoSource(TabularSource): pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type query_timeout: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any :param query: A query to retrieve data from source. Type: string (or Expression with resultType string). :type query: any @@ -37412,7 +37963,7 @@ class ZohoSource(TabularSource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, 'query': {'key': 'query', 'type': 'object'}, } diff --git a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/_models_py3.py b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/_models_py3.py index 3fce09934332..83a693a2c5e4 100644 --- a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/_models_py3.py +++ b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/_models_py3.py @@ -49,7 +49,7 @@ class Activity(msrest.serialization.Model): """A pipeline activity. You probably want to use the sub-classes and not this class directly. Known - sub-classes are: ControlActivity, ExecutionActivity. + sub-classes are: ControlActivity, ExecuteWranglingDataflowActivity, ExecutionActivity. All required parameters must be populated in order to send to Azure. @@ -83,7 +83,7 @@ class Activity(msrest.serialization.Model): } _subtype_map = { - 'type': {'Container': 'ControlActivity', 'Execution': 'ExecutionActivity'} + 'type': {'Container': 'ControlActivity', 'ExecuteWranglingDataflow': 'ExecuteWranglingDataflowActivity', 'Execution': 'ExecutionActivity'} } def __init__( @@ -780,8 +780,8 @@ class TabularSource(CopySource): pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type query_timeout: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any """ _validation = { @@ -796,7 +796,7 @@ class TabularSource(CopySource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, } _subtype_map = { @@ -812,7 +812,7 @@ def __init__( max_concurrent_connections: Optional[Any] = None, disable_metrics_collection: Optional[Any] = None, query_timeout: Optional[Any] = None, - additional_columns: Optional[List["AdditionalColumns"]] = None, + additional_columns: Optional[Any] = None, **kwargs ): super(TabularSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, max_concurrent_connections=max_concurrent_connections, disable_metrics_collection=disable_metrics_collection, **kwargs) @@ -847,8 +847,8 @@ class AmazonMWSSource(TabularSource): pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type query_timeout: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any :param query: A query to retrieve data from source. Type: string (or Expression with resultType string). :type query: any @@ -866,7 +866,7 @@ class AmazonMWSSource(TabularSource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, 'query': {'key': 'query', 'type': 'object'}, } @@ -879,7 +879,7 @@ def __init__( max_concurrent_connections: Optional[Any] = None, disable_metrics_collection: Optional[Any] = None, query_timeout: Optional[Any] = None, - additional_columns: Optional[List["AdditionalColumns"]] = None, + additional_columns: Optional[Any] = None, query: Optional[Any] = None, **kwargs ): @@ -999,8 +999,8 @@ class AmazonRedshiftSource(TabularSource): pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type query_timeout: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any :param query: Database query. Type: string (or Expression with resultType string). :type query: any :param redshift_unload_settings: The Amazon S3 settings needed for the interim Amazon S3 when @@ -1021,7 +1021,7 @@ class AmazonRedshiftSource(TabularSource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, 'query': {'key': 'query', 'type': 'object'}, 'redshift_unload_settings': {'key': 'redshiftUnloadSettings', 'type': 'RedshiftUnloadSettings'}, } @@ -1035,7 +1035,7 @@ def __init__( max_concurrent_connections: Optional[Any] = None, disable_metrics_collection: Optional[Any] = None, query_timeout: Optional[Any] = None, - additional_columns: Optional[List["AdditionalColumns"]] = None, + additional_columns: Optional[Any] = None, query: Optional[Any] = None, redshift_unload_settings: Optional["RedshiftUnloadSettings"] = None, **kwargs @@ -2288,8 +2288,8 @@ class AvroSource(CopySource): :param store_settings: Avro store settings. :type store_settings: ~azure.mgmt.datafactory.models.StoreReadSettings :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any """ _validation = { @@ -2304,7 +2304,7 @@ class AvroSource(CopySource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'store_settings': {'key': 'storeSettings', 'type': 'StoreReadSettings'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, } def __init__( @@ -2316,7 +2316,7 @@ def __init__( max_concurrent_connections: Optional[Any] = None, disable_metrics_collection: Optional[Any] = None, store_settings: Optional["StoreReadSettings"] = None, - additional_columns: Optional[List["AdditionalColumns"]] = None, + additional_columns: Optional[Any] = None, **kwargs ): super(AvroSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, max_concurrent_connections=max_concurrent_connections, disable_metrics_collection=disable_metrics_collection, **kwargs) @@ -4515,8 +4515,8 @@ class AzureDataExplorerSource(CopySource): pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])).. :type query_timeout: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any """ _validation = { @@ -4534,7 +4534,7 @@ class AzureDataExplorerSource(CopySource): 'query': {'key': 'query', 'type': 'object'}, 'no_truncation': {'key': 'noTruncation', 'type': 'object'}, 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, } def __init__( @@ -4548,7 +4548,7 @@ def __init__( disable_metrics_collection: Optional[Any] = None, no_truncation: Optional[Any] = None, query_timeout: Optional[Any] = None, - additional_columns: Optional[List["AdditionalColumns"]] = None, + additional_columns: Optional[Any] = None, **kwargs ): super(AzureDataExplorerSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, max_concurrent_connections=max_concurrent_connections, disable_metrics_collection=disable_metrics_collection, **kwargs) @@ -5934,8 +5934,8 @@ class AzureMariaDBSource(TabularSource): pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type query_timeout: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any :param query: A query to retrieve data from source. Type: string (or Expression with resultType string). :type query: any @@ -5953,7 +5953,7 @@ class AzureMariaDBSource(TabularSource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, 'query': {'key': 'query', 'type': 'object'}, } @@ -5966,7 +5966,7 @@ def __init__( max_concurrent_connections: Optional[Any] = None, disable_metrics_collection: Optional[Any] = None, query_timeout: Optional[Any] = None, - additional_columns: Optional[List["AdditionalColumns"]] = None, + additional_columns: Optional[Any] = None, query: Optional[Any] = None, **kwargs ): @@ -6698,8 +6698,8 @@ class AzureMySqlSource(TabularSource): pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type query_timeout: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any :param query: Database query. Type: string (or Expression with resultType string). :type query: any """ @@ -6716,7 +6716,7 @@ class AzureMySqlSource(TabularSource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, 'query': {'key': 'query', 'type': 'object'}, } @@ -6729,7 +6729,7 @@ def __init__( max_concurrent_connections: Optional[Any] = None, disable_metrics_collection: Optional[Any] = None, query_timeout: Optional[Any] = None, - additional_columns: Optional[List["AdditionalColumns"]] = None, + additional_columns: Optional[Any] = None, query: Optional[Any] = None, **kwargs ): @@ -6971,8 +6971,8 @@ class AzurePostgreSqlSource(TabularSource): pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type query_timeout: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any :param query: A query to retrieve data from source. Type: string (or Expression with resultType string). :type query: any @@ -6990,7 +6990,7 @@ class AzurePostgreSqlSource(TabularSource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, 'query': {'key': 'query', 'type': 'object'}, } @@ -7003,7 +7003,7 @@ def __init__( max_concurrent_connections: Optional[Any] = None, disable_metrics_collection: Optional[Any] = None, query_timeout: Optional[Any] = None, - additional_columns: Optional[List["AdditionalColumns"]] = None, + additional_columns: Optional[Any] = None, query: Optional[Any] = None, **kwargs ): @@ -7863,6 +7863,14 @@ class AzureSqlSink(CopySink): :param table_option: The option to handle sink table, such as autoCreate. For now only 'autoCreate' value is supported. Type: string (or Expression with resultType string). :type table_option: any + :param sql_writer_use_table_lock: Whether to use table lock during bulk copy. Type: boolean (or + Expression with resultType boolean). + :type sql_writer_use_table_lock: any + :param write_behavior: Write behavior when copying data into Azure SQL. Type: + SqlWriteBehaviorEnum (or Expression with resultType SqlWriteBehaviorEnum). + :type write_behavior: any + :param upsert_settings: SQL upsert settings. + :type upsert_settings: ~azure.mgmt.datafactory.models.SqlUpsertSettings """ _validation = { @@ -7884,6 +7892,9 @@ class AzureSqlSink(CopySink): 'stored_procedure_parameters': {'key': 'storedProcedureParameters', 'type': '{StoredProcedureParameter}'}, 'stored_procedure_table_type_parameter_name': {'key': 'storedProcedureTableTypeParameterName', 'type': 'object'}, 'table_option': {'key': 'tableOption', 'type': 'object'}, + 'sql_writer_use_table_lock': {'key': 'sqlWriterUseTableLock', 'type': 'object'}, + 'write_behavior': {'key': 'writeBehavior', 'type': 'object'}, + 'upsert_settings': {'key': 'upsertSettings', 'type': 'SqlUpsertSettings'}, } def __init__( @@ -7902,6 +7913,9 @@ def __init__( stored_procedure_parameters: Optional[Dict[str, "StoredProcedureParameter"]] = None, stored_procedure_table_type_parameter_name: Optional[Any] = None, table_option: Optional[Any] = None, + sql_writer_use_table_lock: Optional[Any] = None, + write_behavior: Optional[Any] = None, + upsert_settings: Optional["SqlUpsertSettings"] = None, **kwargs ): super(AzureSqlSink, self).__init__(additional_properties=additional_properties, write_batch_size=write_batch_size, write_batch_timeout=write_batch_timeout, sink_retry_count=sink_retry_count, sink_retry_wait=sink_retry_wait, max_concurrent_connections=max_concurrent_connections, disable_metrics_collection=disable_metrics_collection, **kwargs) @@ -7912,6 +7926,9 @@ def __init__( self.stored_procedure_parameters = stored_procedure_parameters self.stored_procedure_table_type_parameter_name = stored_procedure_table_type_parameter_name self.table_option = table_option + self.sql_writer_use_table_lock = sql_writer_use_table_lock + self.write_behavior = write_behavior + self.upsert_settings = upsert_settings class AzureSqlSource(TabularSource): @@ -7940,8 +7957,8 @@ class AzureSqlSource(TabularSource): pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type query_timeout: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any :param sql_reader_query: SQL reader query. Type: string (or Expression with resultType string). :type sql_reader_query: any :param sql_reader_stored_procedure_name: Name of the stored procedure for a SQL Database @@ -7973,7 +7990,7 @@ class AzureSqlSource(TabularSource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, 'sql_reader_query': {'key': 'sqlReaderQuery', 'type': 'object'}, 'sql_reader_stored_procedure_name': {'key': 'sqlReaderStoredProcedureName', 'type': 'object'}, 'stored_procedure_parameters': {'key': 'storedProcedureParameters', 'type': '{StoredProcedureParameter}'}, @@ -7991,7 +8008,7 @@ def __init__( max_concurrent_connections: Optional[Any] = None, disable_metrics_collection: Optional[Any] = None, query_timeout: Optional[Any] = None, - additional_columns: Optional[List["AdditionalColumns"]] = None, + additional_columns: Optional[Any] = None, sql_reader_query: Optional[Any] = None, sql_reader_stored_procedure_name: Optional[Any] = None, stored_procedure_parameters: Optional[Dict[str, "StoredProcedureParameter"]] = None, @@ -8348,8 +8365,8 @@ class AzureTableSource(TabularSource): pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type query_timeout: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any :param azure_table_source_query: Azure Table source query. Type: string (or Expression with resultType string). :type azure_table_source_query: any @@ -8370,7 +8387,7 @@ class AzureTableSource(TabularSource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, 'azure_table_source_query': {'key': 'azureTableSourceQuery', 'type': 'object'}, 'azure_table_source_ignore_table_not_found': {'key': 'azureTableSourceIgnoreTableNotFound', 'type': 'object'}, } @@ -8384,7 +8401,7 @@ def __init__( max_concurrent_connections: Optional[Any] = None, disable_metrics_collection: Optional[Any] = None, query_timeout: Optional[Any] = None, - additional_columns: Optional[List["AdditionalColumns"]] = None, + additional_columns: Optional[Any] = None, azure_table_source_query: Optional[Any] = None, azure_table_source_ignore_table_not_found: Optional[Any] = None, **kwargs @@ -9281,8 +9298,8 @@ class CassandraSource(TabularSource): pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type query_timeout: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any :param query: Database query. Should be a SQL-92 query expression or Cassandra Query Language (CQL) command. Type: string (or Expression with resultType string). :type query: any @@ -9308,7 +9325,7 @@ class CassandraSource(TabularSource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, 'query': {'key': 'query', 'type': 'object'}, 'consistency_level': {'key': 'consistencyLevel', 'type': 'str'}, } @@ -9322,7 +9339,7 @@ def __init__( max_concurrent_connections: Optional[Any] = None, disable_metrics_collection: Optional[Any] = None, query_timeout: Optional[Any] = None, - additional_columns: Optional[List["AdditionalColumns"]] = None, + additional_columns: Optional[Any] = None, query: Optional[Any] = None, consistency_level: Optional[Union[str, "CassandraSourceReadConsistencyLevels"]] = None, **kwargs @@ -9893,8 +9910,8 @@ class CommonDataServiceForAppsSource(CopySource): Service for Apps (online & on-premises). Type: string (or Expression with resultType string). :type query: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any """ _validation = { @@ -9909,7 +9926,7 @@ class CommonDataServiceForAppsSource(CopySource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'query': {'key': 'query', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, } def __init__( @@ -9921,7 +9938,7 @@ def __init__( max_concurrent_connections: Optional[Any] = None, disable_metrics_collection: Optional[Any] = None, query: Optional[Any] = None, - additional_columns: Optional[List["AdditionalColumns"]] = None, + additional_columns: Optional[Any] = None, **kwargs ): super(CommonDataServiceForAppsSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, max_concurrent_connections=max_concurrent_connections, disable_metrics_collection=disable_metrics_collection, **kwargs) @@ -10197,8 +10214,8 @@ class ConcurSource(TabularSource): pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type query_timeout: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any :param query: A query to retrieve data from source. Type: string (or Expression with resultType string). :type query: any @@ -10216,7 +10233,7 @@ class ConcurSource(TabularSource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, 'query': {'key': 'query', 'type': 'object'}, } @@ -10229,7 +10246,7 @@ def __init__( max_concurrent_connections: Optional[Any] = None, disable_metrics_collection: Optional[Any] = None, query_timeout: Optional[Any] = None, - additional_columns: Optional[List["AdditionalColumns"]] = None, + additional_columns: Optional[Any] = None, query: Optional[Any] = None, **kwargs ): @@ -10700,6 +10717,9 @@ class CosmosDbMongoDbApiLinkedService(LinkedService): :type parameters: dict[str, ~azure.mgmt.datafactory.models.ParameterSpecification] :param annotations: List of tags that can be used for describing the linked service. :type annotations: list[any] + :param is_server_version_above32: Whether the CosmosDB (MongoDB API) server version is higher + than 3.2. The default value is false. Type: boolean (or Expression with resultType boolean). + :type is_server_version_above32: any :param connection_string: Required. The CosmosDB (MongoDB API) connection string. Type: string, SecureString or AzureKeyVaultSecretReference. Type: string, SecureString or AzureKeyVaultSecretReference. @@ -10722,6 +10742,7 @@ class CosmosDbMongoDbApiLinkedService(LinkedService): 'description': {'key': 'description', 'type': 'str'}, 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'is_server_version_above32': {'key': 'typeProperties.isServerVersionAbove32', 'type': 'object'}, 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'object'}, 'database': {'key': 'typeProperties.database', 'type': 'object'}, } @@ -10736,10 +10757,12 @@ def __init__( description: Optional[str] = None, parameters: Optional[Dict[str, "ParameterSpecification"]] = None, annotations: Optional[List[Any]] = None, + is_server_version_above32: Optional[Any] = None, **kwargs ): super(CosmosDbMongoDbApiLinkedService, self).__init__(additional_properties=additional_properties, connect_via=connect_via, description=description, parameters=parameters, annotations=annotations, **kwargs) self.type = 'CosmosDbMongoDbApi' # type: str + self.is_server_version_above32 = is_server_version_above32 self.connection_string = connection_string self.database = database @@ -10849,8 +10872,8 @@ class CosmosDbMongoDbApiSource(CopySource): pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type query_timeout: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any """ _validation = { @@ -10868,7 +10891,7 @@ class CosmosDbMongoDbApiSource(CopySource): 'cursor_methods': {'key': 'cursorMethods', 'type': 'MongoDbCursorMethodsProperties'}, 'batch_size': {'key': 'batchSize', 'type': 'object'}, 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, } def __init__( @@ -10883,7 +10906,7 @@ def __init__( cursor_methods: Optional["MongoDbCursorMethodsProperties"] = None, batch_size: Optional[Any] = None, query_timeout: Optional[Any] = None, - additional_columns: Optional[List["AdditionalColumns"]] = None, + additional_columns: Optional[Any] = None, **kwargs ): super(CosmosDbMongoDbApiSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, max_concurrent_connections=max_concurrent_connections, disable_metrics_collection=disable_metrics_collection, **kwargs) @@ -11066,8 +11089,8 @@ class CosmosDbSqlApiSource(CopySource): Expression with resultType boolean). :type detect_datetime: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any """ _validation = { @@ -11085,7 +11108,7 @@ class CosmosDbSqlApiSource(CopySource): 'page_size': {'key': 'pageSize', 'type': 'object'}, 'preferred_regions': {'key': 'preferredRegions', 'type': 'object'}, 'detect_datetime': {'key': 'detectDatetime', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, } def __init__( @@ -11100,7 +11123,7 @@ def __init__( page_size: Optional[Any] = None, preferred_regions: Optional[Any] = None, detect_datetime: Optional[Any] = None, - additional_columns: Optional[List["AdditionalColumns"]] = None, + additional_columns: Optional[Any] = None, **kwargs ): super(CosmosDbSqlApiSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, max_concurrent_connections=max_concurrent_connections, disable_metrics_collection=disable_metrics_collection, **kwargs) @@ -11203,8 +11226,8 @@ class CouchbaseSource(TabularSource): pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type query_timeout: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any :param query: A query to retrieve data from source. Type: string (or Expression with resultType string). :type query: any @@ -11222,7 +11245,7 @@ class CouchbaseSource(TabularSource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, 'query': {'key': 'query', 'type': 'object'}, } @@ -11235,7 +11258,7 @@ def __init__( max_concurrent_connections: Optional[Any] = None, disable_metrics_collection: Optional[Any] = None, query_timeout: Optional[Any] = None, - additional_columns: Optional[List["AdditionalColumns"]] = None, + additional_columns: Optional[Any] = None, query: Optional[Any] = None, **kwargs ): @@ -12162,7 +12185,7 @@ class DataFlow(msrest.serialization.Model): """Azure Data Factory nested object which contains a flow with data movements and transformations. You probably want to use the sub-classes and not this class directly. Known - sub-classes are: MappingDataFlow. + sub-classes are: MappingDataFlow, WranglingDataFlow. All required parameters must be populated in order to send to Azure. @@ -12189,7 +12212,7 @@ class DataFlow(msrest.serialization.Model): } _subtype_map = { - 'type': {'MappingDataFlow': 'MappingDataFlow'} + 'type': {'MappingDataFlow': 'MappingDataFlow', 'WranglingDataFlow': 'WranglingDataFlow'} } def __init__( @@ -12944,8 +12967,9 @@ class DatasetCompression(msrest.serialization.Model): :param additional_properties: Unmatched properties from the message are deserialized to this collection. :type additional_properties: dict[str, any] - :param type: Required. Type of dataset compression.Constant filled by server. - :type type: str + :param type: Required. Type of dataset compression. Type: string (or Expression with resultType + string).Constant filled by server. + :type type: any """ _validation = { @@ -12954,7 +12978,7 @@ class DatasetCompression(msrest.serialization.Model): _attribute_map = { 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, + 'type': {'key': 'type', 'type': 'object'}, } _subtype_map = { @@ -12980,8 +13004,9 @@ class DatasetBZip2Compression(DatasetCompression): :param additional_properties: Unmatched properties from the message are deserialized to this collection. :type additional_properties: dict[str, any] - :param type: Required. Type of dataset compression.Constant filled by server. - :type type: str + :param type: Required. Type of dataset compression. Type: string (or Expression with resultType + string).Constant filled by server. + :type type: any """ _validation = { @@ -12990,7 +13015,7 @@ class DatasetBZip2Compression(DatasetCompression): _attribute_map = { 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, + 'type': {'key': 'type', 'type': 'object'}, } def __init__( @@ -13068,8 +13093,9 @@ class DatasetDeflateCompression(DatasetCompression): :param additional_properties: Unmatched properties from the message are deserialized to this collection. :type additional_properties: dict[str, any] - :param type: Required. Type of dataset compression.Constant filled by server. - :type type: str + :param type: Required. Type of dataset compression. Type: string (or Expression with resultType + string).Constant filled by server. + :type type: any :param level: The Deflate compression level. :type level: any """ @@ -13080,7 +13106,7 @@ class DatasetDeflateCompression(DatasetCompression): _attribute_map = { 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, + 'type': {'key': 'type', 'type': 'object'}, 'level': {'key': 'level', 'type': 'object'}, } @@ -13125,8 +13151,9 @@ class DatasetGZipCompression(DatasetCompression): :param additional_properties: Unmatched properties from the message are deserialized to this collection. :type additional_properties: dict[str, any] - :param type: Required. Type of dataset compression.Constant filled by server. - :type type: str + :param type: Required. Type of dataset compression. Type: string (or Expression with resultType + string).Constant filled by server. + :type type: any :param level: The GZip compression level. :type level: any """ @@ -13137,7 +13164,7 @@ class DatasetGZipCompression(DatasetCompression): _attribute_map = { 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, + 'type': {'key': 'type', 'type': 'object'}, 'level': {'key': 'level', 'type': 'object'}, } @@ -13310,8 +13337,9 @@ class DatasetTarCompression(DatasetCompression): :param additional_properties: Unmatched properties from the message are deserialized to this collection. :type additional_properties: dict[str, any] - :param type: Required. Type of dataset compression.Constant filled by server. - :type type: str + :param type: Required. Type of dataset compression. Type: string (or Expression with resultType + string).Constant filled by server. + :type type: any """ _validation = { @@ -13320,7 +13348,7 @@ class DatasetTarCompression(DatasetCompression): _attribute_map = { 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, + 'type': {'key': 'type', 'type': 'object'}, } def __init__( @@ -13341,8 +13369,9 @@ class DatasetTarGZipCompression(DatasetCompression): :param additional_properties: Unmatched properties from the message are deserialized to this collection. :type additional_properties: dict[str, any] - :param type: Required. Type of dataset compression.Constant filled by server. - :type type: str + :param type: Required. Type of dataset compression. Type: string (or Expression with resultType + string).Constant filled by server. + :type type: any :param level: The TarGZip compression level. :type level: any """ @@ -13353,7 +13382,7 @@ class DatasetTarGZipCompression(DatasetCompression): _attribute_map = { 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, + 'type': {'key': 'type', 'type': 'object'}, 'level': {'key': 'level', 'type': 'object'}, } @@ -13377,8 +13406,9 @@ class DatasetZipDeflateCompression(DatasetCompression): :param additional_properties: Unmatched properties from the message are deserialized to this collection. :type additional_properties: dict[str, any] - :param type: Required. Type of dataset compression.Constant filled by server. - :type type: str + :param type: Required. Type of dataset compression. Type: string (or Expression with resultType + string).Constant filled by server. + :type type: any :param level: The ZipDeflate compression level. :type level: any """ @@ -13389,7 +13419,7 @@ class DatasetZipDeflateCompression(DatasetCompression): _attribute_map = { 'additional_properties': {'key': '', 'type': '{object}'}, - 'type': {'key': 'type', 'type': 'str'}, + 'type': {'key': 'type', 'type': 'object'}, 'level': {'key': 'level', 'type': 'object'}, } @@ -13534,8 +13564,8 @@ class Db2Source(TabularSource): pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type query_timeout: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any :param query: Database query. Type: string (or Expression with resultType string). :type query: any """ @@ -13552,7 +13582,7 @@ class Db2Source(TabularSource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, 'query': {'key': 'query', 'type': 'object'}, } @@ -13565,7 +13595,7 @@ def __init__( max_concurrent_connections: Optional[Any] = None, disable_metrics_collection: Optional[Any] = None, query_timeout: Optional[Any] = None, - additional_columns: Optional[List["AdditionalColumns"]] = None, + additional_columns: Optional[Any] = None, query: Optional[Any] = None, **kwargs ): @@ -14028,8 +14058,8 @@ class DelimitedTextSource(CopySource): :param format_settings: DelimitedText format settings. :type format_settings: ~azure.mgmt.datafactory.models.DelimitedTextReadSettings :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any """ _validation = { @@ -14045,7 +14075,7 @@ class DelimitedTextSource(CopySource): 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'store_settings': {'key': 'storeSettings', 'type': 'StoreReadSettings'}, 'format_settings': {'key': 'formatSettings', 'type': 'DelimitedTextReadSettings'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, } def __init__( @@ -14058,7 +14088,7 @@ def __init__( disable_metrics_collection: Optional[Any] = None, store_settings: Optional["StoreReadSettings"] = None, format_settings: Optional["DelimitedTextReadSettings"] = None, - additional_columns: Optional[List["AdditionalColumns"]] = None, + additional_columns: Optional[Any] = None, **kwargs ): super(DelimitedTextSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, max_concurrent_connections=max_concurrent_connections, disable_metrics_collection=disable_metrics_collection, **kwargs) @@ -14373,8 +14403,8 @@ class DocumentDbCollectionSource(CopySource): pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type query_timeout: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any """ _validation = { @@ -14391,7 +14421,7 @@ class DocumentDbCollectionSource(CopySource): 'query': {'key': 'query', 'type': 'object'}, 'nesting_separator': {'key': 'nestingSeparator', 'type': 'object'}, 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, } def __init__( @@ -14405,7 +14435,7 @@ def __init__( query: Optional[Any] = None, nesting_separator: Optional[Any] = None, query_timeout: Optional[Any] = None, - additional_columns: Optional[List["AdditionalColumns"]] = None, + additional_columns: Optional[Any] = None, **kwargs ): super(DocumentDbCollectionSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, max_concurrent_connections=max_concurrent_connections, disable_metrics_collection=disable_metrics_collection, **kwargs) @@ -14507,8 +14537,8 @@ class DrillSource(TabularSource): pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type query_timeout: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any :param query: A query to retrieve data from source. Type: string (or Expression with resultType string). :type query: any @@ -14526,7 +14556,7 @@ class DrillSource(TabularSource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, 'query': {'key': 'query', 'type': 'object'}, } @@ -14539,7 +14569,7 @@ def __init__( max_concurrent_connections: Optional[Any] = None, disable_metrics_collection: Optional[Any] = None, query_timeout: Optional[Any] = None, - additional_columns: Optional[List["AdditionalColumns"]] = None, + additional_columns: Optional[Any] = None, query: Optional[Any] = None, **kwargs ): @@ -14873,8 +14903,8 @@ class DynamicsAXSource(TabularSource): pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type query_timeout: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any :param query: A query to retrieve data from source. Type: string (or Expression with resultType string). :type query: any @@ -14897,7 +14927,7 @@ class DynamicsAXSource(TabularSource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, 'query': {'key': 'query', 'type': 'object'}, 'http_request_timeout': {'key': 'httpRequestTimeout', 'type': 'object'}, } @@ -14911,7 +14941,7 @@ def __init__( max_concurrent_connections: Optional[Any] = None, disable_metrics_collection: Optional[Any] = None, query_timeout: Optional[Any] = None, - additional_columns: Optional[List["AdditionalColumns"]] = None, + additional_columns: Optional[Any] = None, query: Optional[Any] = None, http_request_timeout: Optional[Any] = None, **kwargs @@ -15229,8 +15259,8 @@ class DynamicsCrmSource(CopySource): (online & on-premises). Type: string (or Expression with resultType string). :type query: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any """ _validation = { @@ -15245,7 +15275,7 @@ class DynamicsCrmSource(CopySource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'query': {'key': 'query', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, } def __init__( @@ -15257,7 +15287,7 @@ def __init__( max_concurrent_connections: Optional[Any] = None, disable_metrics_collection: Optional[Any] = None, query: Optional[Any] = None, - additional_columns: Optional[List["AdditionalColumns"]] = None, + additional_columns: Optional[Any] = None, **kwargs ): super(DynamicsCrmSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, max_concurrent_connections=max_concurrent_connections, disable_metrics_collection=disable_metrics_collection, **kwargs) @@ -15572,8 +15602,8 @@ class DynamicsSource(CopySource): (online & on-premises). Type: string (or Expression with resultType string). :type query: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any """ _validation = { @@ -15588,7 +15618,7 @@ class DynamicsSource(CopySource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'query': {'key': 'query', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, } def __init__( @@ -15600,7 +15630,7 @@ def __init__( max_concurrent_connections: Optional[Any] = None, disable_metrics_collection: Optional[Any] = None, query: Optional[Any] = None, - additional_columns: Optional[List["AdditionalColumns"]] = None, + additional_columns: Optional[Any] = None, **kwargs ): super(DynamicsSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, max_concurrent_connections=max_concurrent_connections, disable_metrics_collection=disable_metrics_collection, **kwargs) @@ -15794,8 +15824,8 @@ class EloquaSource(TabularSource): pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type query_timeout: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any :param query: A query to retrieve data from source. Type: string (or Expression with resultType string). :type query: any @@ -15813,7 +15843,7 @@ class EloquaSource(TabularSource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, 'query': {'key': 'query', 'type': 'object'}, } @@ -15826,7 +15856,7 @@ def __init__( max_concurrent_connections: Optional[Any] = None, disable_metrics_collection: Optional[Any] = None, query_timeout: Optional[Any] = None, - additional_columns: Optional[List["AdditionalColumns"]] = None, + additional_columns: Optional[Any] = None, query: Optional[Any] = None, **kwargs ): @@ -16074,8 +16104,8 @@ class ExcelSource(CopySource): :param store_settings: Excel store settings. :type store_settings: ~azure.mgmt.datafactory.models.StoreReadSettings :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any """ _validation = { @@ -16090,7 +16120,7 @@ class ExcelSource(CopySource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'store_settings': {'key': 'storeSettings', 'type': 'StoreReadSettings'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, } def __init__( @@ -16102,7 +16132,7 @@ def __init__( max_concurrent_connections: Optional[Any] = None, disable_metrics_collection: Optional[Any] = None, store_settings: Optional["StoreReadSettings"] = None, - additional_columns: Optional[List["AdditionalColumns"]] = None, + additional_columns: Optional[Any] = None, **kwargs ): super(ExcelSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, max_concurrent_connections=max_concurrent_connections, disable_metrics_collection=disable_metrics_collection, **kwargs) @@ -16207,6 +16237,67 @@ def __init__( self.run_concurrently = run_concurrently +class ExecuteDataFlowActivityTypeProperties(msrest.serialization.Model): + """Execute data flow activity properties. + + All required parameters must be populated in order to send to Azure. + + :param data_flow: Required. Data flow reference. + :type data_flow: ~azure.mgmt.datafactory.models.DataFlowReference + :param staging: Staging info for execute data flow activity. + :type staging: ~azure.mgmt.datafactory.models.DataFlowStagingInfo + :param integration_runtime: The integration runtime reference. + :type integration_runtime: ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :param compute: Compute properties for data flow activity. + :type compute: ~azure.mgmt.datafactory.models.ExecuteDataFlowActivityTypePropertiesCompute + :param trace_level: Trace level setting used for data flow monitoring output. Supported values + are: 'coarse', 'fine', and 'none'. Type: string (or Expression with resultType string). + :type trace_level: any + :param continue_on_error: Continue on error setting used for data flow execution. Enables + processing to continue if a sink fails. Type: boolean (or Expression with resultType boolean). + :type continue_on_error: any + :param run_concurrently: Concurrent run setting used for data flow execution. Allows sinks with + the same save order to be processed concurrently. Type: boolean (or Expression with resultType + boolean). + :type run_concurrently: any + """ + + _validation = { + 'data_flow': {'required': True}, + } + + _attribute_map = { + 'data_flow': {'key': 'dataFlow', 'type': 'DataFlowReference'}, + 'staging': {'key': 'staging', 'type': 'DataFlowStagingInfo'}, + 'integration_runtime': {'key': 'integrationRuntime', 'type': 'IntegrationRuntimeReference'}, + 'compute': {'key': 'compute', 'type': 'ExecuteDataFlowActivityTypePropertiesCompute'}, + 'trace_level': {'key': 'traceLevel', 'type': 'object'}, + 'continue_on_error': {'key': 'continueOnError', 'type': 'object'}, + 'run_concurrently': {'key': 'runConcurrently', 'type': 'object'}, + } + + def __init__( + self, + *, + data_flow: "DataFlowReference", + staging: Optional["DataFlowStagingInfo"] = None, + integration_runtime: Optional["IntegrationRuntimeReference"] = None, + compute: Optional["ExecuteDataFlowActivityTypePropertiesCompute"] = None, + trace_level: Optional[Any] = None, + continue_on_error: Optional[Any] = None, + run_concurrently: Optional[Any] = None, + **kwargs + ): + super(ExecuteDataFlowActivityTypeProperties, self).__init__(**kwargs) + self.data_flow = data_flow + self.staging = staging + self.integration_runtime = integration_runtime + self.compute = compute + self.trace_level = trace_level + self.continue_on_error = continue_on_error + self.run_concurrently = run_concurrently + + class ExecuteDataFlowActivityTypePropertiesCompute(msrest.serialization.Model): """Compute properties for data flow activity. @@ -16301,6 +16392,65 @@ def __init__( self.wait_on_completion = wait_on_completion +class ExecutePowerQueryActivityTypeProperties(ExecuteDataFlowActivityTypeProperties): + """Execute power query data flow activity properties. + + All required parameters must be populated in order to send to Azure. + + :param data_flow: Required. Data flow reference. + :type data_flow: ~azure.mgmt.datafactory.models.DataFlowReference + :param staging: Staging info for execute data flow activity. + :type staging: ~azure.mgmt.datafactory.models.DataFlowStagingInfo + :param integration_runtime: The integration runtime reference. + :type integration_runtime: ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :param compute: Compute properties for data flow activity. + :type compute: ~azure.mgmt.datafactory.models.ExecuteDataFlowActivityTypePropertiesCompute + :param trace_level: Trace level setting used for data flow monitoring output. Supported values + are: 'coarse', 'fine', and 'none'. Type: string (or Expression with resultType string). + :type trace_level: any + :param continue_on_error: Continue on error setting used for data flow execution. Enables + processing to continue if a sink fails. Type: boolean (or Expression with resultType boolean). + :type continue_on_error: any + :param run_concurrently: Concurrent run setting used for data flow execution. Allows sinks with + the same save order to be processed concurrently. Type: boolean (or Expression with resultType + boolean). + :type run_concurrently: any + :param sinks: List of Power Query activity sinks mapped to a queryName. + :type sinks: dict[str, ~azure.mgmt.datafactory.models.PowerQuerySink] + """ + + _validation = { + 'data_flow': {'required': True}, + } + + _attribute_map = { + 'data_flow': {'key': 'dataFlow', 'type': 'DataFlowReference'}, + 'staging': {'key': 'staging', 'type': 'DataFlowStagingInfo'}, + 'integration_runtime': {'key': 'integrationRuntime', 'type': 'IntegrationRuntimeReference'}, + 'compute': {'key': 'compute', 'type': 'ExecuteDataFlowActivityTypePropertiesCompute'}, + 'trace_level': {'key': 'traceLevel', 'type': 'object'}, + 'continue_on_error': {'key': 'continueOnError', 'type': 'object'}, + 'run_concurrently': {'key': 'runConcurrently', 'type': 'object'}, + 'sinks': {'key': 'sinks', 'type': '{PowerQuerySink}'}, + } + + def __init__( + self, + *, + data_flow: "DataFlowReference", + staging: Optional["DataFlowStagingInfo"] = None, + integration_runtime: Optional["IntegrationRuntimeReference"] = None, + compute: Optional["ExecuteDataFlowActivityTypePropertiesCompute"] = None, + trace_level: Optional[Any] = None, + continue_on_error: Optional[Any] = None, + run_concurrently: Optional[Any] = None, + sinks: Optional[Dict[str, "PowerQuerySink"]] = None, + **kwargs + ): + super(ExecutePowerQueryActivityTypeProperties, self).__init__(data_flow=data_flow, staging=staging, integration_runtime=integration_runtime, compute=compute, trace_level=trace_level, continue_on_error=continue_on_error, run_concurrently=run_concurrently, **kwargs) + self.sinks = sinks + + class ExecuteSSISPackageActivity(ExecutionActivity): """Execute SSIS package activity. @@ -16426,6 +16576,104 @@ def __init__( self.log_location = log_location +class ExecuteWranglingDataflowActivity(Activity): + """Execute power query activity. + + All required parameters must be populated in order to send to Azure. + + :param additional_properties: Unmatched properties from the message are deserialized to this + collection. + :type additional_properties: dict[str, any] + :param name: Required. Activity name. + :type name: str + :param type: Required. Type of activity.Constant filled by server. + :type type: str + :param description: Activity description. + :type description: str + :param depends_on: Activity depends on condition. + :type depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] + :param user_properties: Activity user properties. + :type user_properties: list[~azure.mgmt.datafactory.models.UserProperty] + :param policy: Activity policy. + :type policy: ~azure.mgmt.datafactory.models.ActivityPolicy + :param data_flow: Required. Data flow reference. + :type data_flow: ~azure.mgmt.datafactory.models.DataFlowReference + :param staging: Staging info for execute data flow activity. + :type staging: ~azure.mgmt.datafactory.models.DataFlowStagingInfo + :param integration_runtime: The integration runtime reference. + :type integration_runtime: ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :param compute: Compute properties for data flow activity. + :type compute: ~azure.mgmt.datafactory.models.ExecuteDataFlowActivityTypePropertiesCompute + :param trace_level: Trace level setting used for data flow monitoring output. Supported values + are: 'coarse', 'fine', and 'none'. Type: string (or Expression with resultType string). + :type trace_level: any + :param continue_on_error: Continue on error setting used for data flow execution. Enables + processing to continue if a sink fails. Type: boolean (or Expression with resultType boolean). + :type continue_on_error: any + :param run_concurrently: Concurrent run setting used for data flow execution. Allows sinks with + the same save order to be processed concurrently. Type: boolean (or Expression with resultType + boolean). + :type run_concurrently: any + :param sinks: List of Power Query activity sinks mapped to a queryName. + :type sinks: dict[str, ~azure.mgmt.datafactory.models.PowerQuerySink] + """ + + _validation = { + 'name': {'required': True}, + 'type': {'required': True}, + 'data_flow': {'required': True}, + } + + _attribute_map = { + 'additional_properties': {'key': '', 'type': '{object}'}, + 'name': {'key': 'name', 'type': 'str'}, + 'type': {'key': 'type', 'type': 'str'}, + 'description': {'key': 'description', 'type': 'str'}, + 'depends_on': {'key': 'dependsOn', 'type': '[ActivityDependency]'}, + 'user_properties': {'key': 'userProperties', 'type': '[UserProperty]'}, + 'policy': {'key': 'policy', 'type': 'ActivityPolicy'}, + 'data_flow': {'key': 'typeProperties.dataFlow', 'type': 'DataFlowReference'}, + 'staging': {'key': 'typeProperties.staging', 'type': 'DataFlowStagingInfo'}, + 'integration_runtime': {'key': 'typeProperties.integrationRuntime', 'type': 'IntegrationRuntimeReference'}, + 'compute': {'key': 'typeProperties.compute', 'type': 'ExecuteDataFlowActivityTypePropertiesCompute'}, + 'trace_level': {'key': 'typeProperties.traceLevel', 'type': 'object'}, + 'continue_on_error': {'key': 'typeProperties.continueOnError', 'type': 'object'}, + 'run_concurrently': {'key': 'typeProperties.runConcurrently', 'type': 'object'}, + 'sinks': {'key': 'typeProperties.sinks', 'type': '{PowerQuerySink}'}, + } + + def __init__( + self, + *, + name: str, + data_flow: "DataFlowReference", + additional_properties: Optional[Dict[str, Any]] = None, + description: Optional[str] = None, + depends_on: Optional[List["ActivityDependency"]] = None, + user_properties: Optional[List["UserProperty"]] = None, + policy: Optional["ActivityPolicy"] = None, + staging: Optional["DataFlowStagingInfo"] = None, + integration_runtime: Optional["IntegrationRuntimeReference"] = None, + compute: Optional["ExecuteDataFlowActivityTypePropertiesCompute"] = None, + trace_level: Optional[Any] = None, + continue_on_error: Optional[Any] = None, + run_concurrently: Optional[Any] = None, + sinks: Optional[Dict[str, "PowerQuerySink"]] = None, + **kwargs + ): + super(ExecuteWranglingDataflowActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, **kwargs) + self.type = 'ExecuteWranglingDataflow' # type: str + self.policy = policy + self.data_flow = data_flow + self.staging = staging + self.integration_runtime = integration_runtime + self.compute = compute + self.trace_level = trace_level + self.continue_on_error = continue_on_error + self.run_concurrently = run_concurrently + self.sinks = sinks + + class ExposureControlBatchRequest(msrest.serialization.Model): """A list of exposure control features. @@ -16795,6 +17043,10 @@ class FactoryGitHubConfiguration(FactoryRepoConfiguration): :type last_commit_id: str :param host_name: GitHub Enterprise host name. For example: https://github.mydomain.com. :type host_name: str + :param client_id: GitHub bring your own app client id. + :type client_id: str + :param client_secret: GitHub bring your own app client secret information. + :type client_secret: ~azure.mgmt.datafactory.models.GitHubClientSecret """ _validation = { @@ -16813,6 +17065,8 @@ class FactoryGitHubConfiguration(FactoryRepoConfiguration): 'root_folder': {'key': 'rootFolder', 'type': 'str'}, 'last_commit_id': {'key': 'lastCommitId', 'type': 'str'}, 'host_name': {'key': 'hostName', 'type': 'str'}, + 'client_id': {'key': 'clientId', 'type': 'str'}, + 'client_secret': {'key': 'clientSecret', 'type': 'GitHubClientSecret'}, } def __init__( @@ -16824,11 +17078,15 @@ def __init__( root_folder: str, last_commit_id: Optional[str] = None, host_name: Optional[str] = None, + client_id: Optional[str] = None, + client_secret: Optional["GitHubClientSecret"] = None, **kwargs ): super(FactoryGitHubConfiguration, self).__init__(account_name=account_name, repository_name=repository_name, collaboration_branch=collaboration_branch, root_folder=root_folder, last_commit_id=last_commit_id, **kwargs) self.type = 'FactoryGitHubConfiguration' # type: str self.host_name = host_name + self.client_id = client_id + self.client_secret = client_secret class FactoryIdentity(msrest.serialization.Model): @@ -17475,8 +17733,8 @@ class FileSystemSource(CopySource): true. Type: boolean (or Expression with resultType boolean). :type recursive: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any """ _validation = { @@ -17491,7 +17749,7 @@ class FileSystemSource(CopySource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'recursive': {'key': 'recursive', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, } def __init__( @@ -17503,7 +17761,7 @@ def __init__( max_concurrent_connections: Optional[Any] = None, disable_metrics_collection: Optional[Any] = None, recursive: Optional[Any] = None, - additional_columns: Optional[List["AdditionalColumns"]] = None, + additional_columns: Optional[Any] = None, **kwargs ): super(FileSystemSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, max_concurrent_connections=max_concurrent_connections, disable_metrics_collection=disable_metrics_collection, **kwargs) @@ -18333,8 +18591,8 @@ class GoogleAdWordsSource(TabularSource): pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type query_timeout: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any :param query: A query to retrieve data from source. Type: string (or Expression with resultType string). :type query: any @@ -18352,7 +18610,7 @@ class GoogleAdWordsSource(TabularSource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, 'query': {'key': 'query', 'type': 'object'}, } @@ -18365,7 +18623,7 @@ def __init__( max_concurrent_connections: Optional[Any] = None, disable_metrics_collection: Optional[Any] = None, query_timeout: Optional[Any] = None, - additional_columns: Optional[List["AdditionalColumns"]] = None, + additional_columns: Optional[Any] = None, query: Optional[Any] = None, **kwargs ): @@ -18605,8 +18863,8 @@ class GoogleBigQuerySource(TabularSource): pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type query_timeout: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any :param query: A query to retrieve data from source. Type: string (or Expression with resultType string). :type query: any @@ -18624,7 +18882,7 @@ class GoogleBigQuerySource(TabularSource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, 'query': {'key': 'query', 'type': 'object'}, } @@ -18637,7 +18895,7 @@ def __init__( max_concurrent_connections: Optional[Any] = None, disable_metrics_collection: Optional[Any] = None, query_timeout: Optional[Any] = None, - additional_columns: Optional[List["AdditionalColumns"]] = None, + additional_columns: Optional[Any] = None, query: Optional[Any] = None, **kwargs ): @@ -18965,8 +19223,8 @@ class GreenplumSource(TabularSource): pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type query_timeout: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any :param query: A query to retrieve data from source. Type: string (or Expression with resultType string). :type query: any @@ -18984,7 +19242,7 @@ class GreenplumSource(TabularSource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, 'query': {'key': 'query', 'type': 'object'}, } @@ -18997,7 +19255,7 @@ def __init__( max_concurrent_connections: Optional[Any] = None, disable_metrics_collection: Optional[Any] = None, query_timeout: Optional[Any] = None, - additional_columns: Optional[List["AdditionalColumns"]] = None, + additional_columns: Optional[Any] = None, query: Optional[Any] = None, **kwargs ): @@ -19294,8 +19552,8 @@ class HBaseSource(TabularSource): pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type query_timeout: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any :param query: A query to retrieve data from source. Type: string (or Expression with resultType string). :type query: any @@ -19313,7 +19571,7 @@ class HBaseSource(TabularSource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, 'query': {'key': 'query', 'type': 'object'}, } @@ -19326,7 +19584,7 @@ def __init__( max_concurrent_connections: Optional[Any] = None, disable_metrics_collection: Optional[Any] = None, query_timeout: Optional[Any] = None, - additional_columns: Optional[List["AdditionalColumns"]] = None, + additional_columns: Optional[Any] = None, query: Optional[Any] = None, **kwargs ): @@ -20732,8 +20990,8 @@ class HiveSource(TabularSource): pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type query_timeout: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any :param query: A query to retrieve data from source. Type: string (or Expression with resultType string). :type query: any @@ -20751,7 +21009,7 @@ class HiveSource(TabularSource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, 'query': {'key': 'query', 'type': 'object'}, } @@ -20764,7 +21022,7 @@ def __init__( max_concurrent_connections: Optional[Any] = None, disable_metrics_collection: Optional[Any] = None, query_timeout: Optional[Any] = None, - additional_columns: Optional[List["AdditionalColumns"]] = None, + additional_columns: Optional[Any] = None, query: Optional[Any] = None, **kwargs ): @@ -21352,8 +21610,8 @@ class HubspotSource(TabularSource): pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type query_timeout: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any :param query: A query to retrieve data from source. Type: string (or Expression with resultType string). :type query: any @@ -21371,7 +21629,7 @@ class HubspotSource(TabularSource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, 'query': {'key': 'query', 'type': 'object'}, } @@ -21384,7 +21642,7 @@ def __init__( max_concurrent_connections: Optional[Any] = None, disable_metrics_collection: Optional[Any] = None, query_timeout: Optional[Any] = None, - additional_columns: Optional[List["AdditionalColumns"]] = None, + additional_columns: Optional[Any] = None, query: Optional[Any] = None, **kwargs ): @@ -21684,8 +21942,8 @@ class ImpalaSource(TabularSource): pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type query_timeout: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any :param query: A query to retrieve data from source. Type: string (or Expression with resultType string). :type query: any @@ -21703,7 +21961,7 @@ class ImpalaSource(TabularSource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, 'query': {'key': 'query', 'type': 'object'}, } @@ -21716,7 +21974,7 @@ def __init__( max_concurrent_connections: Optional[Any] = None, disable_metrics_collection: Optional[Any] = None, query_timeout: Optional[Any] = None, - additional_columns: Optional[List["AdditionalColumns"]] = None, + additional_columns: Optional[Any] = None, query: Optional[Any] = None, **kwargs ): @@ -21904,8 +22162,8 @@ class InformixSource(TabularSource): pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type query_timeout: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any :param query: Database query. Type: string (or Expression with resultType string). :type query: any """ @@ -21922,7 +22180,7 @@ class InformixSource(TabularSource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, 'query': {'key': 'query', 'type': 'object'}, } @@ -21935,7 +22193,7 @@ def __init__( max_concurrent_connections: Optional[Any] = None, disable_metrics_collection: Optional[Any] = None, query_timeout: Optional[Any] = None, - additional_columns: Optional[List["AdditionalColumns"]] = None, + additional_columns: Optional[Any] = None, query: Optional[Any] = None, **kwargs ): @@ -22246,6 +22504,9 @@ class IntegrationRuntimeDataFlowProperties(msrest.serialization.Model): :param time_to_live: Time to live (in minutes) setting of the cluster which will execute data flow job. :type time_to_live: int + :param cleanup: Cluster will not be recycled and it will be used in next data flow activity run + until TTL (time to live) is reached if this is set as false. Default is true. + :type cleanup: bool """ _validation = { @@ -22257,6 +22518,7 @@ class IntegrationRuntimeDataFlowProperties(msrest.serialization.Model): 'compute_type': {'key': 'computeType', 'type': 'str'}, 'core_count': {'key': 'coreCount', 'type': 'int'}, 'time_to_live': {'key': 'timeToLive', 'type': 'int'}, + 'cleanup': {'key': 'cleanup', 'type': 'bool'}, } def __init__( @@ -22266,6 +22528,7 @@ def __init__( compute_type: Optional[Union[str, "DataFlowComputeType"]] = None, core_count: Optional[int] = None, time_to_live: Optional[int] = None, + cleanup: Optional[bool] = None, **kwargs ): super(IntegrationRuntimeDataFlowProperties, self).__init__(**kwargs) @@ -22273,6 +22536,7 @@ def __init__( self.compute_type = compute_type self.core_count = core_count self.time_to_live = time_to_live + self.cleanup = cleanup class IntegrationRuntimeDataProxyProperties(msrest.serialization.Model): @@ -22488,6 +22752,103 @@ def __init__( self.received_bytes = None +class IntegrationRuntimeOutboundNetworkDependenciesCategoryEndpoint(msrest.serialization.Model): + """Azure-SSIS integration runtime outbound network dependency endpoints for one category. + + :param category: The category of outbound network dependency. + :type category: str + :param endpoints: The endpoints for outbound network dependency. + :type endpoints: + list[~azure.mgmt.datafactory.models.IntegrationRuntimeOutboundNetworkDependenciesEndpoint] + """ + + _attribute_map = { + 'category': {'key': 'category', 'type': 'str'}, + 'endpoints': {'key': 'endpoints', 'type': '[IntegrationRuntimeOutboundNetworkDependenciesEndpoint]'}, + } + + def __init__( + self, + *, + category: Optional[str] = None, + endpoints: Optional[List["IntegrationRuntimeOutboundNetworkDependenciesEndpoint"]] = None, + **kwargs + ): + super(IntegrationRuntimeOutboundNetworkDependenciesCategoryEndpoint, self).__init__(**kwargs) + self.category = category + self.endpoints = endpoints + + +class IntegrationRuntimeOutboundNetworkDependenciesEndpoint(msrest.serialization.Model): + """The endpoint for Azure-SSIS integration runtime outbound network dependency. + + :param domain_name: The domain name of endpoint. + :type domain_name: str + :param endpoint_details: The details of endpoint. + :type endpoint_details: + list[~azure.mgmt.datafactory.models.IntegrationRuntimeOutboundNetworkDependenciesEndpointDetails] + """ + + _attribute_map = { + 'domain_name': {'key': 'domainName', 'type': 'str'}, + 'endpoint_details': {'key': 'endpointDetails', 'type': '[IntegrationRuntimeOutboundNetworkDependenciesEndpointDetails]'}, + } + + def __init__( + self, + *, + domain_name: Optional[str] = None, + endpoint_details: Optional[List["IntegrationRuntimeOutboundNetworkDependenciesEndpointDetails"]] = None, + **kwargs + ): + super(IntegrationRuntimeOutboundNetworkDependenciesEndpoint, self).__init__(**kwargs) + self.domain_name = domain_name + self.endpoint_details = endpoint_details + + +class IntegrationRuntimeOutboundNetworkDependenciesEndpointDetails(msrest.serialization.Model): + """The details of Azure-SSIS integration runtime outbound network dependency endpoint. + + :param port: The port of endpoint. + :type port: int + """ + + _attribute_map = { + 'port': {'key': 'port', 'type': 'int'}, + } + + def __init__( + self, + *, + port: Optional[int] = None, + **kwargs + ): + super(IntegrationRuntimeOutboundNetworkDependenciesEndpointDetails, self).__init__(**kwargs) + self.port = port + + +class IntegrationRuntimeOutboundNetworkDependenciesEndpointsResponse(msrest.serialization.Model): + """Azure-SSIS integration runtime outbound network dependency endpoints. + + :param value: The list of outbound network dependency endpoints. + :type value: + list[~azure.mgmt.datafactory.models.IntegrationRuntimeOutboundNetworkDependenciesCategoryEndpoint] + """ + + _attribute_map = { + 'value': {'key': 'value', 'type': '[IntegrationRuntimeOutboundNetworkDependenciesCategoryEndpoint]'}, + } + + def __init__( + self, + *, + value: Optional[List["IntegrationRuntimeOutboundNetworkDependenciesCategoryEndpoint"]] = None, + **kwargs + ): + super(IntegrationRuntimeOutboundNetworkDependenciesEndpointsResponse, self).__init__(**kwargs) + self.value = value + + class IntegrationRuntimeReference(msrest.serialization.Model): """Integration runtime reference type. @@ -22853,6 +23214,9 @@ class IntegrationRuntimeVNetProperties(msrest.serialization.Model): :param public_i_ps: Resource IDs of the public IP addresses that this integration runtime will use. :type public_i_ps: list[str] + :param subnet_id: The ID of subnet, to which this Azure-SSIS integration runtime will be + joined. + :type subnet_id: str """ _attribute_map = { @@ -22860,6 +23224,7 @@ class IntegrationRuntimeVNetProperties(msrest.serialization.Model): 'v_net_id': {'key': 'vNetId', 'type': 'str'}, 'subnet': {'key': 'subnet', 'type': 'str'}, 'public_i_ps': {'key': 'publicIPs', 'type': '[str]'}, + 'subnet_id': {'key': 'subnetId', 'type': 'str'}, } def __init__( @@ -22869,6 +23234,7 @@ def __init__( v_net_id: Optional[str] = None, subnet: Optional[str] = None, public_i_ps: Optional[List[str]] = None, + subnet_id: Optional[str] = None, **kwargs ): super(IntegrationRuntimeVNetProperties, self).__init__(**kwargs) @@ -22876,6 +23242,7 @@ def __init__( self.v_net_id = v_net_id self.subnet = subnet self.public_i_ps = public_i_ps + self.subnet_id = subnet_id class JiraLinkedService(LinkedService): @@ -23070,8 +23437,8 @@ class JiraSource(TabularSource): pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type query_timeout: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any :param query: A query to retrieve data from source. Type: string (or Expression with resultType string). :type query: any @@ -23089,7 +23456,7 @@ class JiraSource(TabularSource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, 'query': {'key': 'query', 'type': 'object'}, } @@ -23102,7 +23469,7 @@ def __init__( max_concurrent_connections: Optional[Any] = None, disable_metrics_collection: Optional[Any] = None, query_timeout: Optional[Any] = None, - additional_columns: Optional[List["AdditionalColumns"]] = None, + additional_columns: Optional[Any] = None, query: Optional[Any] = None, **kwargs ): @@ -23402,8 +23769,8 @@ class JsonSource(CopySource): :param format_settings: Json format settings. :type format_settings: ~azure.mgmt.datafactory.models.JsonReadSettings :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any """ _validation = { @@ -23419,7 +23786,7 @@ class JsonSource(CopySource): 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'store_settings': {'key': 'storeSettings', 'type': 'StoreReadSettings'}, 'format_settings': {'key': 'formatSettings', 'type': 'JsonReadSettings'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, } def __init__( @@ -23432,7 +23799,7 @@ def __init__( disable_metrics_collection: Optional[Any] = None, store_settings: Optional["StoreReadSettings"] = None, format_settings: Optional["JsonReadSettings"] = None, - additional_columns: Optional[List["AdditionalColumns"]] = None, + additional_columns: Optional[Any] = None, **kwargs ): super(JsonSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, max_concurrent_connections=max_concurrent_connections, disable_metrics_collection=disable_metrics_collection, **kwargs) @@ -24175,8 +24542,8 @@ class MagentoSource(TabularSource): pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type query_timeout: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any :param query: A query to retrieve data from source. Type: string (or Expression with resultType string). :type query: any @@ -24194,7 +24561,7 @@ class MagentoSource(TabularSource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, 'query': {'key': 'query', 'type': 'object'}, } @@ -24207,7 +24574,7 @@ def __init__( max_concurrent_connections: Optional[Any] = None, disable_metrics_collection: Optional[Any] = None, query_timeout: Optional[Any] = None, - additional_columns: Optional[List["AdditionalColumns"]] = None, + additional_columns: Optional[Any] = None, query: Optional[Any] = None, **kwargs ): @@ -24967,8 +25334,8 @@ class MariaDBSource(TabularSource): pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type query_timeout: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any :param query: A query to retrieve data from source. Type: string (or Expression with resultType string). :type query: any @@ -24986,7 +25353,7 @@ class MariaDBSource(TabularSource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, 'query': {'key': 'query', 'type': 'object'}, } @@ -24999,7 +25366,7 @@ def __init__( max_concurrent_connections: Optional[Any] = None, disable_metrics_collection: Optional[Any] = None, query_timeout: Optional[Any] = None, - additional_columns: Optional[List["AdditionalColumns"]] = None, + additional_columns: Optional[Any] = None, query: Optional[Any] = None, **kwargs ): @@ -25260,8 +25627,8 @@ class MarketoSource(TabularSource): pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type query_timeout: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any :param query: A query to retrieve data from source. Type: string (or Expression with resultType string). :type query: any @@ -25279,7 +25646,7 @@ class MarketoSource(TabularSource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, 'query': {'key': 'query', 'type': 'object'}, } @@ -25292,7 +25659,7 @@ def __init__( max_concurrent_connections: Optional[Any] = None, disable_metrics_collection: Optional[Any] = None, query_timeout: Optional[Any] = None, - additional_columns: Optional[List["AdditionalColumns"]] = None, + additional_columns: Optional[Any] = None, query: Optional[Any] = None, **kwargs ): @@ -25505,8 +25872,8 @@ class MicrosoftAccessSource(CopySource): :param query: Database query. Type: string (or Expression with resultType string). :type query: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any """ _validation = { @@ -25521,7 +25888,7 @@ class MicrosoftAccessSource(CopySource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'query': {'key': 'query', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, } def __init__( @@ -25533,7 +25900,7 @@ def __init__( max_concurrent_connections: Optional[Any] = None, disable_metrics_collection: Optional[Any] = None, query: Optional[Any] = None, - additional_columns: Optional[List["AdditionalColumns"]] = None, + additional_columns: Optional[Any] = None, **kwargs ): super(MicrosoftAccessSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, max_concurrent_connections=max_concurrent_connections, disable_metrics_collection=disable_metrics_collection, **kwargs) @@ -25848,8 +26215,8 @@ class MongoDbAtlasSource(CopySource): pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type query_timeout: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any """ _validation = { @@ -25867,7 +26234,7 @@ class MongoDbAtlasSource(CopySource): 'cursor_methods': {'key': 'cursorMethods', 'type': 'MongoDbCursorMethodsProperties'}, 'batch_size': {'key': 'batchSize', 'type': 'object'}, 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, } def __init__( @@ -25882,7 +26249,7 @@ def __init__( cursor_methods: Optional["MongoDbCursorMethodsProperties"] = None, batch_size: Optional[Any] = None, query_timeout: Optional[Any] = None, - additional_columns: Optional[List["AdditionalColumns"]] = None, + additional_columns: Optional[Any] = None, **kwargs ): super(MongoDbAtlasSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, max_concurrent_connections=max_concurrent_connections, disable_metrics_collection=disable_metrics_collection, **kwargs) @@ -26148,8 +26515,8 @@ class MongoDbSource(CopySource): with resultType string). :type query: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any """ _validation = { @@ -26164,7 +26531,7 @@ class MongoDbSource(CopySource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'query': {'key': 'query', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, } def __init__( @@ -26176,7 +26543,7 @@ def __init__( max_concurrent_connections: Optional[Any] = None, disable_metrics_collection: Optional[Any] = None, query: Optional[Any] = None, - additional_columns: Optional[List["AdditionalColumns"]] = None, + additional_columns: Optional[Any] = None, **kwargs ): super(MongoDbSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, max_concurrent_connections=max_concurrent_connections, disable_metrics_collection=disable_metrics_collection, **kwargs) @@ -26421,8 +26788,8 @@ class MongoDbV2Source(CopySource): pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type query_timeout: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any """ _validation = { @@ -26440,7 +26807,7 @@ class MongoDbV2Source(CopySource): 'cursor_methods': {'key': 'cursorMethods', 'type': 'MongoDbCursorMethodsProperties'}, 'batch_size': {'key': 'batchSize', 'type': 'object'}, 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, } def __init__( @@ -26455,7 +26822,7 @@ def __init__( cursor_methods: Optional["MongoDbCursorMethodsProperties"] = None, batch_size: Optional[Any] = None, query_timeout: Optional[Any] = None, - additional_columns: Optional[List["AdditionalColumns"]] = None, + additional_columns: Optional[Any] = None, **kwargs ): super(MongoDbV2Source, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, max_concurrent_connections=max_concurrent_connections, disable_metrics_collection=disable_metrics_collection, **kwargs) @@ -26558,8 +26925,8 @@ class MySqlSource(TabularSource): pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type query_timeout: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any :param query: Database query. Type: string (or Expression with resultType string). :type query: any """ @@ -26576,7 +26943,7 @@ class MySqlSource(TabularSource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, 'query': {'key': 'query', 'type': 'object'}, } @@ -26589,7 +26956,7 @@ def __init__( max_concurrent_connections: Optional[Any] = None, disable_metrics_collection: Optional[Any] = None, query_timeout: Optional[Any] = None, - additional_columns: Optional[List["AdditionalColumns"]] = None, + additional_columns: Optional[Any] = None, query: Optional[Any] = None, **kwargs ): @@ -26793,8 +27160,8 @@ class NetezzaSource(TabularSource): pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type query_timeout: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any :param query: A query to retrieve data from source. Type: string (or Expression with resultType string). :type query: any @@ -26817,7 +27184,7 @@ class NetezzaSource(TabularSource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, 'query': {'key': 'query', 'type': 'object'}, 'partition_option': {'key': 'partitionOption', 'type': 'object'}, 'partition_settings': {'key': 'partitionSettings', 'type': 'NetezzaPartitionSettings'}, @@ -26832,7 +27199,7 @@ def __init__( max_concurrent_connections: Optional[Any] = None, disable_metrics_collection: Optional[Any] = None, query_timeout: Optional[Any] = None, - additional_columns: Optional[List["AdditionalColumns"]] = None, + additional_columns: Optional[Any] = None, query: Optional[Any] = None, partition_option: Optional[Any] = None, partition_settings: Optional["NetezzaPartitionSettings"] = None, @@ -27162,8 +27529,8 @@ class ODataSource(CopySource): ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type http_request_timeout: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any """ _validation = { @@ -27179,7 +27546,7 @@ class ODataSource(CopySource): 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'query': {'key': 'query', 'type': 'object'}, 'http_request_timeout': {'key': 'httpRequestTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, } def __init__( @@ -27192,7 +27559,7 @@ def __init__( disable_metrics_collection: Optional[Any] = None, query: Optional[Any] = None, http_request_timeout: Optional[Any] = None, - additional_columns: Optional[List["AdditionalColumns"]] = None, + additional_columns: Optional[Any] = None, **kwargs ): super(ODataSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, max_concurrent_connections=max_concurrent_connections, disable_metrics_collection=disable_metrics_collection, **kwargs) @@ -27380,8 +27747,8 @@ class OdbcSource(TabularSource): pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type query_timeout: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any :param query: Database query. Type: string (or Expression with resultType string). :type query: any """ @@ -27398,7 +27765,7 @@ class OdbcSource(TabularSource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, 'query': {'key': 'query', 'type': 'object'}, } @@ -27411,7 +27778,7 @@ def __init__( max_concurrent_connections: Optional[Any] = None, disable_metrics_collection: Optional[Any] = None, query_timeout: Optional[Any] = None, - additional_columns: Optional[List["AdditionalColumns"]] = None, + additional_columns: Optional[Any] = None, query: Optional[Any] = None, **kwargs ): @@ -28533,8 +28900,8 @@ class OracleServiceCloudSource(TabularSource): pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type query_timeout: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any :param query: A query to retrieve data from source. Type: string (or Expression with resultType string). :type query: any @@ -28552,7 +28919,7 @@ class OracleServiceCloudSource(TabularSource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, 'query': {'key': 'query', 'type': 'object'}, } @@ -28565,7 +28932,7 @@ def __init__( max_concurrent_connections: Optional[Any] = None, disable_metrics_collection: Optional[Any] = None, query_timeout: Optional[Any] = None, - additional_columns: Optional[List["AdditionalColumns"]] = None, + additional_columns: Optional[Any] = None, query: Optional[Any] = None, **kwargs ): @@ -28675,8 +29042,8 @@ class OracleSource(CopySource): :param partition_settings: The settings that will be leveraged for Oracle source partitioning. :type partition_settings: ~azure.mgmt.datafactory.models.OraclePartitionSettings :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any """ _validation = { @@ -28694,7 +29061,7 @@ class OracleSource(CopySource): 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, 'partition_option': {'key': 'partitionOption', 'type': 'object'}, 'partition_settings': {'key': 'partitionSettings', 'type': 'OraclePartitionSettings'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, } def __init__( @@ -28709,7 +29076,7 @@ def __init__( query_timeout: Optional[Any] = None, partition_option: Optional[Any] = None, partition_settings: Optional["OraclePartitionSettings"] = None, - additional_columns: Optional[List["AdditionalColumns"]] = None, + additional_columns: Optional[Any] = None, **kwargs ): super(OracleSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, max_concurrent_connections=max_concurrent_connections, disable_metrics_collection=disable_metrics_collection, **kwargs) @@ -29011,8 +29378,8 @@ class OrcSource(CopySource): :param store_settings: ORC store settings. :type store_settings: ~azure.mgmt.datafactory.models.StoreReadSettings :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any """ _validation = { @@ -29027,7 +29394,7 @@ class OrcSource(CopySource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'store_settings': {'key': 'storeSettings', 'type': 'StoreReadSettings'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, } def __init__( @@ -29039,7 +29406,7 @@ def __init__( max_concurrent_connections: Optional[Any] = None, disable_metrics_collection: Optional[Any] = None, store_settings: Optional["StoreReadSettings"] = None, - additional_columns: Optional[List["AdditionalColumns"]] = None, + additional_columns: Optional[Any] = None, **kwargs ): super(OrcSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, max_concurrent_connections=max_concurrent_connections, disable_metrics_collection=disable_metrics_collection, **kwargs) @@ -29367,8 +29734,8 @@ class ParquetSource(CopySource): :param store_settings: Parquet store settings. :type store_settings: ~azure.mgmt.datafactory.models.StoreReadSettings :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any """ _validation = { @@ -29383,7 +29750,7 @@ class ParquetSource(CopySource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'store_settings': {'key': 'storeSettings', 'type': 'StoreReadSettings'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, } def __init__( @@ -29395,7 +29762,7 @@ def __init__( max_concurrent_connections: Optional[Any] = None, disable_metrics_collection: Optional[Any] = None, store_settings: Optional["StoreReadSettings"] = None, - additional_columns: Optional[List["AdditionalColumns"]] = None, + additional_columns: Optional[Any] = None, **kwargs ): super(ParquetSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, max_concurrent_connections=max_concurrent_connections, disable_metrics_collection=disable_metrics_collection, **kwargs) @@ -29632,8 +29999,8 @@ class PaypalSource(TabularSource): pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type query_timeout: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any :param query: A query to retrieve data from source. Type: string (or Expression with resultType string). :type query: any @@ -29651,7 +30018,7 @@ class PaypalSource(TabularSource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, 'query': {'key': 'query', 'type': 'object'}, } @@ -29664,7 +30031,7 @@ def __init__( max_concurrent_connections: Optional[Any] = None, disable_metrics_collection: Optional[Any] = None, query_timeout: Optional[Any] = None, - additional_columns: Optional[List["AdditionalColumns"]] = None, + additional_columns: Optional[Any] = None, query: Optional[Any] = None, **kwargs ): @@ -29903,8 +30270,8 @@ class PhoenixSource(TabularSource): pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type query_timeout: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any :param query: A query to retrieve data from source. Type: string (or Expression with resultType string). :type query: any @@ -29922,7 +30289,7 @@ class PhoenixSource(TabularSource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, 'query': {'key': 'query', 'type': 'object'}, } @@ -29935,7 +30302,7 @@ def __init__( max_concurrent_connections: Optional[Any] = None, disable_metrics_collection: Optional[Any] = None, query_timeout: Optional[Any] = None, - additional_columns: Optional[List["AdditionalColumns"]] = None, + additional_columns: Optional[Any] = None, query: Optional[Any] = None, **kwargs ): @@ -30478,8 +30845,8 @@ class PostgreSqlSource(TabularSource): pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type query_timeout: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any :param query: Database query. Type: string (or Expression with resultType string). :type query: any """ @@ -30496,7 +30863,7 @@ class PostgreSqlSource(TabularSource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, 'query': {'key': 'query', 'type': 'object'}, } @@ -30509,7 +30876,7 @@ def __init__( max_concurrent_connections: Optional[Any] = None, disable_metrics_collection: Optional[Any] = None, query_timeout: Optional[Any] = None, - additional_columns: Optional[List["AdditionalColumns"]] = None, + additional_columns: Optional[Any] = None, query: Optional[Any] = None, **kwargs ): @@ -30598,6 +30965,100 @@ def __init__( self.schema_type_properties_schema = schema_type_properties_schema +class PowerQuerySink(DataFlowSink): + """Power query sink. + + All required parameters must be populated in order to send to Azure. + + :param name: Required. Transformation name. + :type name: str + :param description: Transformation description. + :type description: str + :param dataset: Dataset reference. + :type dataset: ~azure.mgmt.datafactory.models.DatasetReference + :param linked_service: Linked service reference. + :type linked_service: ~azure.mgmt.datafactory.models.LinkedServiceReference + :param schema_linked_service: Schema linked service reference. + :type schema_linked_service: ~azure.mgmt.datafactory.models.LinkedServiceReference + :param script: sink script. + :type script: str + """ + + _validation = { + 'name': {'required': True}, + } + + _attribute_map = { + 'name': {'key': 'name', 'type': 'str'}, + 'description': {'key': 'description', 'type': 'str'}, + 'dataset': {'key': 'dataset', 'type': 'DatasetReference'}, + 'linked_service': {'key': 'linkedService', 'type': 'LinkedServiceReference'}, + 'schema_linked_service': {'key': 'schemaLinkedService', 'type': 'LinkedServiceReference'}, + 'script': {'key': 'script', 'type': 'str'}, + } + + def __init__( + self, + *, + name: str, + description: Optional[str] = None, + dataset: Optional["DatasetReference"] = None, + linked_service: Optional["LinkedServiceReference"] = None, + schema_linked_service: Optional["LinkedServiceReference"] = None, + script: Optional[str] = None, + **kwargs + ): + super(PowerQuerySink, self).__init__(name=name, description=description, dataset=dataset, linked_service=linked_service, schema_linked_service=schema_linked_service, **kwargs) + self.script = script + + +class PowerQuerySource(DataFlowSource): + """Power query source. + + All required parameters must be populated in order to send to Azure. + + :param name: Required. Transformation name. + :type name: str + :param description: Transformation description. + :type description: str + :param dataset: Dataset reference. + :type dataset: ~azure.mgmt.datafactory.models.DatasetReference + :param linked_service: Linked service reference. + :type linked_service: ~azure.mgmt.datafactory.models.LinkedServiceReference + :param schema_linked_service: Schema linked service reference. + :type schema_linked_service: ~azure.mgmt.datafactory.models.LinkedServiceReference + :param script: source script. + :type script: str + """ + + _validation = { + 'name': {'required': True}, + } + + _attribute_map = { + 'name': {'key': 'name', 'type': 'str'}, + 'description': {'key': 'description', 'type': 'str'}, + 'dataset': {'key': 'dataset', 'type': 'DatasetReference'}, + 'linked_service': {'key': 'linkedService', 'type': 'LinkedServiceReference'}, + 'schema_linked_service': {'key': 'schemaLinkedService', 'type': 'LinkedServiceReference'}, + 'script': {'key': 'script', 'type': 'str'}, + } + + def __init__( + self, + *, + name: str, + description: Optional[str] = None, + dataset: Optional["DatasetReference"] = None, + linked_service: Optional["LinkedServiceReference"] = None, + schema_linked_service: Optional["LinkedServiceReference"] = None, + script: Optional[str] = None, + **kwargs + ): + super(PowerQuerySource, self).__init__(name=name, description=description, dataset=dataset, linked_service=linked_service, schema_linked_service=schema_linked_service, **kwargs) + self.script = script + + class PrestoLinkedService(LinkedService): """Presto server linked service. @@ -30838,8 +31299,8 @@ class PrestoSource(TabularSource): pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type query_timeout: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any :param query: A query to retrieve data from source. Type: string (or Expression with resultType string). :type query: any @@ -30857,7 +31318,7 @@ class PrestoSource(TabularSource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, 'query': {'key': 'query', 'type': 'object'}, } @@ -30870,7 +31331,7 @@ def __init__( max_concurrent_connections: Optional[Any] = None, disable_metrics_collection: Optional[Any] = None, query_timeout: Optional[Any] = None, - additional_columns: Optional[List["AdditionalColumns"]] = None, + additional_columns: Optional[Any] = None, query: Optional[Any] = None, **kwargs ): @@ -31368,8 +31829,8 @@ class QuickBooksSource(TabularSource): pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type query_timeout: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any :param query: A query to retrieve data from source. Type: string (or Expression with resultType string). :type query: any @@ -31387,7 +31848,7 @@ class QuickBooksSource(TabularSource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, 'query': {'key': 'query', 'type': 'object'}, } @@ -31400,7 +31861,7 @@ def __init__( max_concurrent_connections: Optional[Any] = None, disable_metrics_collection: Optional[Any] = None, query_timeout: Optional[Any] = None, - additional_columns: Optional[List["AdditionalColumns"]] = None, + additional_columns: Optional[Any] = None, query: Optional[Any] = None, **kwargs ): @@ -31592,8 +32053,8 @@ class RelationalSource(CopySource): :param query: Database query. Type: string (or Expression with resultType string). :type query: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any """ _validation = { @@ -31608,7 +32069,7 @@ class RelationalSource(CopySource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'query': {'key': 'query', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, } def __init__( @@ -31620,7 +32081,7 @@ def __init__( max_concurrent_connections: Optional[Any] = None, disable_metrics_collection: Optional[Any] = None, query: Optional[Any] = None, - additional_columns: Optional[List["AdditionalColumns"]] = None, + additional_columns: Optional[Any] = None, **kwargs ): super(RelationalSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, max_concurrent_connections=max_concurrent_connections, disable_metrics_collection=disable_metrics_collection, **kwargs) @@ -31995,8 +32456,8 @@ class ResponsysSource(TabularSource): pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type query_timeout: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any :param query: A query to retrieve data from source. Type: string (or Expression with resultType string). :type query: any @@ -32014,7 +32475,7 @@ class ResponsysSource(TabularSource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, 'query': {'key': 'query', 'type': 'object'}, } @@ -32027,7 +32488,7 @@ def __init__( max_concurrent_connections: Optional[Any] = None, disable_metrics_collection: Optional[Any] = None, query_timeout: Optional[Any] = None, - additional_columns: Optional[List["AdditionalColumns"]] = None, + additional_columns: Optional[Any] = None, query: Optional[Any] = None, **kwargs ): @@ -32389,8 +32850,8 @@ class RestSource(CopySource): :param request_interval: The time to await before sending next page request. :type request_interval: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any """ _validation = { @@ -32410,7 +32871,7 @@ class RestSource(CopySource): 'pagination_rules': {'key': 'paginationRules', 'type': 'object'}, 'http_request_timeout': {'key': 'httpRequestTimeout', 'type': 'object'}, 'request_interval': {'key': 'requestInterval', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, } def __init__( @@ -32427,7 +32888,7 @@ def __init__( pagination_rules: Optional[Any] = None, http_request_timeout: Optional[Any] = None, request_interval: Optional[Any] = None, - additional_columns: Optional[List["AdditionalColumns"]] = None, + additional_columns: Optional[Any] = None, **kwargs ): super(RestSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, max_concurrent_connections=max_concurrent_connections, disable_metrics_collection=disable_metrics_collection, **kwargs) @@ -32876,8 +33337,8 @@ class SalesforceMarketingCloudSource(TabularSource): pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type query_timeout: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any :param query: A query to retrieve data from source. Type: string (or Expression with resultType string). :type query: any @@ -32895,7 +33356,7 @@ class SalesforceMarketingCloudSource(TabularSource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, 'query': {'key': 'query', 'type': 'object'}, } @@ -32908,7 +33369,7 @@ def __init__( max_concurrent_connections: Optional[Any] = None, disable_metrics_collection: Optional[Any] = None, query_timeout: Optional[Any] = None, - additional_columns: Optional[List["AdditionalColumns"]] = None, + additional_columns: Optional[Any] = None, query: Optional[Any] = None, **kwargs ): @@ -33256,8 +33717,8 @@ class SalesforceServiceCloudSource(CopySource): include: "Query", "QueryAll". :type read_behavior: str or ~azure.mgmt.datafactory.models.SalesforceSourceReadBehavior :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any """ _validation = { @@ -33273,7 +33734,7 @@ class SalesforceServiceCloudSource(CopySource): 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'query': {'key': 'query', 'type': 'object'}, 'read_behavior': {'key': 'readBehavior', 'type': 'str'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, } def __init__( @@ -33286,7 +33747,7 @@ def __init__( disable_metrics_collection: Optional[Any] = None, query: Optional[Any] = None, read_behavior: Optional[Union[str, "SalesforceSourceReadBehavior"]] = None, - additional_columns: Optional[List["AdditionalColumns"]] = None, + additional_columns: Optional[Any] = None, **kwargs ): super(SalesforceServiceCloudSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, max_concurrent_connections=max_concurrent_connections, disable_metrics_collection=disable_metrics_collection, **kwargs) @@ -33405,8 +33866,8 @@ class SalesforceSource(TabularSource): pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type query_timeout: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any :param query: Database query. Type: string (or Expression with resultType string). :type query: any :param read_behavior: The read behavior for the operation. Default is Query. Possible values @@ -33426,7 +33887,7 @@ class SalesforceSource(TabularSource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, 'query': {'key': 'query', 'type': 'object'}, 'read_behavior': {'key': 'readBehavior', 'type': 'str'}, } @@ -33440,7 +33901,7 @@ def __init__( max_concurrent_connections: Optional[Any] = None, disable_metrics_collection: Optional[Any] = None, query_timeout: Optional[Any] = None, - additional_columns: Optional[List["AdditionalColumns"]] = None, + additional_columns: Optional[Any] = None, query: Optional[Any] = None, read_behavior: Optional[Union[str, "SalesforceSourceReadBehavior"]] = None, **kwargs @@ -33626,8 +34087,8 @@ class SapBwSource(TabularSource): pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type query_timeout: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any :param query: MDX query. Type: string (or Expression with resultType string). :type query: any """ @@ -33644,7 +34105,7 @@ class SapBwSource(TabularSource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, 'query': {'key': 'query', 'type': 'object'}, } @@ -33657,7 +34118,7 @@ def __init__( max_concurrent_connections: Optional[Any] = None, disable_metrics_collection: Optional[Any] = None, query_timeout: Optional[Any] = None, - additional_columns: Optional[List["AdditionalColumns"]] = None, + additional_columns: Optional[Any] = None, query: Optional[Any] = None, **kwargs ): @@ -33911,8 +34372,8 @@ class SapCloudForCustomerSource(TabularSource): pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type query_timeout: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any :param query: SAP Cloud for Customer OData query. For example, "$top=1". Type: string (or Expression with resultType string). :type query: any @@ -33935,7 +34396,7 @@ class SapCloudForCustomerSource(TabularSource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, 'query': {'key': 'query', 'type': 'object'}, 'http_request_timeout': {'key': 'httpRequestTimeout', 'type': 'object'}, } @@ -33949,7 +34410,7 @@ def __init__( max_concurrent_connections: Optional[Any] = None, disable_metrics_collection: Optional[Any] = None, query_timeout: Optional[Any] = None, - additional_columns: Optional[List["AdditionalColumns"]] = None, + additional_columns: Optional[Any] = None, query: Optional[Any] = None, http_request_timeout: Optional[Any] = None, **kwargs @@ -34129,8 +34590,8 @@ class SapEccSource(TabularSource): pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type query_timeout: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any :param query: SAP ECC OData query. For example, "$top=1". Type: string (or Expression with resultType string). :type query: any @@ -34153,7 +34614,7 @@ class SapEccSource(TabularSource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, 'query': {'key': 'query', 'type': 'object'}, 'http_request_timeout': {'key': 'httpRequestTimeout', 'type': 'object'}, } @@ -34167,7 +34628,7 @@ def __init__( max_concurrent_connections: Optional[Any] = None, disable_metrics_collection: Optional[Any] = None, query_timeout: Optional[Any] = None, - additional_columns: Optional[List["AdditionalColumns"]] = None, + additional_columns: Optional[Any] = None, query: Optional[Any] = None, http_request_timeout: Optional[Any] = None, **kwargs @@ -34309,8 +34770,8 @@ class SapHanaSource(TabularSource): pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type query_timeout: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any :param query: SAP HANA Sql query. Type: string (or Expression with resultType string). :type query: any :param packet_size: The packet size of data read from SAP HANA. Type: integer(or Expression @@ -34336,7 +34797,7 @@ class SapHanaSource(TabularSource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, 'query': {'key': 'query', 'type': 'object'}, 'packet_size': {'key': 'packetSize', 'type': 'object'}, 'partition_option': {'key': 'partitionOption', 'type': 'object'}, @@ -34352,7 +34813,7 @@ def __init__( max_concurrent_connections: Optional[Any] = None, disable_metrics_collection: Optional[Any] = None, query_timeout: Optional[Any] = None, - additional_columns: Optional[List["AdditionalColumns"]] = None, + additional_columns: Optional[Any] = None, query: Optional[Any] = None, packet_size: Optional[Any] = None, partition_option: Optional[Any] = None, @@ -34583,8 +35044,8 @@ class SapOpenHubSource(TabularSource): pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type query_timeout: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any :param exclude_last_request: Whether to exclude the records of the last request. The default value is true. Type: boolean (or Expression with resultType boolean). :type exclude_last_request: any @@ -34613,7 +35074,7 @@ class SapOpenHubSource(TabularSource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, 'exclude_last_request': {'key': 'excludeLastRequest', 'type': 'object'}, 'base_request_id': {'key': 'baseRequestId', 'type': 'object'}, 'custom_rfc_read_table_function_module': {'key': 'customRfcReadTableFunctionModule', 'type': 'object'}, @@ -34629,7 +35090,7 @@ def __init__( max_concurrent_connections: Optional[Any] = None, disable_metrics_collection: Optional[Any] = None, query_timeout: Optional[Any] = None, - additional_columns: Optional[List["AdditionalColumns"]] = None, + additional_columns: Optional[Any] = None, exclude_last_request: Optional[Any] = None, base_request_id: Optional[Any] = None, custom_rfc_read_table_function_module: Optional[Any] = None, @@ -35010,8 +35471,8 @@ class SapTableSource(TabularSource): pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type query_timeout: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any :param row_count: The number of rows to be retrieved. Type: integer(or Expression with resultType integer). :type row_count: any @@ -35055,7 +35516,7 @@ class SapTableSource(TabularSource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, 'row_count': {'key': 'rowCount', 'type': 'object'}, 'row_skips': {'key': 'rowSkips', 'type': 'object'}, 'rfc_table_fields': {'key': 'rfcTableFields', 'type': 'object'}, @@ -35076,7 +35537,7 @@ def __init__( max_concurrent_connections: Optional[Any] = None, disable_metrics_collection: Optional[Any] = None, query_timeout: Optional[Any] = None, - additional_columns: Optional[List["AdditionalColumns"]] = None, + additional_columns: Optional[Any] = None, row_count: Optional[Any] = None, row_skips: Optional[Any] = None, rfc_table_fields: Optional[Any] = None, @@ -35826,8 +36287,8 @@ class ServiceNowSource(TabularSource): pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type query_timeout: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any :param query: A query to retrieve data from source. Type: string (or Expression with resultType string). :type query: any @@ -35845,7 +36306,7 @@ class ServiceNowSource(TabularSource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, 'query': {'key': 'query', 'type': 'object'}, } @@ -35858,7 +36319,7 @@ def __init__( max_concurrent_connections: Optional[Any] = None, disable_metrics_collection: Optional[Any] = None, query_timeout: Optional[Any] = None, - additional_columns: Optional[List["AdditionalColumns"]] = None, + additional_columns: Optional[Any] = None, query: Optional[Any] = None, **kwargs ): @@ -36688,8 +37149,8 @@ class ShopifySource(TabularSource): pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type query_timeout: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any :param query: A query to retrieve data from source. Type: string (or Expression with resultType string). :type query: any @@ -36707,7 +37168,7 @@ class ShopifySource(TabularSource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, 'query': {'key': 'query', 'type': 'object'}, } @@ -36720,7 +37181,7 @@ def __init__( max_concurrent_connections: Optional[Any] = None, disable_metrics_collection: Optional[Any] = None, query_timeout: Optional[Any] = None, - additional_columns: Optional[List["AdditionalColumns"]] = None, + additional_columns: Optional[Any] = None, query: Optional[Any] = None, **kwargs ): @@ -37366,8 +37827,8 @@ class SparkSource(TabularSource): pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type query_timeout: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any :param query: A query to retrieve data from source. Type: string (or Expression with resultType string). :type query: any @@ -37385,7 +37846,7 @@ class SparkSource(TabularSource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, 'query': {'key': 'query', 'type': 'object'}, } @@ -37398,7 +37859,7 @@ def __init__( max_concurrent_connections: Optional[Any] = None, disable_metrics_collection: Optional[Any] = None, query_timeout: Optional[Any] = None, - additional_columns: Optional[List["AdditionalColumns"]] = None, + additional_columns: Optional[Any] = None, query: Optional[Any] = None, **kwargs ): @@ -37494,6 +37955,14 @@ class SqlDWSink(CopySink): :param table_option: The option to handle sink table, such as autoCreate. For now only 'autoCreate' value is supported. Type: string (or Expression with resultType string). :type table_option: any + :param sql_writer_use_table_lock: Whether to use table lock during bulk copy. Type: boolean (or + Expression with resultType boolean). + :type sql_writer_use_table_lock: any + :param write_behavior: Write behavior when copying data into azure SQL DW. Type: + SqlDWWriteBehaviorEnum (or Expression with resultType SqlDWWriteBehaviorEnum). + :type write_behavior: any + :param upsert_settings: SQL DW upsert settings. + :type upsert_settings: ~azure.mgmt.datafactory.models.SqlDWUpsertSettings """ _validation = { @@ -37515,6 +37984,9 @@ class SqlDWSink(CopySink): 'allow_copy_command': {'key': 'allowCopyCommand', 'type': 'object'}, 'copy_command_settings': {'key': 'copyCommandSettings', 'type': 'DWCopyCommandSettings'}, 'table_option': {'key': 'tableOption', 'type': 'object'}, + 'sql_writer_use_table_lock': {'key': 'sqlWriterUseTableLock', 'type': 'object'}, + 'write_behavior': {'key': 'writeBehavior', 'type': 'object'}, + 'upsert_settings': {'key': 'upsertSettings', 'type': 'SqlDWUpsertSettings'}, } def __init__( @@ -37533,6 +38005,9 @@ def __init__( allow_copy_command: Optional[Any] = None, copy_command_settings: Optional["DWCopyCommandSettings"] = None, table_option: Optional[Any] = None, + sql_writer_use_table_lock: Optional[Any] = None, + write_behavior: Optional[Any] = None, + upsert_settings: Optional["SqlDWUpsertSettings"] = None, **kwargs ): super(SqlDWSink, self).__init__(additional_properties=additional_properties, write_batch_size=write_batch_size, write_batch_timeout=write_batch_timeout, sink_retry_count=sink_retry_count, sink_retry_wait=sink_retry_wait, max_concurrent_connections=max_concurrent_connections, disable_metrics_collection=disable_metrics_collection, **kwargs) @@ -37543,6 +38018,9 @@ def __init__( self.allow_copy_command = allow_copy_command self.copy_command_settings = copy_command_settings self.table_option = table_option + self.sql_writer_use_table_lock = sql_writer_use_table_lock + self.write_behavior = write_behavior + self.upsert_settings = upsert_settings class SqlDWSource(TabularSource): @@ -37571,8 +38049,8 @@ class SqlDWSource(TabularSource): pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type query_timeout: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any :param sql_reader_query: SQL Data Warehouse reader query. Type: string (or Expression with resultType string). :type sql_reader_query: any @@ -37603,7 +38081,7 @@ class SqlDWSource(TabularSource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, 'sql_reader_query': {'key': 'sqlReaderQuery', 'type': 'object'}, 'sql_reader_stored_procedure_name': {'key': 'sqlReaderStoredProcedureName', 'type': 'object'}, 'stored_procedure_parameters': {'key': 'storedProcedureParameters', 'type': 'object'}, @@ -37620,7 +38098,7 @@ def __init__( max_concurrent_connections: Optional[Any] = None, disable_metrics_collection: Optional[Any] = None, query_timeout: Optional[Any] = None, - additional_columns: Optional[List["AdditionalColumns"]] = None, + additional_columns: Optional[Any] = None, sql_reader_query: Optional[Any] = None, sql_reader_stored_procedure_name: Optional[Any] = None, stored_procedure_parameters: Optional[Any] = None, @@ -37637,6 +38115,34 @@ def __init__( self.partition_settings = partition_settings +class SqlDWUpsertSettings(msrest.serialization.Model): + """Sql DW upsert option settings. + + :param interim_schema_name: Schema name for interim table. Type: string (or Expression with + resultType string). + :type interim_schema_name: any + :param keys: Key column names for unique row identification. Type: array of strings (or + Expression with resultType array of strings). + :type keys: any + """ + + _attribute_map = { + 'interim_schema_name': {'key': 'interimSchemaName', 'type': 'object'}, + 'keys': {'key': 'keys', 'type': 'object'}, + } + + def __init__( + self, + *, + interim_schema_name: Optional[Any] = None, + keys: Optional[Any] = None, + **kwargs + ): + super(SqlDWUpsertSettings, self).__init__(**kwargs) + self.interim_schema_name = interim_schema_name + self.keys = keys + + class SqlMISink(CopySink): """A copy activity Azure SQL Managed Instance sink. @@ -37683,6 +38189,14 @@ class SqlMISink(CopySink): :param table_option: The option to handle sink table, such as autoCreate. For now only 'autoCreate' value is supported. Type: string (or Expression with resultType string). :type table_option: any + :param sql_writer_use_table_lock: Whether to use table lock during bulk copy. Type: boolean (or + Expression with resultType boolean). + :type sql_writer_use_table_lock: any + :param write_behavior: White behavior when copying data into azure SQL MI. Type: + SqlWriteBehaviorEnum (or Expression with resultType SqlWriteBehaviorEnum). + :type write_behavior: any + :param upsert_settings: SQL upsert settings. + :type upsert_settings: ~azure.mgmt.datafactory.models.SqlUpsertSettings """ _validation = { @@ -37704,6 +38218,9 @@ class SqlMISink(CopySink): 'stored_procedure_parameters': {'key': 'storedProcedureParameters', 'type': '{StoredProcedureParameter}'}, 'stored_procedure_table_type_parameter_name': {'key': 'storedProcedureTableTypeParameterName', 'type': 'object'}, 'table_option': {'key': 'tableOption', 'type': 'object'}, + 'sql_writer_use_table_lock': {'key': 'sqlWriterUseTableLock', 'type': 'object'}, + 'write_behavior': {'key': 'writeBehavior', 'type': 'object'}, + 'upsert_settings': {'key': 'upsertSettings', 'type': 'SqlUpsertSettings'}, } def __init__( @@ -37722,6 +38239,9 @@ def __init__( stored_procedure_parameters: Optional[Dict[str, "StoredProcedureParameter"]] = None, stored_procedure_table_type_parameter_name: Optional[Any] = None, table_option: Optional[Any] = None, + sql_writer_use_table_lock: Optional[Any] = None, + write_behavior: Optional[Any] = None, + upsert_settings: Optional["SqlUpsertSettings"] = None, **kwargs ): super(SqlMISink, self).__init__(additional_properties=additional_properties, write_batch_size=write_batch_size, write_batch_timeout=write_batch_timeout, sink_retry_count=sink_retry_count, sink_retry_wait=sink_retry_wait, max_concurrent_connections=max_concurrent_connections, disable_metrics_collection=disable_metrics_collection, **kwargs) @@ -37732,6 +38252,9 @@ def __init__( self.stored_procedure_parameters = stored_procedure_parameters self.stored_procedure_table_type_parameter_name = stored_procedure_table_type_parameter_name self.table_option = table_option + self.sql_writer_use_table_lock = sql_writer_use_table_lock + self.write_behavior = write_behavior + self.upsert_settings = upsert_settings class SqlMISource(TabularSource): @@ -37760,8 +38283,8 @@ class SqlMISource(TabularSource): pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type query_timeout: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any :param sql_reader_query: SQL reader query. Type: string (or Expression with resultType string). :type sql_reader_query: any :param sql_reader_stored_procedure_name: Name of the stored procedure for a Azure SQL Managed @@ -37793,7 +38316,7 @@ class SqlMISource(TabularSource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, 'sql_reader_query': {'key': 'sqlReaderQuery', 'type': 'object'}, 'sql_reader_stored_procedure_name': {'key': 'sqlReaderStoredProcedureName', 'type': 'object'}, 'stored_procedure_parameters': {'key': 'storedProcedureParameters', 'type': '{StoredProcedureParameter}'}, @@ -37811,7 +38334,7 @@ def __init__( max_concurrent_connections: Optional[Any] = None, disable_metrics_collection: Optional[Any] = None, query_timeout: Optional[Any] = None, - additional_columns: Optional[List["AdditionalColumns"]] = None, + additional_columns: Optional[Any] = None, sql_reader_query: Optional[Any] = None, sql_reader_stored_procedure_name: Optional[Any] = None, stored_procedure_parameters: Optional[Dict[str, "StoredProcedureParameter"]] = None, @@ -37993,6 +38516,14 @@ class SqlServerSink(CopySink): :param table_option: The option to handle sink table, such as autoCreate. For now only 'autoCreate' value is supported. Type: string (or Expression with resultType string). :type table_option: any + :param sql_writer_use_table_lock: Whether to use table lock during bulk copy. Type: boolean (or + Expression with resultType boolean). + :type sql_writer_use_table_lock: any + :param write_behavior: Write behavior when copying data into sql server. Type: + SqlWriteBehaviorEnum (or Expression with resultType SqlWriteBehaviorEnum). + :type write_behavior: any + :param upsert_settings: SQL upsert settings. + :type upsert_settings: ~azure.mgmt.datafactory.models.SqlUpsertSettings """ _validation = { @@ -38014,6 +38545,9 @@ class SqlServerSink(CopySink): 'stored_procedure_parameters': {'key': 'storedProcedureParameters', 'type': '{StoredProcedureParameter}'}, 'stored_procedure_table_type_parameter_name': {'key': 'storedProcedureTableTypeParameterName', 'type': 'object'}, 'table_option': {'key': 'tableOption', 'type': 'object'}, + 'sql_writer_use_table_lock': {'key': 'sqlWriterUseTableLock', 'type': 'object'}, + 'write_behavior': {'key': 'writeBehavior', 'type': 'object'}, + 'upsert_settings': {'key': 'upsertSettings', 'type': 'SqlUpsertSettings'}, } def __init__( @@ -38032,6 +38566,9 @@ def __init__( stored_procedure_parameters: Optional[Dict[str, "StoredProcedureParameter"]] = None, stored_procedure_table_type_parameter_name: Optional[Any] = None, table_option: Optional[Any] = None, + sql_writer_use_table_lock: Optional[Any] = None, + write_behavior: Optional[Any] = None, + upsert_settings: Optional["SqlUpsertSettings"] = None, **kwargs ): super(SqlServerSink, self).__init__(additional_properties=additional_properties, write_batch_size=write_batch_size, write_batch_timeout=write_batch_timeout, sink_retry_count=sink_retry_count, sink_retry_wait=sink_retry_wait, max_concurrent_connections=max_concurrent_connections, disable_metrics_collection=disable_metrics_collection, **kwargs) @@ -38042,6 +38579,9 @@ def __init__( self.stored_procedure_parameters = stored_procedure_parameters self.stored_procedure_table_type_parameter_name = stored_procedure_table_type_parameter_name self.table_option = table_option + self.sql_writer_use_table_lock = sql_writer_use_table_lock + self.write_behavior = write_behavior + self.upsert_settings = upsert_settings class SqlServerSource(TabularSource): @@ -38070,8 +38610,8 @@ class SqlServerSource(TabularSource): pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type query_timeout: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any :param sql_reader_query: SQL reader query. Type: string (or Expression with resultType string). :type sql_reader_query: any :param sql_reader_stored_procedure_name: Name of the stored procedure for a SQL Database @@ -38103,7 +38643,7 @@ class SqlServerSource(TabularSource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, 'sql_reader_query': {'key': 'sqlReaderQuery', 'type': 'object'}, 'sql_reader_stored_procedure_name': {'key': 'sqlReaderStoredProcedureName', 'type': 'object'}, 'stored_procedure_parameters': {'key': 'storedProcedureParameters', 'type': '{StoredProcedureParameter}'}, @@ -38121,7 +38661,7 @@ def __init__( max_concurrent_connections: Optional[Any] = None, disable_metrics_collection: Optional[Any] = None, query_timeout: Optional[Any] = None, - additional_columns: Optional[List["AdditionalColumns"]] = None, + additional_columns: Optional[Any] = None, sql_reader_query: Optional[Any] = None, sql_reader_stored_procedure_name: Optional[Any] = None, stored_procedure_parameters: Optional[Dict[str, "StoredProcedureParameter"]] = None, @@ -38337,6 +38877,14 @@ class SqlSink(CopySink): :param table_option: The option to handle sink table, such as autoCreate. For now only 'autoCreate' value is supported. Type: string (or Expression with resultType string). :type table_option: any + :param sql_writer_use_table_lock: Whether to use table lock during bulk copy. Type: boolean (or + Expression with resultType boolean). + :type sql_writer_use_table_lock: any + :param write_behavior: Write behavior when copying data into sql. Type: SqlWriteBehaviorEnum + (or Expression with resultType SqlWriteBehaviorEnum). + :type write_behavior: any + :param upsert_settings: SQL upsert settings. + :type upsert_settings: ~azure.mgmt.datafactory.models.SqlUpsertSettings """ _validation = { @@ -38358,6 +38906,9 @@ class SqlSink(CopySink): 'stored_procedure_parameters': {'key': 'storedProcedureParameters', 'type': '{StoredProcedureParameter}'}, 'stored_procedure_table_type_parameter_name': {'key': 'storedProcedureTableTypeParameterName', 'type': 'object'}, 'table_option': {'key': 'tableOption', 'type': 'object'}, + 'sql_writer_use_table_lock': {'key': 'sqlWriterUseTableLock', 'type': 'object'}, + 'write_behavior': {'key': 'writeBehavior', 'type': 'object'}, + 'upsert_settings': {'key': 'upsertSettings', 'type': 'SqlUpsertSettings'}, } def __init__( @@ -38376,6 +38927,9 @@ def __init__( stored_procedure_parameters: Optional[Dict[str, "StoredProcedureParameter"]] = None, stored_procedure_table_type_parameter_name: Optional[Any] = None, table_option: Optional[Any] = None, + sql_writer_use_table_lock: Optional[Any] = None, + write_behavior: Optional[Any] = None, + upsert_settings: Optional["SqlUpsertSettings"] = None, **kwargs ): super(SqlSink, self).__init__(additional_properties=additional_properties, write_batch_size=write_batch_size, write_batch_timeout=write_batch_timeout, sink_retry_count=sink_retry_count, sink_retry_wait=sink_retry_wait, max_concurrent_connections=max_concurrent_connections, disable_metrics_collection=disable_metrics_collection, **kwargs) @@ -38386,6 +38940,9 @@ def __init__( self.stored_procedure_parameters = stored_procedure_parameters self.stored_procedure_table_type_parameter_name = stored_procedure_table_type_parameter_name self.table_option = table_option + self.sql_writer_use_table_lock = sql_writer_use_table_lock + self.write_behavior = write_behavior + self.upsert_settings = upsert_settings class SqlSource(TabularSource): @@ -38414,8 +38971,8 @@ class SqlSource(TabularSource): pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type query_timeout: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any :param sql_reader_query: SQL reader query. Type: string (or Expression with resultType string). :type sql_reader_query: any :param sql_reader_stored_procedure_name: Name of the stored procedure for a SQL Database @@ -38449,7 +39006,7 @@ class SqlSource(TabularSource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, 'sql_reader_query': {'key': 'sqlReaderQuery', 'type': 'object'}, 'sql_reader_stored_procedure_name': {'key': 'sqlReaderStoredProcedureName', 'type': 'object'}, 'stored_procedure_parameters': {'key': 'storedProcedureParameters', 'type': '{StoredProcedureParameter}'}, @@ -38467,7 +39024,7 @@ def __init__( max_concurrent_connections: Optional[Any] = None, disable_metrics_collection: Optional[Any] = None, query_timeout: Optional[Any] = None, - additional_columns: Optional[List["AdditionalColumns"]] = None, + additional_columns: Optional[Any] = None, sql_reader_query: Optional[Any] = None, sql_reader_stored_procedure_name: Optional[Any] = None, stored_procedure_parameters: Optional[Dict[str, "StoredProcedureParameter"]] = None, @@ -38486,6 +39043,40 @@ def __init__( self.partition_settings = partition_settings +class SqlUpsertSettings(msrest.serialization.Model): + """Sql upsert option settings. + + :param use_temp_db: Specifies whether to use temp db for upsert interim table. Type: boolean + (or Expression with resultType boolean). + :type use_temp_db: any + :param interim_schema_name: Schema name for interim table. Type: string (or Expression with + resultType string). + :type interim_schema_name: any + :param keys: Key column names for unique row identification. Type: array of strings (or + Expression with resultType array of strings). + :type keys: any + """ + + _attribute_map = { + 'use_temp_db': {'key': 'useTempDB', 'type': 'object'}, + 'interim_schema_name': {'key': 'interimSchemaName', 'type': 'object'}, + 'keys': {'key': 'keys', 'type': 'object'}, + } + + def __init__( + self, + *, + use_temp_db: Optional[Any] = None, + interim_schema_name: Optional[Any] = None, + keys: Optional[Any] = None, + **kwargs + ): + super(SqlUpsertSettings, self).__init__(**kwargs) + self.use_temp_db = use_temp_db + self.interim_schema_name = interim_schema_name + self.keys = keys + + class SquareLinkedService(LinkedService): """Square Service linked service. @@ -38680,8 +39271,8 @@ class SquareSource(TabularSource): pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type query_timeout: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any :param query: A query to retrieve data from source. Type: string (or Expression with resultType string). :type query: any @@ -38699,7 +39290,7 @@ class SquareSource(TabularSource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, 'query': {'key': 'query', 'type': 'object'}, } @@ -38712,7 +39303,7 @@ def __init__( max_concurrent_connections: Optional[Any] = None, disable_metrics_collection: Optional[Any] = None, query_timeout: Optional[Any] = None, - additional_columns: Optional[List["AdditionalColumns"]] = None, + additional_columns: Optional[Any] = None, query: Optional[Any] = None, **kwargs ): @@ -39785,8 +40376,8 @@ class SybaseSource(TabularSource): pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type query_timeout: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any :param query: Database query. Type: string (or Expression with resultType string). :type query: any """ @@ -39803,7 +40394,7 @@ class SybaseSource(TabularSource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, 'query': {'key': 'query', 'type': 'object'}, } @@ -39816,7 +40407,7 @@ def __init__( max_concurrent_connections: Optional[Any] = None, disable_metrics_collection: Optional[Any] = None, query_timeout: Optional[Any] = None, - additional_columns: Optional[List["AdditionalColumns"]] = None, + additional_columns: Optional[Any] = None, query: Optional[Any] = None, **kwargs ): @@ -40189,8 +40780,8 @@ class TeradataSource(TabularSource): pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type query_timeout: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any :param query: Teradata query. Type: string (or Expression with resultType string). :type query: any :param partition_option: The partition mechanism that will be used for teradata read in @@ -40213,7 +40804,7 @@ class TeradataSource(TabularSource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, 'query': {'key': 'query', 'type': 'object'}, 'partition_option': {'key': 'partitionOption', 'type': 'object'}, 'partition_settings': {'key': 'partitionSettings', 'type': 'TeradataPartitionSettings'}, @@ -40228,7 +40819,7 @@ def __init__( max_concurrent_connections: Optional[Any] = None, disable_metrics_collection: Optional[Any] = None, query_timeout: Optional[Any] = None, - additional_columns: Optional[List["AdditionalColumns"]] = None, + additional_columns: Optional[Any] = None, query: Optional[Any] = None, partition_option: Optional[Any] = None, partition_settings: Optional["TeradataPartitionSettings"] = None, @@ -41398,8 +41989,8 @@ class VerticaSource(TabularSource): pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type query_timeout: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any :param query: A query to retrieve data from source. Type: string (or Expression with resultType string). :type query: any @@ -41417,7 +42008,7 @@ class VerticaSource(TabularSource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, 'query': {'key': 'query', 'type': 'object'}, } @@ -41430,7 +42021,7 @@ def __init__( max_concurrent_connections: Optional[Any] = None, disable_metrics_collection: Optional[Any] = None, query_timeout: Optional[Any] = None, - additional_columns: Optional[List["AdditionalColumns"]] = None, + additional_columns: Optional[Any] = None, query: Optional[Any] = None, **kwargs ): @@ -42072,8 +42663,8 @@ class WebSource(CopySource): false. Type: boolean (or Expression with resultType boolean). :type disable_metrics_collection: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any """ _validation = { @@ -42087,7 +42678,7 @@ class WebSource(CopySource): 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, } def __init__( @@ -42098,7 +42689,7 @@ def __init__( source_retry_wait: Optional[Any] = None, max_concurrent_connections: Optional[Any] = None, disable_metrics_collection: Optional[Any] = None, - additional_columns: Optional[List["AdditionalColumns"]] = None, + additional_columns: Optional[Any] = None, **kwargs ): super(WebSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, max_concurrent_connections=max_concurrent_connections, disable_metrics_collection=disable_metrics_collection, **kwargs) @@ -42182,6 +42773,55 @@ def __init__( self.path = path +class WranglingDataFlow(DataFlow): + """Power Query data flow. + + All required parameters must be populated in order to send to Azure. + + :param type: Required. Type of data flow.Constant filled by server. + :type type: str + :param description: The description of the data flow. + :type description: str + :param annotations: List of tags that can be used for describing the data flow. + :type annotations: list[any] + :param folder: The folder that this data flow is in. If not specified, Data flow will appear at + the root level. + :type folder: ~azure.mgmt.datafactory.models.DataFlowFolder + :param sources: List of sources in Power Query. + :type sources: list[~azure.mgmt.datafactory.models.PowerQuerySource] + :param script: Power query mashup script. + :type script: str + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'type': {'key': 'type', 'type': 'str'}, + 'description': {'key': 'description', 'type': 'str'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'folder': {'key': 'folder', 'type': 'DataFlowFolder'}, + 'sources': {'key': 'typeProperties.sources', 'type': '[PowerQuerySource]'}, + 'script': {'key': 'typeProperties.script', 'type': 'str'}, + } + + def __init__( + self, + *, + description: Optional[str] = None, + annotations: Optional[List[Any]] = None, + folder: Optional["DataFlowFolder"] = None, + sources: Optional[List["PowerQuerySource"]] = None, + script: Optional[str] = None, + **kwargs + ): + super(WranglingDataFlow, self).__init__(description=description, annotations=annotations, folder=folder, **kwargs) + self.type = 'WranglingDataFlow' # type: str + self.sources = sources + self.script = script + + class XeroLinkedService(LinkedService): """Xero Service linked service. @@ -42372,8 +43012,8 @@ class XeroSource(TabularSource): pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type query_timeout: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any :param query: A query to retrieve data from source. Type: string (or Expression with resultType string). :type query: any @@ -42391,7 +43031,7 @@ class XeroSource(TabularSource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, 'query': {'key': 'query', 'type': 'object'}, } @@ -42404,7 +43044,7 @@ def __init__( max_concurrent_connections: Optional[Any] = None, disable_metrics_collection: Optional[Any] = None, query_timeout: Optional[Any] = None, - additional_columns: Optional[List["AdditionalColumns"]] = None, + additional_columns: Optional[Any] = None, query: Optional[Any] = None, **kwargs ): @@ -42589,8 +43229,8 @@ class XmlSource(CopySource): :param format_settings: Xml format settings. :type format_settings: ~azure.mgmt.datafactory.models.XmlReadSettings :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any """ _validation = { @@ -42606,7 +43246,7 @@ class XmlSource(CopySource): 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'store_settings': {'key': 'storeSettings', 'type': 'StoreReadSettings'}, 'format_settings': {'key': 'formatSettings', 'type': 'XmlReadSettings'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, } def __init__( @@ -42619,7 +43259,7 @@ def __init__( disable_metrics_collection: Optional[Any] = None, store_settings: Optional["StoreReadSettings"] = None, format_settings: Optional["XmlReadSettings"] = None, - additional_columns: Optional[List["AdditionalColumns"]] = None, + additional_columns: Optional[Any] = None, **kwargs ): super(XmlSource, self).__init__(additional_properties=additional_properties, source_retry_count=source_retry_count, source_retry_wait=source_retry_wait, max_concurrent_connections=max_concurrent_connections, disable_metrics_collection=disable_metrics_collection, **kwargs) @@ -42849,8 +43489,8 @@ class ZohoSource(TabularSource): pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). :type query_timeout: any :param additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects (or Expression with resultType array of objects). - :type additional_columns: list[~azure.mgmt.datafactory.models.AdditionalColumns] + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :type additional_columns: any :param query: A query to retrieve data from source. Type: string (or Expression with resultType string). :type query: any @@ -42868,7 +43508,7 @@ class ZohoSource(TabularSource): 'max_concurrent_connections': {'key': 'maxConcurrentConnections', 'type': 'object'}, 'disable_metrics_collection': {'key': 'disableMetricsCollection', 'type': 'object'}, 'query_timeout': {'key': 'queryTimeout', 'type': 'object'}, - 'additional_columns': {'key': 'additionalColumns', 'type': '[AdditionalColumns]'}, + 'additional_columns': {'key': 'additionalColumns', 'type': 'object'}, 'query': {'key': 'query', 'type': 'object'}, } @@ -42881,7 +43521,7 @@ def __init__( max_concurrent_connections: Optional[Any] = None, disable_metrics_collection: Optional[Any] = None, query_timeout: Optional[Any] = None, - additional_columns: Optional[List["AdditionalColumns"]] = None, + additional_columns: Optional[Any] = None, query: Optional[Any] = None, **kwargs ): diff --git a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/_integration_runtimes_operations.py b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/_integration_runtimes_operations.py index 9d19f1015005..2a61d232c982 100644 --- a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/_integration_runtimes_operations.py +++ b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/operations/_integration_runtimes_operations.py @@ -465,6 +465,69 @@ def get_status( return deserialized get_status.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.DataFactory/factories/{factoryName}/integrationRuntimes/{integrationRuntimeName}/getStatus'} # type: ignore + def list_outbound_network_dependencies_endpoints( + self, + resource_group_name, # type: str + factory_name, # type: str + integration_runtime_name, # type: str + **kwargs # type: Any + ): + # type: (...) -> "_models.IntegrationRuntimeOutboundNetworkDependenciesEndpointsResponse" + """Gets the list of outbound network dependencies for a given Azure-SSIS integration runtime. + + :param resource_group_name: The resource group name. + :type resource_group_name: str + :param factory_name: The factory name. + :type factory_name: str + :param integration_runtime_name: The integration runtime name. + :type integration_runtime_name: str + :keyword callable cls: A custom type or function that will be passed the direct response + :return: IntegrationRuntimeOutboundNetworkDependenciesEndpointsResponse, or the result of cls(response) + :rtype: ~azure.mgmt.datafactory.models.IntegrationRuntimeOutboundNetworkDependenciesEndpointsResponse + :raises: ~azure.core.exceptions.HttpResponseError + """ + cls = kwargs.pop('cls', None) # type: ClsType["_models.IntegrationRuntimeOutboundNetworkDependenciesEndpointsResponse"] + error_map = { + 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError + } + error_map.update(kwargs.pop('error_map', {})) + api_version = "2018-06-01" + accept = "application/json" + + # Construct URL + url = self.list_outbound_network_dependencies_endpoints.metadata['url'] # type: ignore + path_format_arguments = { + 'subscriptionId': self._serialize.url("self._config.subscription_id", self._config.subscription_id, 'str'), + 'resourceGroupName': self._serialize.url("resource_group_name", resource_group_name, 'str', max_length=90, min_length=1, pattern=r'^[-\w\._\(\)]+$'), + 'factoryName': self._serialize.url("factory_name", factory_name, 'str', max_length=63, min_length=3, pattern=r'^[A-Za-z0-9]+(?:-[A-Za-z0-9]+)*$'), + 'integrationRuntimeName': self._serialize.url("integration_runtime_name", integration_runtime_name, 'str', max_length=63, min_length=3, pattern=r'^[A-Za-z0-9]+(?:-[A-Za-z0-9]+)*$'), + } + url = self._client.format_url(url, **path_format_arguments) + + # Construct parameters + query_parameters = {} # type: Dict[str, Any] + query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') + + # Construct headers + header_parameters = {} # type: Dict[str, Any] + header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') + + request = self._client.get(url, query_parameters, header_parameters) + pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs) + response = pipeline_response.http_response + + if response.status_code not in [200]: + map_error(status_code=response.status_code, response=response, error_map=error_map) + raise HttpResponseError(response=response, error_format=ARMErrorFormat) + + deserialized = self._deserialize('IntegrationRuntimeOutboundNetworkDependenciesEndpointsResponse', pipeline_response) + + if cls: + return cls(pipeline_response, deserialized, {}) + + return deserialized + list_outbound_network_dependencies_endpoints.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.DataFactory/factories/{factoryName}/integrationRuntimes/{integrationRuntimeName}/outboundNetworkDependenciesEndpoints'} # type: ignore + def get_connection_info( self, resource_group_name, # type: str