From d301c272d86df09d34e22a4a194db731290aa591 Mon Sep 17 00:00:00 2001 From: SDKAuto Date: Wed, 17 Jan 2024 06:53:57 +0000 Subject: [PATCH] CodeGen from PR 27412 in Azure/azure-rest-api-specs Merge d5fd66ba9d2f9fe255715d36381db680fdac68ab into fa40a4232663140524a7d7d93becb53597090f55 --- .../azure-mgmt-datafactory/_meta.json | 4 +- .../azure/mgmt/datafactory/_version.py | 2 +- .../azure/mgmt/datafactory/models/__init__.py | 36 + .../_data_factory_management_client_enums.py | 15 + .../mgmt/datafactory/models/_models_py3.py | 3488 ++++++++++++++--- 5 files changed, 2907 insertions(+), 638 deletions(-) diff --git a/sdk/datafactory/azure-mgmt-datafactory/_meta.json b/sdk/datafactory/azure-mgmt-datafactory/_meta.json index 041d591b083e..737f4ab78f08 100644 --- a/sdk/datafactory/azure-mgmt-datafactory/_meta.json +++ b/sdk/datafactory/azure-mgmt-datafactory/_meta.json @@ -1,11 +1,11 @@ { - "commit": "ed84b11847785792767b0b84cc6f98f4ea08ca77", + "commit": "8d83e9affbd67e7b79a6423ea095fba1623c98ea", "repository_url": "https://github.com/Azure/azure-rest-api-specs", "autorest": "3.9.7", "use": [ "@autorest/python@6.7.1", "@autorest/modelerfour@4.26.2" ], - "autorest_command": "autorest specification/datafactory/resource-manager/readme.md --generate-sample=True --include-x-ms-examples-original-file=True --python --python-sdks-folder=/home/vsts/work/1/azure-sdk-for-python/sdk --use=@autorest/python@6.7.1 --use=@autorest/modelerfour@4.26.2 --version=3.9.7 --version-tolerant=False", + "autorest_command": "autorest specification/datafactory/resource-manager/readme.md --generate-sample=True --include-x-ms-examples-original-file=True --python --python-sdks-folder=/mnt/vss/_work/1/s/azure-sdk-for-python/sdk --use=@autorest/python@6.7.1 --use=@autorest/modelerfour@4.26.2 --version=3.9.7 --version-tolerant=False", "readme": "specification/datafactory/resource-manager/readme.md" } \ No newline at end of file diff --git a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/_version.py b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/_version.py index 77f53a3589c6..c47f66669f1b 100644 --- a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/_version.py +++ b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/_version.py @@ -6,4 +6,4 @@ # Changes may cause incorrect behavior and will be lost if the code is regenerated. # -------------------------------------------------------------------------- -VERSION = "4.0.0" +VERSION = "1.0.0" diff --git a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py index 702f57a07cff..8387fcc75d17 100644 --- a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py +++ b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py @@ -587,8 +587,16 @@ from ._models_py3 import SalesforceServiceCloudObjectDataset from ._models_py3 import SalesforceServiceCloudSink from ._models_py3 import SalesforceServiceCloudSource +from ._models_py3 import SalesforceServiceCloudV2LinkedService +from ._models_py3 import SalesforceServiceCloudV2ObjectDataset +from ._models_py3 import SalesforceServiceCloudV2Sink +from ._models_py3 import SalesforceServiceCloudV2Source from ._models_py3 import SalesforceSink from ._models_py3 import SalesforceSource +from ._models_py3 import SalesforceV2LinkedService +from ._models_py3 import SalesforceV2ObjectDataset +from ._models_py3 import SalesforceV2Sink +from ._models_py3 import SalesforceV2Source from ._models_py3 import SapBWLinkedService from ._models_py3 import SapBwCubeDataset from ._models_py3 import SapBwSource @@ -650,6 +658,10 @@ from ._models_py3 import SnowflakeLinkedService from ._models_py3 import SnowflakeSink from ._models_py3 import SnowflakeSource +from ._models_py3 import SnowflakeV2Dataset +from ._models_py3 import SnowflakeV2LinkedService +from ._models_py3 import SnowflakeV2Sink +from ._models_py3 import SnowflakeV2Source from ._models_py3 import SparkConfigurationParametrizationReference from ._models_py3 import SparkLinkedService from ._models_py3 import SparkObjectDataset @@ -734,6 +746,10 @@ from ._models_py3 import VerticaSource from ._models_py3 import VerticaTableDataset from ._models_py3 import WaitActivity +from ._models_py3 import WarehouseLinkedService +from ._models_py3 import WarehouseSink +from ._models_py3 import WarehouseSource +from ._models_py3 import WarehouseTableDataset from ._models_py3 import WebActivity from ._models_py3 import WebActivityAuthentication from ._models_py3 import WebAnonymousAuthentication @@ -839,6 +855,7 @@ from ._data_factory_management_client_enums import RunQueryOrderByField from ._data_factory_management_client_enums import SalesforceSinkWriteBehavior from ._data_factory_management_client_enums import SalesforceSourceReadBehavior +from ._data_factory_management_client_enums import SalesforceV2SinkWriteBehavior from ._data_factory_management_client_enums import SapCloudForCustomerSinkWriteBehavior from ._data_factory_management_client_enums import SapHanaAuthenticationType from ._data_factory_management_client_enums import SapHanaPartitionOption @@ -851,6 +868,7 @@ from ._data_factory_management_client_enums import ServiceNowAuthenticationType from ._data_factory_management_client_enums import ServicePrincipalCredentialType from ._data_factory_management_client_enums import SftpAuthenticationType +from ._data_factory_management_client_enums import SnowflakeAuthenticationType from ._data_factory_management_client_enums import SparkAuthenticationType from ._data_factory_management_client_enums import SparkConfigurationReferenceType from ._data_factory_management_client_enums import SparkJobReferenceType @@ -1464,8 +1482,16 @@ "SalesforceServiceCloudObjectDataset", "SalesforceServiceCloudSink", "SalesforceServiceCloudSource", + "SalesforceServiceCloudV2LinkedService", + "SalesforceServiceCloudV2ObjectDataset", + "SalesforceServiceCloudV2Sink", + "SalesforceServiceCloudV2Source", "SalesforceSink", "SalesforceSource", + "SalesforceV2LinkedService", + "SalesforceV2ObjectDataset", + "SalesforceV2Sink", + "SalesforceV2Source", "SapBWLinkedService", "SapBwCubeDataset", "SapBwSource", @@ -1527,6 +1553,10 @@ "SnowflakeLinkedService", "SnowflakeSink", "SnowflakeSource", + "SnowflakeV2Dataset", + "SnowflakeV2LinkedService", + "SnowflakeV2Sink", + "SnowflakeV2Source", "SparkConfigurationParametrizationReference", "SparkLinkedService", "SparkObjectDataset", @@ -1611,6 +1641,10 @@ "VerticaSource", "VerticaTableDataset", "WaitActivity", + "WarehouseLinkedService", + "WarehouseSink", + "WarehouseSource", + "WarehouseTableDataset", "WebActivity", "WebActivityAuthentication", "WebAnonymousAuthentication", @@ -1715,6 +1749,7 @@ "RunQueryOrderByField", "SalesforceSinkWriteBehavior", "SalesforceSourceReadBehavior", + "SalesforceV2SinkWriteBehavior", "SapCloudForCustomerSinkWriteBehavior", "SapHanaAuthenticationType", "SapHanaPartitionOption", @@ -1727,6 +1762,7 @@ "ServiceNowAuthenticationType", "ServicePrincipalCredentialType", "SftpAuthenticationType", + "SnowflakeAuthenticationType", "SparkAuthenticationType", "SparkConfigurationReferenceType", "SparkJobReferenceType", diff --git a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/_data_factory_management_client_enums.py b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/_data_factory_management_client_enums.py index b0ad5505dd4a..7f01ed8b6fe7 100644 --- a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/_data_factory_management_client_enums.py +++ b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/_data_factory_management_client_enums.py @@ -721,6 +721,13 @@ class SalesforceSourceReadBehavior(str, Enum, metaclass=CaseInsensitiveEnumMeta) QUERY_ALL = "QueryAll" +class SalesforceV2SinkWriteBehavior(str, Enum, metaclass=CaseInsensitiveEnumMeta): + """The write behavior for the operation. Default is Insert.""" + + INSERT = "Insert" + UPSERT = "Upsert" + + class SapCloudForCustomerSinkWriteBehavior(str, Enum, metaclass=CaseInsensitiveEnumMeta): """The write behavior for the operation. Default is 'Insert'.""" @@ -827,6 +834,14 @@ class SftpAuthenticationType(str, Enum, metaclass=CaseInsensitiveEnumMeta): MULTI_FACTOR = "MultiFactor" +class SnowflakeAuthenticationType(str, Enum, metaclass=CaseInsensitiveEnumMeta): + """The type used for authentication. Type: string.""" + + BASIC = "Basic" + KEY_PAIR = "KeyPair" + AAD_SERVICE_PRINCIPAL = "AADServicePrincipal" + + class SparkAuthenticationType(str, Enum, metaclass=CaseInsensitiveEnumMeta): """The authentication method used to access the Spark server.""" diff --git a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/_models_py3.py b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/_models_py3.py index 6b8186c03b06..5c984d7d254e 100644 --- a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/_models_py3.py +++ b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/_models_py3.py @@ -494,14 +494,15 @@ class LinkedService(_serialization.Model): PhoenixLinkedService, PostgreSqlLinkedService, PrestoLinkedService, QuickBooksLinkedService, QuickbaseLinkedService, ResponsysLinkedService, RestServiceLinkedService, SalesforceLinkedService, SalesforceMarketingCloudLinkedService, - SalesforceServiceCloudLinkedService, SapBWLinkedService, SapCloudForCustomerLinkedService, + SalesforceServiceCloudLinkedService, SalesforceServiceCloudV2LinkedService, + SalesforceV2LinkedService, SapBWLinkedService, SapCloudForCustomerLinkedService, SapEccLinkedService, SapHanaLinkedService, SapOdpLinkedService, SapOpenHubLinkedService, SapTableLinkedService, ServiceNowLinkedService, SftpServerLinkedService, SharePointOnlineListLinkedService, ShopifyLinkedService, SmartsheetLinkedService, - SnowflakeLinkedService, SparkLinkedService, SqlServerLinkedService, SquareLinkedService, - SybaseLinkedService, TeamDeskLinkedService, TeradataLinkedService, TwilioLinkedService, - VerticaLinkedService, WebLinkedService, XeroLinkedService, ZendeskLinkedService, - ZohoLinkedService + SnowflakeLinkedService, SnowflakeV2LinkedService, SparkLinkedService, SqlServerLinkedService, + SquareLinkedService, SybaseLinkedService, TeamDeskLinkedService, TeradataLinkedService, + TwilioLinkedService, VerticaLinkedService, WarehouseLinkedService, WebLinkedService, + XeroLinkedService, ZendeskLinkedService, ZohoLinkedService All required parameters must be populated in order to send to Azure. @@ -624,6 +625,8 @@ class LinkedService(_serialization.Model): "Salesforce": "SalesforceLinkedService", "SalesforceMarketingCloud": "SalesforceMarketingCloudLinkedService", "SalesforceServiceCloud": "SalesforceServiceCloudLinkedService", + "SalesforceServiceCloudV2": "SalesforceServiceCloudV2LinkedService", + "SalesforceV2": "SalesforceV2LinkedService", "SapBW": "SapBWLinkedService", "SapCloudForCustomer": "SapCloudForCustomerLinkedService", "SapEcc": "SapEccLinkedService", @@ -637,6 +640,7 @@ class LinkedService(_serialization.Model): "Shopify": "ShopifyLinkedService", "Smartsheet": "SmartsheetLinkedService", "Snowflake": "SnowflakeLinkedService", + "SnowflakeV2": "SnowflakeV2LinkedService", "Spark": "SparkLinkedService", "SqlServer": "SqlServerLinkedService", "Square": "SquareLinkedService", @@ -645,6 +649,7 @@ class LinkedService(_serialization.Model): "Teradata": "TeradataLinkedService", "Twilio": "TwilioLinkedService", "Vertica": "VerticaLinkedService", + "Warehouse": "WarehouseLinkedService", "Web": "WebLinkedService", "Xero": "XeroLinkedService", "Zendesk": "ZendeskLinkedService", @@ -865,13 +870,14 @@ class Dataset(_serialization.Model): OracleTableDataset, OrcDataset, ParquetDataset, PaypalObjectDataset, PhoenixObjectDataset, PostgreSqlTableDataset, PrestoObjectDataset, QuickBooksObjectDataset, RelationalTableDataset, ResponsysObjectDataset, RestResourceDataset, SalesforceMarketingCloudObjectDataset, - SalesforceObjectDataset, SalesforceServiceCloudObjectDataset, SapBwCubeDataset, + SalesforceObjectDataset, SalesforceServiceCloudObjectDataset, + SalesforceServiceCloudV2ObjectDataset, SalesforceV2ObjectDataset, SapBwCubeDataset, SapCloudForCustomerResourceDataset, SapEccResourceDataset, SapHanaTableDataset, SapOdpResourceDataset, SapOpenHubTableDataset, SapTableResourceDataset, ServiceNowObjectDataset, SharePointOnlineListResourceDataset, ShopifyObjectDataset, - SnowflakeDataset, SparkObjectDataset, SqlServerTableDataset, SquareObjectDataset, - SybaseTableDataset, TeradataTableDataset, VerticaTableDataset, WebTableDataset, - XeroObjectDataset, XmlDataset, ZohoObjectDataset + SnowflakeDataset, SnowflakeV2Dataset, SparkObjectDataset, SqlServerTableDataset, + SquareObjectDataset, SybaseTableDataset, TeradataTableDataset, VerticaTableDataset, + WarehouseTableDataset, WebTableDataset, XeroObjectDataset, XmlDataset, ZohoObjectDataset All required parameters must be populated in order to send to Azure. @@ -994,6 +1000,8 @@ class Dataset(_serialization.Model): "SalesforceMarketingCloudObject": "SalesforceMarketingCloudObjectDataset", "SalesforceObject": "SalesforceObjectDataset", "SalesforceServiceCloudObject": "SalesforceServiceCloudObjectDataset", + "SalesforceServiceCloudV2Object": "SalesforceServiceCloudV2ObjectDataset", + "SalesforceV2Object": "SalesforceV2ObjectDataset", "SapBwCube": "SapBwCubeDataset", "SapCloudForCustomerResource": "SapCloudForCustomerResourceDataset", "SapEccResource": "SapEccResourceDataset", @@ -1005,12 +1013,14 @@ class Dataset(_serialization.Model): "SharePointOnlineListResource": "SharePointOnlineListResourceDataset", "ShopifyObject": "ShopifyObjectDataset", "SnowflakeTable": "SnowflakeDataset", + "SnowflakeV2Table": "SnowflakeV2Dataset", "SparkObject": "SparkObjectDataset", "SqlServerTable": "SqlServerTableDataset", "SquareObject": "SquareObjectDataset", "SybaseTable": "SybaseTableDataset", "TeradataTable": "TeradataTableDataset", "VerticaTable": "VerticaTableDataset", + "WarehouseTable": "WarehouseTableDataset", "WebTable": "WebTableDataset", "XeroObject": "XeroObjectDataset", "Xml": "XmlDataset", @@ -1178,8 +1188,8 @@ class CopySource(_serialization.Model): ExcelSource, FileSystemSource, HdfsSource, HttpSource, JsonSource, LakeHouseTableSource, MicrosoftAccessSource, MongoDbAtlasSource, MongoDbSource, MongoDbV2Source, ODataSource, Office365Source, OracleSource, OrcSource, ParquetSource, RelationalSource, RestSource, - SalesforceServiceCloudSource, SharePointOnlineListSource, SnowflakeSource, TabularSource, - WebSource, XmlSource + SalesforceServiceCloudSource, SalesforceServiceCloudV2Source, SharePointOnlineListSource, + SnowflakeSource, SnowflakeV2Source, TabularSource, WebSource, XmlSource All required parameters must be populated in order to send to Azure. @@ -1250,8 +1260,10 @@ class CopySource(_serialization.Model): "RelationalSource": "RelationalSource", "RestSource": "RestSource", "SalesforceServiceCloudSource": "SalesforceServiceCloudSource", + "SalesforceServiceCloudV2Source": "SalesforceServiceCloudV2Source", "SharePointOnlineListSource": "SharePointOnlineListSource", "SnowflakeSource": "SnowflakeSource", + "SnowflakeV2Source": "SnowflakeV2Source", "TabularSource": "TabularSource", "WebSource": "WebSource", "XmlSource": "XmlSource", @@ -1305,10 +1317,11 @@ class TabularSource(CopySource): HubspotSource, ImpalaSource, InformixSource, JiraSource, MagentoSource, MariaDBSource, MarketoSource, MySqlSource, NetezzaSource, OdbcSource, OracleServiceCloudSource, PaypalSource, PhoenixSource, PostgreSqlSource, PrestoSource, QuickBooksSource, ResponsysSource, - SalesforceMarketingCloudSource, SalesforceSource, SapBwSource, SapCloudForCustomerSource, - SapEccSource, SapHanaSource, SapOdpSource, SapOpenHubSource, SapTableSource, ServiceNowSource, - ShopifySource, SparkSource, SqlDWSource, SqlMISource, SqlServerSource, SqlSource, SquareSource, - SybaseSource, TeradataSource, VerticaSource, XeroSource, ZohoSource + SalesforceMarketingCloudSource, SalesforceSource, SalesforceV2Source, SapBwSource, + SapCloudForCustomerSource, SapEccSource, SapHanaSource, SapOdpSource, SapOpenHubSource, + SapTableSource, ServiceNowSource, ShopifySource, SparkSource, SqlDWSource, SqlMISource, + SqlServerSource, SqlSource, SquareSource, SybaseSource, TeradataSource, VerticaSource, + WarehouseSource, XeroSource, ZohoSource All required parameters must be populated in order to send to Azure. @@ -1393,6 +1406,7 @@ class TabularSource(CopySource): "ResponsysSource": "ResponsysSource", "SalesforceMarketingCloudSource": "SalesforceMarketingCloudSource", "SalesforceSource": "SalesforceSource", + "SalesforceV2Source": "SalesforceV2Source", "SapBwSource": "SapBwSource", "SapCloudForCustomerSource": "SapCloudForCustomerSource", "SapEccSource": "SapEccSource", @@ -1411,6 +1425,7 @@ class TabularSource(CopySource): "SybaseSource": "SybaseSource", "TeradataSource": "TeradataSource", "VerticaSource": "VerticaSource", + "WarehouseSource": "WarehouseSource", "XeroSource": "XeroSource", "ZohoSource": "ZohoSource", } @@ -4344,8 +4359,9 @@ class CopySink(_serialization.Model): CommonDataServiceForAppsSink, CosmosDbMongoDbApiSink, CosmosDbSqlApiSink, DelimitedTextSink, DocumentDbCollectionSink, DynamicsCrmSink, DynamicsSink, FileSystemSink, InformixSink, JsonSink, LakeHouseTableSink, MicrosoftAccessSink, MongoDbAtlasSink, MongoDbV2Sink, OdbcSink, - OracleSink, OrcSink, ParquetSink, RestSink, SalesforceServiceCloudSink, SalesforceSink, - SapCloudForCustomerSink, SnowflakeSink, SqlDWSink, SqlMISink, SqlServerSink, SqlSink + OracleSink, OrcSink, ParquetSink, RestSink, SalesforceServiceCloudSink, + SalesforceServiceCloudV2Sink, SalesforceSink, SalesforceV2Sink, SapCloudForCustomerSink, + SnowflakeSink, SnowflakeV2Sink, SqlDWSink, SqlMISink, SqlServerSink, SqlSink, WarehouseSink All required parameters must be populated in order to send to Azure. @@ -4424,13 +4440,17 @@ class CopySink(_serialization.Model): "ParquetSink": "ParquetSink", "RestSink": "RestSink", "SalesforceServiceCloudSink": "SalesforceServiceCloudSink", + "SalesforceServiceCloudV2Sink": "SalesforceServiceCloudV2Sink", "SalesforceSink": "SalesforceSink", + "SalesforceV2Sink": "SalesforceV2Sink", "SapCloudForCustomerSink": "SapCloudForCustomerSink", "SnowflakeSink": "SnowflakeSink", + "SnowflakeV2Sink": "SnowflakeV2Sink", "SqlDWSink": "SqlDWSink", "SqlMISink": "SqlMISink", "SqlServerSink": "SqlServerSink", "SqlSink": "SqlSink", + "WarehouseSink": "WarehouseSink", } } @@ -5889,6 +5909,9 @@ class StoreWriteSettings(_serialization.Model): :vartype disable_metrics_collection: JSON :ivar copy_behavior: The type of copy behavior for copy sink. :vartype copy_behavior: JSON + :ivar metadata: Specify the custom metadata to be added to sink data. Type: array of objects + (or Expression with resultType array of objects). + :vartype metadata: list[~azure.mgmt.datafactory.models.MetadataItem] """ _validation = { @@ -5901,6 +5924,7 @@ class StoreWriteSettings(_serialization.Model): "max_concurrent_connections": {"key": "maxConcurrentConnections", "type": "object"}, "disable_metrics_collection": {"key": "disableMetricsCollection", "type": "object"}, "copy_behavior": {"key": "copyBehavior", "type": "object"}, + "metadata": {"key": "metadata", "type": "[MetadataItem]"}, } _subtype_map = { @@ -5922,6 +5946,7 @@ def __init__( max_concurrent_connections: Optional[JSON] = None, disable_metrics_collection: Optional[JSON] = None, copy_behavior: Optional[JSON] = None, + metadata: Optional[List["_models.MetadataItem"]] = None, **kwargs: Any ) -> None: """ @@ -5936,6 +5961,9 @@ def __init__( :paramtype disable_metrics_collection: JSON :keyword copy_behavior: The type of copy behavior for copy sink. :paramtype copy_behavior: JSON + :keyword metadata: Specify the custom metadata to be added to sink data. Type: array of objects + (or Expression with resultType array of objects). + :paramtype metadata: list[~azure.mgmt.datafactory.models.MetadataItem] """ super().__init__(**kwargs) self.additional_properties = additional_properties @@ -5943,6 +5971,7 @@ def __init__( self.max_concurrent_connections = max_concurrent_connections self.disable_metrics_collection = disable_metrics_collection self.copy_behavior = copy_behavior + self.metadata = metadata class AzureBlobFSWriteSettings(StoreWriteSettings): @@ -5963,6 +5992,9 @@ class AzureBlobFSWriteSettings(StoreWriteSettings): :vartype disable_metrics_collection: JSON :ivar copy_behavior: The type of copy behavior for copy sink. :vartype copy_behavior: JSON + :ivar metadata: Specify the custom metadata to be added to sink data. Type: array of objects + (or Expression with resultType array of objects). + :vartype metadata: list[~azure.mgmt.datafactory.models.MetadataItem] :ivar block_size_in_mb: Indicates the block size(MB) when writing data to blob. Type: integer (or Expression with resultType integer). :vartype block_size_in_mb: JSON @@ -5978,6 +6010,7 @@ class AzureBlobFSWriteSettings(StoreWriteSettings): "max_concurrent_connections": {"key": "maxConcurrentConnections", "type": "object"}, "disable_metrics_collection": {"key": "disableMetricsCollection", "type": "object"}, "copy_behavior": {"key": "copyBehavior", "type": "object"}, + "metadata": {"key": "metadata", "type": "[MetadataItem]"}, "block_size_in_mb": {"key": "blockSizeInMB", "type": "object"}, } @@ -5988,6 +6021,7 @@ def __init__( max_concurrent_connections: Optional[JSON] = None, disable_metrics_collection: Optional[JSON] = None, copy_behavior: Optional[JSON] = None, + metadata: Optional[List["_models.MetadataItem"]] = None, block_size_in_mb: Optional[JSON] = None, **kwargs: Any ) -> None: @@ -6003,6 +6037,9 @@ def __init__( :paramtype disable_metrics_collection: JSON :keyword copy_behavior: The type of copy behavior for copy sink. :paramtype copy_behavior: JSON + :keyword metadata: Specify the custom metadata to be added to sink data. Type: array of objects + (or Expression with resultType array of objects). + :paramtype metadata: list[~azure.mgmt.datafactory.models.MetadataItem] :keyword block_size_in_mb: Indicates the block size(MB) when writing data to blob. Type: integer (or Expression with resultType integer). :paramtype block_size_in_mb: JSON @@ -6012,6 +6049,7 @@ def __init__( max_concurrent_connections=max_concurrent_connections, disable_metrics_collection=disable_metrics_collection, copy_behavior=copy_behavior, + metadata=metadata, **kwargs ) self.type: str = "AzureBlobFSWriteSettings" @@ -6443,6 +6481,9 @@ class AzureBlobStorageWriteSettings(StoreWriteSettings): :vartype disable_metrics_collection: JSON :ivar copy_behavior: The type of copy behavior for copy sink. :vartype copy_behavior: JSON + :ivar metadata: Specify the custom metadata to be added to sink data. Type: array of objects + (or Expression with resultType array of objects). + :vartype metadata: list[~azure.mgmt.datafactory.models.MetadataItem] :ivar block_size_in_mb: Indicates the block size(MB) when writing data to blob. Type: integer (or Expression with resultType integer). :vartype block_size_in_mb: JSON @@ -6458,6 +6499,7 @@ class AzureBlobStorageWriteSettings(StoreWriteSettings): "max_concurrent_connections": {"key": "maxConcurrentConnections", "type": "object"}, "disable_metrics_collection": {"key": "disableMetricsCollection", "type": "object"}, "copy_behavior": {"key": "copyBehavior", "type": "object"}, + "metadata": {"key": "metadata", "type": "[MetadataItem]"}, "block_size_in_mb": {"key": "blockSizeInMB", "type": "object"}, } @@ -6468,6 +6510,7 @@ def __init__( max_concurrent_connections: Optional[JSON] = None, disable_metrics_collection: Optional[JSON] = None, copy_behavior: Optional[JSON] = None, + metadata: Optional[List["_models.MetadataItem"]] = None, block_size_in_mb: Optional[JSON] = None, **kwargs: Any ) -> None: @@ -6483,6 +6526,9 @@ def __init__( :paramtype disable_metrics_collection: JSON :keyword copy_behavior: The type of copy behavior for copy sink. :paramtype copy_behavior: JSON + :keyword metadata: Specify the custom metadata to be added to sink data. Type: array of objects + (or Expression with resultType array of objects). + :paramtype metadata: list[~azure.mgmt.datafactory.models.MetadataItem] :keyword block_size_in_mb: Indicates the block size(MB) when writing data to blob. Type: integer (or Expression with resultType integer). :paramtype block_size_in_mb: JSON @@ -6492,6 +6538,7 @@ def __init__( max_concurrent_connections=max_concurrent_connections, disable_metrics_collection=disable_metrics_collection, copy_behavior=copy_behavior, + metadata=metadata, **kwargs ) self.type: str = "AzureBlobStorageWriteSettings" @@ -8960,6 +9007,9 @@ class AzureDataLakeStoreWriteSettings(StoreWriteSettings): :vartype disable_metrics_collection: JSON :ivar copy_behavior: The type of copy behavior for copy sink. :vartype copy_behavior: JSON + :ivar metadata: Specify the custom metadata to be added to sink data. Type: array of objects + (or Expression with resultType array of objects). + :vartype metadata: list[~azure.mgmt.datafactory.models.MetadataItem] :ivar expiry_date_time: Specifies the expiry time of the written files. The time is applied to the UTC time zone in the format of "2018-12-01T05:00:00Z". Default value is NULL. Type: string (or Expression with resultType string). @@ -8976,6 +9026,7 @@ class AzureDataLakeStoreWriteSettings(StoreWriteSettings): "max_concurrent_connections": {"key": "maxConcurrentConnections", "type": "object"}, "disable_metrics_collection": {"key": "disableMetricsCollection", "type": "object"}, "copy_behavior": {"key": "copyBehavior", "type": "object"}, + "metadata": {"key": "metadata", "type": "[MetadataItem]"}, "expiry_date_time": {"key": "expiryDateTime", "type": "object"}, } @@ -8986,6 +9037,7 @@ def __init__( max_concurrent_connections: Optional[JSON] = None, disable_metrics_collection: Optional[JSON] = None, copy_behavior: Optional[JSON] = None, + metadata: Optional[List["_models.MetadataItem"]] = None, expiry_date_time: Optional[JSON] = None, **kwargs: Any ) -> None: @@ -9001,6 +9053,9 @@ def __init__( :paramtype disable_metrics_collection: JSON :keyword copy_behavior: The type of copy behavior for copy sink. :paramtype copy_behavior: JSON + :keyword metadata: Specify the custom metadata to be added to sink data. Type: array of objects + (or Expression with resultType array of objects). + :paramtype metadata: list[~azure.mgmt.datafactory.models.MetadataItem] :keyword expiry_date_time: Specifies the expiry time of the written files. The time is applied to the UTC time zone in the format of "2018-12-01T05:00:00Z". Default value is NULL. Type: string (or Expression with resultType string). @@ -9011,6 +9066,7 @@ def __init__( max_concurrent_connections=max_concurrent_connections, disable_metrics_collection=disable_metrics_collection, copy_behavior=copy_behavior, + metadata=metadata, **kwargs ) self.type: str = "AzureDataLakeStoreWriteSettings" @@ -9387,6 +9443,9 @@ class AzureFileStorageWriteSettings(StoreWriteSettings): :vartype disable_metrics_collection: JSON :ivar copy_behavior: The type of copy behavior for copy sink. :vartype copy_behavior: JSON + :ivar metadata: Specify the custom metadata to be added to sink data. Type: array of objects + (or Expression with resultType array of objects). + :vartype metadata: list[~azure.mgmt.datafactory.models.MetadataItem] """ _validation = { @@ -9399,6 +9458,7 @@ class AzureFileStorageWriteSettings(StoreWriteSettings): "max_concurrent_connections": {"key": "maxConcurrentConnections", "type": "object"}, "disable_metrics_collection": {"key": "disableMetricsCollection", "type": "object"}, "copy_behavior": {"key": "copyBehavior", "type": "object"}, + "metadata": {"key": "metadata", "type": "[MetadataItem]"}, } def __init__( @@ -9408,6 +9468,7 @@ def __init__( max_concurrent_connections: Optional[JSON] = None, disable_metrics_collection: Optional[JSON] = None, copy_behavior: Optional[JSON] = None, + metadata: Optional[List["_models.MetadataItem"]] = None, **kwargs: Any ) -> None: """ @@ -9422,12 +9483,16 @@ def __init__( :paramtype disable_metrics_collection: JSON :keyword copy_behavior: The type of copy behavior for copy sink. :paramtype copy_behavior: JSON + :keyword metadata: Specify the custom metadata to be added to sink data. Type: array of objects + (or Expression with resultType array of objects). + :paramtype metadata: list[~azure.mgmt.datafactory.models.MetadataItem] """ super().__init__( additional_properties=additional_properties, max_concurrent_connections=max_concurrent_connections, disable_metrics_collection=disable_metrics_collection, copy_behavior=copy_behavior, + metadata=metadata, **kwargs ) self.type: str = "AzureFileStorageWriteSettings" @@ -9470,7 +9535,7 @@ class AzureFunctionActivity(ExecutionActivity): # pylint: disable=too-many-inst :vartype function_name: JSON :ivar headers: Represents the headers that will be sent to the request. For example, to set the language and type on a request: "headers" : { "Accept-Language": "en-us", "Content-Type": - "application/json" }. Type: string (or Expression with resultType string). + "application/json" }. Type: dictionary (or Expression with resultType dictionary). :vartype headers: JSON :ivar body: Represents the payload that will be sent to the endpoint. Required for POST/PUT method, not allowed for GET method Type: string (or Expression with resultType string). @@ -9550,7 +9615,7 @@ def __init__( :paramtype function_name: JSON :keyword headers: Represents the headers that will be sent to the request. For example, to set the language and type on a request: "headers" : { "Accept-Language": "en-us", "Content-Type": - "application/json" }. Type: string (or Expression with resultType string). + "application/json" }. Type: dictionary (or Expression with resultType dictionary). :paramtype headers: JSON :keyword body: Represents the payload that will be sent to the endpoint. Required for POST/PUT method, not allowed for GET method Type: string (or Expression with resultType string). @@ -9594,7 +9659,8 @@ class AzureFunctionLinkedService(LinkedService): # pylint: disable=too-many-ins :ivar annotations: List of tags that can be used for describing the linked service. :vartype annotations: list[JSON] :ivar function_app_url: The endpoint of the Azure Function App. URL will be in the format - https://:code:``.azurewebsites.net. Required. + https://:code:``.azurewebsites.net. Type: string (or Expression with resultType + string). Required. :vartype function_app_url: JSON :ivar function_key: Function or Host key for Azure Function App. :vartype function_key: ~azure.mgmt.datafactory.models.SecretBase @@ -9603,7 +9669,8 @@ class AzureFunctionLinkedService(LinkedService): # pylint: disable=too-many-ins :vartype encrypted_credential: str :ivar credential: The credential reference containing authentication information. :vartype credential: ~azure.mgmt.datafactory.models.CredentialReference - :ivar resource_id: Allowed token audiences for azure function. + :ivar resource_id: Allowed token audiences for azure function. Type: string (or Expression with + resultType string). :vartype resource_id: JSON :ivar authentication: Type of authentication (Required to specify MSI) used to connect to AzureFunction. Type: string (or Expression with resultType string). @@ -9659,7 +9726,8 @@ def __init__( :keyword annotations: List of tags that can be used for describing the linked service. :paramtype annotations: list[JSON] :keyword function_app_url: The endpoint of the Azure Function App. URL will be in the format - https://:code:``.azurewebsites.net. Required. + https://:code:``.azurewebsites.net. Type: string (or Expression with resultType + string). Required. :paramtype function_app_url: JSON :keyword function_key: Function or Host key for Azure Function App. :paramtype function_key: ~azure.mgmt.datafactory.models.SecretBase @@ -9668,7 +9736,8 @@ def __init__( :paramtype encrypted_credential: str :keyword credential: The credential reference containing authentication information. :paramtype credential: ~azure.mgmt.datafactory.models.CredentialReference - :keyword resource_id: Allowed token audiences for azure function. + :keyword resource_id: Allowed token audiences for azure function. Type: string (or Expression + with resultType string). :paramtype resource_id: JSON :keyword authentication: Type of authentication (Required to specify MSI) used to connect to AzureFunction. Type: string (or Expression with resultType string). @@ -10343,7 +10412,7 @@ class AzureMLExecutePipelineActivity(ExecutionActivity): # pylint: disable=too- :vartype ml_pipeline_parameters: JSON :ivar data_path_assignments: Dictionary used for changing data path assignments without retraining. Values will be passed in the dataPathAssignments property of the published pipeline - execution request. Type: object with key value pairs (or Expression with resultType object). + execution request. Type: object (or Expression with resultType object). :vartype data_path_assignments: JSON :ivar ml_parent_run_id: The parent Azure ML Service pipeline run id. This information will be passed in the ParentRunId property of the published pipeline execution request. Type: string @@ -10446,7 +10515,7 @@ def __init__( :paramtype ml_pipeline_parameters: JSON :keyword data_path_assignments: Dictionary used for changing data path assignments without retraining. Values will be passed in the dataPathAssignments property of the published pipeline - execution request. Type: object with key value pairs (or Expression with resultType object). + execution request. Type: object (or Expression with resultType object). :paramtype data_path_assignments: JSON :keyword ml_parent_run_id: The parent Azure ML Service pipeline run id. This information will be passed in the ParentRunId property of the published pipeline execution request. Type: string @@ -13064,7 +13133,8 @@ class AzureSqlSource(TabularSource): # pylint: disable=too-many-instance-attrib :ivar produce_additional_types: Which additional types to produce. :vartype produce_additional_types: JSON :ivar partition_option: The partition mechanism that will be used for Sql read in parallel. - Possible values include: "None", "PhysicalPartitionsOfTable", "DynamicRange". + Possible values include: "None", "PhysicalPartitionsOfTable", "DynamicRange". Type: string (or + Expression with resultType string). :vartype partition_option: JSON :ivar partition_settings: The settings that will be leveraged for Sql source partitioning. :vartype partition_settings: ~azure.mgmt.datafactory.models.SqlPartitionSettings @@ -13150,7 +13220,8 @@ def __init__( :keyword produce_additional_types: Which additional types to produce. :paramtype produce_additional_types: JSON :keyword partition_option: The partition mechanism that will be used for Sql read in parallel. - Possible values include: "None", "PhysicalPartitionsOfTable", "DynamicRange". + Possible values include: "None", "PhysicalPartitionsOfTable", "DynamicRange". Type: string (or + Expression with resultType string). :paramtype partition_option: JSON :keyword partition_settings: The settings that will be leveraged for Sql source partitioning. :paramtype partition_settings: ~azure.mgmt.datafactory.models.SqlPartitionSettings @@ -15679,9 +15750,9 @@ class CmdkeySetup(CustomSetupBase): :ivar type: The type of custom setup. Required. :vartype type: str - :ivar target_name: The server name of data source access. Required. + :ivar target_name: The server name of data source access. Type: string. Required. :vartype target_name: JSON - :ivar user_name: The user name of data source access. Required. + :ivar user_name: The user name of data source access. Type: string. Required. :vartype user_name: JSON :ivar password: The password of data source access. Required. :vartype password: ~azure.mgmt.datafactory.models.SecretBase @@ -15703,9 +15774,9 @@ class CmdkeySetup(CustomSetupBase): def __init__(self, *, target_name: JSON, user_name: JSON, password: "_models.SecretBase", **kwargs: Any) -> None: """ - :keyword target_name: The server name of data source access. Required. + :keyword target_name: The server name of data source access. Type: string. Required. :paramtype target_name: JSON - :keyword user_name: The user name of data source access. Required. + :keyword user_name: The user name of data source access. Type: string. Required. :paramtype user_name: JSON :keyword password: The password of data source access. Required. :paramtype password: ~azure.mgmt.datafactory.models.SecretBase @@ -27071,6 +27142,9 @@ class FileServerWriteSettings(StoreWriteSettings): :vartype disable_metrics_collection: JSON :ivar copy_behavior: The type of copy behavior for copy sink. :vartype copy_behavior: JSON + :ivar metadata: Specify the custom metadata to be added to sink data. Type: array of objects + (or Expression with resultType array of objects). + :vartype metadata: list[~azure.mgmt.datafactory.models.MetadataItem] """ _validation = { @@ -27083,6 +27157,7 @@ class FileServerWriteSettings(StoreWriteSettings): "max_concurrent_connections": {"key": "maxConcurrentConnections", "type": "object"}, "disable_metrics_collection": {"key": "disableMetricsCollection", "type": "object"}, "copy_behavior": {"key": "copyBehavior", "type": "object"}, + "metadata": {"key": "metadata", "type": "[MetadataItem]"}, } def __init__( @@ -27092,6 +27167,7 @@ def __init__( max_concurrent_connections: Optional[JSON] = None, disable_metrics_collection: Optional[JSON] = None, copy_behavior: Optional[JSON] = None, + metadata: Optional[List["_models.MetadataItem"]] = None, **kwargs: Any ) -> None: """ @@ -27106,12 +27182,16 @@ def __init__( :paramtype disable_metrics_collection: JSON :keyword copy_behavior: The type of copy behavior for copy sink. :paramtype copy_behavior: JSON + :keyword metadata: Specify the custom metadata to be added to sink data. Type: array of objects + (or Expression with resultType array of objects). + :paramtype metadata: list[~azure.mgmt.datafactory.models.MetadataItem] """ super().__init__( additional_properties=additional_properties, max_concurrent_connections=max_concurrent_connections, disable_metrics_collection=disable_metrics_collection, copy_behavior=copy_behavior, + metadata=metadata, **kwargs ) self.type: str = "FileServerWriteSettings" @@ -31334,8 +31414,8 @@ class HDInsightOnDemandLinkedService(LinkedService): # pylint: disable=too-many :vartype parameters: dict[str, ~azure.mgmt.datafactory.models.ParameterSpecification] :ivar annotations: List of tags that can be used for describing the linked service. :vartype annotations: list[JSON] - :ivar cluster_size: Number of worker/data nodes in the cluster. Suggestion value: 4. Type: - string (or Expression with resultType string). Required. + :ivar cluster_size: Number of worker/data nodes in the cluster. Suggestion value: 4. Type: int + (or Expression with resultType int). Required. :vartype cluster_size: JSON :ivar time_to_live: The allowed idle time for the on-demand HDInsight cluster. Specifies how long the on-demand HDInsight cluster stays alive after completion of an activity run if there @@ -31552,7 +31632,7 @@ def __init__( # pylint: disable=too-many-locals :keyword annotations: List of tags that can be used for describing the linked service. :paramtype annotations: list[JSON] :keyword cluster_size: Number of worker/data nodes in the cluster. Suggestion value: 4. Type: - string (or Expression with resultType string). Required. + int (or Expression with resultType int). Required. :paramtype cluster_size: JSON :keyword time_to_live: The allowed idle time for the on-demand HDInsight cluster. Specifies how long the on-demand HDInsight cluster stays alive after completion of an activity run if there @@ -32836,7 +32916,7 @@ class HttpLinkedService(LinkedService): # pylint: disable=too-many-instance-att authentication. :vartype password: ~azure.mgmt.datafactory.models.SecretBase :ivar auth_headers: The additional HTTP headers in the request to RESTful API used for - authorization. Type: object (or Expression with resultType object). + authorization. Type: key value pairs (value should be string type). :vartype auth_headers: JSON :ivar embedded_cert_data: Base64 encoded certificate data for ClientCertificate authentication. For on-premises copy with ClientCertificate authentication, either CertThumbprint or @@ -32926,7 +33006,7 @@ def __init__( EmbeddedCertData authentication. :paramtype password: ~azure.mgmt.datafactory.models.SecretBase :keyword auth_headers: The additional HTTP headers in the request to RESTful API used for - authorization. Type: object (or Expression with resultType object). + authorization. Type: key value pairs (value should be string type). :paramtype auth_headers: JSON :keyword embedded_cert_data: Base64 encoded certificate data for ClientCertificate authentication. For on-premises copy with ClientCertificate authentication, either @@ -37272,6 +37352,9 @@ class LakeHouseWriteSettings(StoreWriteSettings): :vartype disable_metrics_collection: JSON :ivar copy_behavior: The type of copy behavior for copy sink. :vartype copy_behavior: JSON + :ivar metadata: Specify the custom metadata to be added to sink data. Type: array of objects + (or Expression with resultType array of objects). + :vartype metadata: list[~azure.mgmt.datafactory.models.MetadataItem] """ _validation = { @@ -37284,6 +37367,7 @@ class LakeHouseWriteSettings(StoreWriteSettings): "max_concurrent_connections": {"key": "maxConcurrentConnections", "type": "object"}, "disable_metrics_collection": {"key": "disableMetricsCollection", "type": "object"}, "copy_behavior": {"key": "copyBehavior", "type": "object"}, + "metadata": {"key": "metadata", "type": "[MetadataItem]"}, } def __init__( @@ -37293,6 +37377,7 @@ def __init__( max_concurrent_connections: Optional[JSON] = None, disable_metrics_collection: Optional[JSON] = None, copy_behavior: Optional[JSON] = None, + metadata: Optional[List["_models.MetadataItem"]] = None, **kwargs: Any ) -> None: """ @@ -37307,12 +37392,16 @@ def __init__( :paramtype disable_metrics_collection: JSON :keyword copy_behavior: The type of copy behavior for copy sink. :paramtype copy_behavior: JSON + :keyword metadata: Specify the custom metadata to be added to sink data. Type: array of objects + (or Expression with resultType array of objects). + :paramtype metadata: list[~azure.mgmt.datafactory.models.MetadataItem] """ super().__init__( additional_properties=additional_properties, max_concurrent_connections=max_concurrent_connections, disable_metrics_collection=disable_metrics_collection, copy_behavior=copy_behavior, + metadata=metadata, **kwargs ) self.type: str = "LakeHouseWriteSettings" @@ -39516,7 +39605,7 @@ def __init__( self.script_lines = script_lines -class MariaDBLinkedService(LinkedService): +class MariaDBLinkedService(LinkedService): # pylint: disable=too-many-instance-attributes """MariaDB server linked service. All required parameters must be populated in order to send to Azure. @@ -39534,11 +39623,23 @@ class MariaDBLinkedService(LinkedService): :vartype parameters: dict[str, ~azure.mgmt.datafactory.models.ParameterSpecification] :ivar annotations: List of tags that can be used for describing the linked service. :vartype annotations: list[JSON] + :ivar driver_version: The version of the MariaDB driver. Type: string. V1 or empty for legacy + driver, V2 for new driver. V1 can support connection string and property bag, V2 can only + support connection string. + :vartype driver_version: JSON :ivar connection_string: An ODBC connection string. Type: string, SecureString or AzureKeyVaultSecretReference. :vartype connection_string: JSON - :ivar pwd: The Azure key vault secret reference of password in connection string. - :vartype pwd: ~azure.mgmt.datafactory.models.AzureKeyVaultSecretReference + :ivar server: Server name for connection. Type: string. + :vartype server: JSON + :ivar port: The port for the connection. Type: integer. + :vartype port: JSON + :ivar username: Username for authentication. Type: string. + :vartype username: JSON + :ivar database: Database name for connection. Type: string. + :vartype database: JSON + :ivar password: The Azure key vault secret reference of password in connection string. + :vartype password: ~azure.mgmt.datafactory.models.AzureKeyVaultSecretReference :ivar encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string. :vartype encrypted_credential: str @@ -39555,8 +39656,13 @@ class MariaDBLinkedService(LinkedService): "description": {"key": "description", "type": "str"}, "parameters": {"key": "parameters", "type": "{ParameterSpecification}"}, "annotations": {"key": "annotations", "type": "[object]"}, + "driver_version": {"key": "typeProperties.driverVersion", "type": "object"}, "connection_string": {"key": "typeProperties.connectionString", "type": "object"}, - "pwd": {"key": "typeProperties.pwd", "type": "AzureKeyVaultSecretReference"}, + "server": {"key": "typeProperties.server", "type": "object"}, + "port": {"key": "typeProperties.port", "type": "object"}, + "username": {"key": "typeProperties.username", "type": "object"}, + "database": {"key": "typeProperties.database", "type": "object"}, + "password": {"key": "typeProperties.password", "type": "AzureKeyVaultSecretReference"}, "encrypted_credential": {"key": "typeProperties.encryptedCredential", "type": "str"}, } @@ -39568,8 +39674,13 @@ def __init__( description: Optional[str] = None, parameters: Optional[Dict[str, "_models.ParameterSpecification"]] = None, annotations: Optional[List[JSON]] = None, + driver_version: Optional[JSON] = None, connection_string: Optional[JSON] = None, - pwd: Optional["_models.AzureKeyVaultSecretReference"] = None, + server: Optional[JSON] = None, + port: Optional[JSON] = None, + username: Optional[JSON] = None, + database: Optional[JSON] = None, + password: Optional["_models.AzureKeyVaultSecretReference"] = None, encrypted_credential: Optional[str] = None, **kwargs: Any ) -> None: @@ -39585,11 +39696,23 @@ def __init__( :paramtype parameters: dict[str, ~azure.mgmt.datafactory.models.ParameterSpecification] :keyword annotations: List of tags that can be used for describing the linked service. :paramtype annotations: list[JSON] + :keyword driver_version: The version of the MariaDB driver. Type: string. V1 or empty for + legacy driver, V2 for new driver. V1 can support connection string and property bag, V2 can + only support connection string. + :paramtype driver_version: JSON :keyword connection_string: An ODBC connection string. Type: string, SecureString or AzureKeyVaultSecretReference. :paramtype connection_string: JSON - :keyword pwd: The Azure key vault secret reference of password in connection string. - :paramtype pwd: ~azure.mgmt.datafactory.models.AzureKeyVaultSecretReference + :keyword server: Server name for connection. Type: string. + :paramtype server: JSON + :keyword port: The port for the connection. Type: integer. + :paramtype port: JSON + :keyword username: Username for authentication. Type: string. + :paramtype username: JSON + :keyword database: Database name for connection. Type: string. + :paramtype database: JSON + :keyword password: The Azure key vault secret reference of password in connection string. + :paramtype password: ~azure.mgmt.datafactory.models.AzureKeyVaultSecretReference :keyword encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string. :paramtype encrypted_credential: str @@ -39603,8 +39726,13 @@ def __init__( **kwargs ) self.type: str = "MariaDB" + self.driver_version = driver_version self.connection_string = connection_string - self.pwd = pwd + self.server = server + self.port = port + self.username = username + self.database = database + self.password = password self.encrypted_credential = encrypted_credential @@ -41871,7 +41999,7 @@ def __init__( self.additional_columns = additional_columns -class MySqlLinkedService(LinkedService): +class MySqlLinkedService(LinkedService): # pylint: disable=too-many-instance-attributes """Linked service for MySQL data source. All required parameters must be populated in order to send to Azure. @@ -41889,9 +42017,27 @@ class MySqlLinkedService(LinkedService): :vartype parameters: dict[str, ~azure.mgmt.datafactory.models.ParameterSpecification] :ivar annotations: List of tags that can be used for describing the linked service. :vartype annotations: list[JSON] + :ivar driver_version: The version of the MySQL driver. Type: string. V1 or empty for legacy + driver, V2 for new driver. V1 can support connection string and property bag, V2 can only + support connection string. + :vartype driver_version: JSON :ivar connection_string: The connection string. Type: string, SecureString or - AzureKeyVaultSecretReference. Required. + AzureKeyVaultSecretReference. :vartype connection_string: JSON + :ivar server: Server name for connection. Type: string. + :vartype server: JSON + :ivar port: The port for the connection. Type: integer. + :vartype port: JSON + :ivar username: Username for authentication. Type: string. + :vartype username: JSON + :ivar database: Database name for connection. Type: string. + :vartype database: JSON + :ivar ssl_mode: SSL mode for connection. Type: integer. 0: disable, 1: prefer, 2: require, 3: + verify-ca, 4: verify-full. + :vartype ssl_mode: JSON + :ivar use_system_trust_store: Use system trust store for connection. Type: integer. 0: enable, + 1: disable. + :vartype use_system_trust_store: JSON :ivar password: The Azure key vault secret reference of password in connection string. :vartype password: ~azure.mgmt.datafactory.models.AzureKeyVaultSecretReference :ivar encrypted_credential: The encrypted credential used for authentication. Credentials are @@ -41901,7 +42047,6 @@ class MySqlLinkedService(LinkedService): _validation = { "type": {"required": True}, - "connection_string": {"required": True}, } _attribute_map = { @@ -41911,7 +42056,14 @@ class MySqlLinkedService(LinkedService): "description": {"key": "description", "type": "str"}, "parameters": {"key": "parameters", "type": "{ParameterSpecification}"}, "annotations": {"key": "annotations", "type": "[object]"}, + "driver_version": {"key": "typeProperties.driverVersion", "type": "object"}, "connection_string": {"key": "typeProperties.connectionString", "type": "object"}, + "server": {"key": "typeProperties.server", "type": "object"}, + "port": {"key": "typeProperties.port", "type": "object"}, + "username": {"key": "typeProperties.username", "type": "object"}, + "database": {"key": "typeProperties.database", "type": "object"}, + "ssl_mode": {"key": "typeProperties.sslMode", "type": "object"}, + "use_system_trust_store": {"key": "typeProperties.useSystemTrustStore", "type": "object"}, "password": {"key": "typeProperties.password", "type": "AzureKeyVaultSecretReference"}, "encrypted_credential": {"key": "typeProperties.encryptedCredential", "type": "str"}, } @@ -41919,12 +42071,19 @@ class MySqlLinkedService(LinkedService): def __init__( self, *, - connection_string: JSON, additional_properties: Optional[Dict[str, JSON]] = None, connect_via: Optional["_models.IntegrationRuntimeReference"] = None, description: Optional[str] = None, parameters: Optional[Dict[str, "_models.ParameterSpecification"]] = None, annotations: Optional[List[JSON]] = None, + driver_version: Optional[JSON] = None, + connection_string: Optional[JSON] = None, + server: Optional[JSON] = None, + port: Optional[JSON] = None, + username: Optional[JSON] = None, + database: Optional[JSON] = None, + ssl_mode: Optional[JSON] = None, + use_system_trust_store: Optional[JSON] = None, password: Optional["_models.AzureKeyVaultSecretReference"] = None, encrypted_credential: Optional[str] = None, **kwargs: Any @@ -41941,9 +42100,27 @@ def __init__( :paramtype parameters: dict[str, ~azure.mgmt.datafactory.models.ParameterSpecification] :keyword annotations: List of tags that can be used for describing the linked service. :paramtype annotations: list[JSON] + :keyword driver_version: The version of the MySQL driver. Type: string. V1 or empty for legacy + driver, V2 for new driver. V1 can support connection string and property bag, V2 can only + support connection string. + :paramtype driver_version: JSON :keyword connection_string: The connection string. Type: string, SecureString or - AzureKeyVaultSecretReference. Required. + AzureKeyVaultSecretReference. :paramtype connection_string: JSON + :keyword server: Server name for connection. Type: string. + :paramtype server: JSON + :keyword port: The port for the connection. Type: integer. + :paramtype port: JSON + :keyword username: Username for authentication. Type: string. + :paramtype username: JSON + :keyword database: Database name for connection. Type: string. + :paramtype database: JSON + :keyword ssl_mode: SSL mode for connection. Type: integer. 0: disable, 1: prefer, 2: require, + 3: verify-ca, 4: verify-full. + :paramtype ssl_mode: JSON + :keyword use_system_trust_store: Use system trust store for connection. Type: integer. 0: + enable, 1: disable. + :paramtype use_system_trust_store: JSON :keyword password: The Azure key vault secret reference of password in connection string. :paramtype password: ~azure.mgmt.datafactory.models.AzureKeyVaultSecretReference :keyword encrypted_credential: The encrypted credential used for authentication. Credentials @@ -41959,7 +42136,14 @@ def __init__( **kwargs ) self.type: str = "MySql" + self.driver_version = driver_version self.connection_string = connection_string + self.server = server + self.port = port + self.username = username + self.database = database + self.ssl_mode = ssl_mode + self.use_system_trust_store = use_system_trust_store self.password = password self.encrypted_credential = encrypted_credential @@ -42611,7 +42795,7 @@ class ODataLinkedService(LinkedService): # pylint: disable=too-many-instance-at :ivar password: Password of the OData service. :vartype password: ~azure.mgmt.datafactory.models.SecretBase :ivar auth_headers: The additional HTTP headers in the request to RESTful API used for - authorization. Type: object (or Expression with resultType object). + authorization. Type: key value pairs (value should be string type). :vartype auth_headers: JSON :ivar tenant: Specify the tenant information (domain name or tenant ID) under which your application resides. Type: string (or Expression with resultType string). @@ -42731,7 +42915,7 @@ def __init__( :keyword password: Password of the OData service. :paramtype password: ~azure.mgmt.datafactory.models.SecretBase :keyword auth_headers: The additional HTTP headers in the request to RESTful API used for - authorization. Type: object (or Expression with resultType object). + authorization. Type: key value pairs (value should be string type). :paramtype auth_headers: JSON :keyword tenant: Specify the tenant information (domain name or tenant ID) under which your application resides. Type: string (or Expression with resultType string). @@ -50553,7 +50737,7 @@ class RestSink(CopySink): # pylint: disable=too-many-instance-attributes string (or Expression with resultType string). :vartype request_method: JSON :ivar additional_headers: The additional HTTP headers in the request to the RESTful API. Type: - string (or Expression with resultType string). + key value pairs (value should be string type). :vartype additional_headers: JSON :ivar http_request_timeout: The timeout (TimeSpan) to get an HTTP response. It is the timeout to get a response, not the timeout to read response data. Default value: 00:01:40. Type: string @@ -50563,7 +50747,8 @@ class RestSink(CopySink): # pylint: disable=too-many-instance-attributes :ivar request_interval: The time to await before sending next request, in milliseconds. :vartype request_interval: JSON :ivar http_compression_type: Http Compression Type to Send data in compressed format with - Optimal Compression Level, Default is None. And The Only Supported option is Gzip. + Optimal Compression Level, Default is None. And The Only Supported option is Gzip. Type: string + (or Expression with resultType string). :vartype http_compression_type: JSON """ @@ -50630,7 +50815,7 @@ def __init__( Type: string (or Expression with resultType string). :paramtype request_method: JSON :keyword additional_headers: The additional HTTP headers in the request to the RESTful API. - Type: string (or Expression with resultType string). + Type: key value pairs (value should be string type). :paramtype additional_headers: JSON :keyword http_request_timeout: The timeout (TimeSpan) to get an HTTP response. It is the timeout to get a response, not the timeout to read response data. Default value: 00:01:40. @@ -50640,7 +50825,8 @@ def __init__( :keyword request_interval: The time to await before sending next request, in milliseconds. :paramtype request_interval: JSON :keyword http_compression_type: Http Compression Type to Send data in compressed format with - Optimal Compression Level, Default is None. And The Only Supported option is Gzip. + Optimal Compression Level, Default is None. And The Only Supported option is Gzip. Type: string + (or Expression with resultType string). :paramtype http_compression_type: JSON """ super().__init__( @@ -50703,7 +50889,7 @@ class RestSource(CopySource): # pylint: disable=too-many-instance-attributes :ivar request_interval: The time to await before sending next page request. :vartype request_interval: JSON :ivar additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). + key value pairs (value should be string type). :vartype additional_columns: JSON """ @@ -50780,7 +50966,7 @@ def __init__( :keyword request_interval: The time to await before sending next page request. :paramtype request_interval: JSON :keyword additional_columns: Specifies the additional columns to be added to source data. Type: - array of objects(AdditionalColumns) (or Expression with resultType array of objects). + key value pairs (value should be string type). :paramtype additional_columns: JSON """ super().__init__( @@ -52035,8 +52221,244 @@ def __init__( self.additional_columns = additional_columns -class SalesforceSink(CopySink): # pylint: disable=too-many-instance-attributes - """A copy activity Salesforce sink. +class SalesforceServiceCloudV2LinkedService(LinkedService): # pylint: disable=too-many-instance-attributes + """Linked service for Salesforce Service Cloud V2. + + All required parameters must be populated in order to send to Azure. + + :ivar additional_properties: Unmatched properties from the message are deserialized to this + collection. + :vartype additional_properties: dict[str, JSON] + :ivar type: Type of linked service. Required. + :vartype type: str + :ivar connect_via: The integration runtime reference. + :vartype connect_via: ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :ivar description: Linked service description. + :vartype description: str + :ivar parameters: Parameters for linked service. + :vartype parameters: dict[str, ~azure.mgmt.datafactory.models.ParameterSpecification] + :ivar annotations: List of tags that can be used for describing the linked service. + :vartype annotations: list[JSON] + :ivar environment_url: The URL of Salesforce Service Cloud instance. For example, + 'https://[domain].my.salesforce.com'. Type: string (or Expression with resultType string). + :vartype environment_url: JSON + :ivar authentication_type: The authentication type to be used to connect to the Salesforce. + Currently, we only support OAuth2ClientCredentials, it is also the default value. + :vartype authentication_type: JSON + :ivar client_id: The client Id for OAuth 2.0 Client Credentials Flow authentication of the + Salesforce instance. Type: string (or Expression with resultType string). + :vartype client_id: JSON + :ivar client_secret: The client secret for OAuth 2.0 Client Credentials Flow authentication of + the Salesforce instance. + :vartype client_secret: ~azure.mgmt.datafactory.models.SecretBase + :ivar api_version: The Salesforce API version used in ADF. The version must be larger than or + equal to 47.0 which is required by Salesforce BULK API 2.0. Type: string (or Expression with + resultType string). + :vartype api_version: JSON + :ivar encrypted_credential: The encrypted credential used for authentication. Credentials are + encrypted using the integration runtime credential manager. Type: string. + :vartype encrypted_credential: str + """ + + _validation = { + "type": {"required": True}, + } + + _attribute_map = { + "additional_properties": {"key": "", "type": "{object}"}, + "type": {"key": "type", "type": "str"}, + "connect_via": {"key": "connectVia", "type": "IntegrationRuntimeReference"}, + "description": {"key": "description", "type": "str"}, + "parameters": {"key": "parameters", "type": "{ParameterSpecification}"}, + "annotations": {"key": "annotations", "type": "[object]"}, + "environment_url": {"key": "typeProperties.environmentUrl", "type": "object"}, + "authentication_type": {"key": "typeProperties.authenticationType", "type": "object"}, + "client_id": {"key": "typeProperties.clientId", "type": "object"}, + "client_secret": {"key": "typeProperties.clientSecret", "type": "SecretBase"}, + "api_version": {"key": "typeProperties.apiVersion", "type": "object"}, + "encrypted_credential": {"key": "typeProperties.encryptedCredential", "type": "str"}, + } + + def __init__( + self, + *, + additional_properties: Optional[Dict[str, JSON]] = None, + connect_via: Optional["_models.IntegrationRuntimeReference"] = None, + description: Optional[str] = None, + parameters: Optional[Dict[str, "_models.ParameterSpecification"]] = None, + annotations: Optional[List[JSON]] = None, + environment_url: Optional[JSON] = None, + authentication_type: Optional[JSON] = None, + client_id: Optional[JSON] = None, + client_secret: Optional["_models.SecretBase"] = None, + api_version: Optional[JSON] = None, + encrypted_credential: Optional[str] = None, + **kwargs: Any + ) -> None: + """ + :keyword additional_properties: Unmatched properties from the message are deserialized to this + collection. + :paramtype additional_properties: dict[str, JSON] + :keyword connect_via: The integration runtime reference. + :paramtype connect_via: ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :keyword description: Linked service description. + :paramtype description: str + :keyword parameters: Parameters for linked service. + :paramtype parameters: dict[str, ~azure.mgmt.datafactory.models.ParameterSpecification] + :keyword annotations: List of tags that can be used for describing the linked service. + :paramtype annotations: list[JSON] + :keyword environment_url: The URL of Salesforce Service Cloud instance. For example, + 'https://[domain].my.salesforce.com'. Type: string (or Expression with resultType string). + :paramtype environment_url: JSON + :keyword authentication_type: The authentication type to be used to connect to the Salesforce. + Currently, we only support OAuth2ClientCredentials, it is also the default value. + :paramtype authentication_type: JSON + :keyword client_id: The client Id for OAuth 2.0 Client Credentials Flow authentication of the + Salesforce instance. Type: string (or Expression with resultType string). + :paramtype client_id: JSON + :keyword client_secret: The client secret for OAuth 2.0 Client Credentials Flow authentication + of the Salesforce instance. + :paramtype client_secret: ~azure.mgmt.datafactory.models.SecretBase + :keyword api_version: The Salesforce API version used in ADF. The version must be larger than + or equal to 47.0 which is required by Salesforce BULK API 2.0. Type: string (or Expression with + resultType string). + :paramtype api_version: JSON + :keyword encrypted_credential: The encrypted credential used for authentication. Credentials + are encrypted using the integration runtime credential manager. Type: string. + :paramtype encrypted_credential: str + """ + super().__init__( + additional_properties=additional_properties, + connect_via=connect_via, + description=description, + parameters=parameters, + annotations=annotations, + **kwargs + ) + self.type: str = "SalesforceServiceCloudV2" + self.environment_url = environment_url + self.authentication_type = authentication_type + self.client_id = client_id + self.client_secret = client_secret + self.api_version = api_version + self.encrypted_credential = encrypted_credential + + +class SalesforceServiceCloudV2ObjectDataset(Dataset): # pylint: disable=too-many-instance-attributes + """The Salesforce Service Cloud V2 object dataset. + + All required parameters must be populated in order to send to Azure. + + :ivar additional_properties: Unmatched properties from the message are deserialized to this + collection. + :vartype additional_properties: dict[str, JSON] + :ivar type: Type of dataset. Required. + :vartype type: str + :ivar description: Dataset description. + :vartype description: str + :ivar structure: Columns that define the structure of the dataset. Type: array (or Expression + with resultType array), itemType: DatasetDataElement. + :vartype structure: JSON + :ivar schema: Columns that define the physical type schema of the dataset. Type: array (or + Expression with resultType array), itemType: DatasetSchemaDataElement. + :vartype schema: JSON + :ivar linked_service_name: Linked service reference. Required. + :vartype linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference + :ivar parameters: Parameters for dataset. + :vartype parameters: dict[str, ~azure.mgmt.datafactory.models.ParameterSpecification] + :ivar annotations: List of tags that can be used for describing the Dataset. + :vartype annotations: list[JSON] + :ivar folder: The folder that this Dataset is in. If not specified, Dataset will appear at the + root level. + :vartype folder: ~azure.mgmt.datafactory.models.DatasetFolder + :ivar object_api_name: The Salesforce Service Cloud V2 object API name. Type: string (or + Expression with resultType string). + :vartype object_api_name: JSON + :ivar report_id: The Salesforce Service Cloud V2 reportId. Type: string (or Expression with + resultType string). + :vartype report_id: JSON + """ + + _validation = { + "type": {"required": True}, + "linked_service_name": {"required": True}, + } + + _attribute_map = { + "additional_properties": {"key": "", "type": "{object}"}, + "type": {"key": "type", "type": "str"}, + "description": {"key": "description", "type": "str"}, + "structure": {"key": "structure", "type": "object"}, + "schema": {"key": "schema", "type": "object"}, + "linked_service_name": {"key": "linkedServiceName", "type": "LinkedServiceReference"}, + "parameters": {"key": "parameters", "type": "{ParameterSpecification}"}, + "annotations": {"key": "annotations", "type": "[object]"}, + "folder": {"key": "folder", "type": "DatasetFolder"}, + "object_api_name": {"key": "typeProperties.objectApiName", "type": "object"}, + "report_id": {"key": "typeProperties.reportId", "type": "object"}, + } + + def __init__( + self, + *, + linked_service_name: "_models.LinkedServiceReference", + additional_properties: Optional[Dict[str, JSON]] = None, + description: Optional[str] = None, + structure: Optional[JSON] = None, + schema: Optional[JSON] = None, + parameters: Optional[Dict[str, "_models.ParameterSpecification"]] = None, + annotations: Optional[List[JSON]] = None, + folder: Optional["_models.DatasetFolder"] = None, + object_api_name: Optional[JSON] = None, + report_id: Optional[JSON] = None, + **kwargs: Any + ) -> None: + """ + :keyword additional_properties: Unmatched properties from the message are deserialized to this + collection. + :paramtype additional_properties: dict[str, JSON] + :keyword description: Dataset description. + :paramtype description: str + :keyword structure: Columns that define the structure of the dataset. Type: array (or + Expression with resultType array), itemType: DatasetDataElement. + :paramtype structure: JSON + :keyword schema: Columns that define the physical type schema of the dataset. Type: array (or + Expression with resultType array), itemType: DatasetSchemaDataElement. + :paramtype schema: JSON + :keyword linked_service_name: Linked service reference. Required. + :paramtype linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference + :keyword parameters: Parameters for dataset. + :paramtype parameters: dict[str, ~azure.mgmt.datafactory.models.ParameterSpecification] + :keyword annotations: List of tags that can be used for describing the Dataset. + :paramtype annotations: list[JSON] + :keyword folder: The folder that this Dataset is in. If not specified, Dataset will appear at + the root level. + :paramtype folder: ~azure.mgmt.datafactory.models.DatasetFolder + :keyword object_api_name: The Salesforce Service Cloud V2 object API name. Type: string (or + Expression with resultType string). + :paramtype object_api_name: JSON + :keyword report_id: The Salesforce Service Cloud V2 reportId. Type: string (or Expression with + resultType string). + :paramtype report_id: JSON + """ + super().__init__( + additional_properties=additional_properties, + description=description, + structure=structure, + schema=schema, + linked_service_name=linked_service_name, + parameters=parameters, + annotations=annotations, + folder=folder, + **kwargs + ) + self.type: str = "SalesforceServiceCloudV2Object" + self.object_api_name = object_api_name + self.report_id = report_id + + +class SalesforceServiceCloudV2Sink(CopySink): # pylint: disable=too-many-instance-attributes + """A copy activity Salesforce Service Cloud V2 sink. All required parameters must be populated in order to send to Azure. @@ -52065,7 +52487,7 @@ class SalesforceSink(CopySink): # pylint: disable=too-many-instance-attributes :vartype disable_metrics_collection: JSON :ivar write_behavior: The write behavior for the operation. Default is Insert. Known values are: "Insert" and "Upsert". - :vartype write_behavior: str or ~azure.mgmt.datafactory.models.SalesforceSinkWriteBehavior + :vartype write_behavior: str or ~azure.mgmt.datafactory.models.SalesforceV2SinkWriteBehavior :ivar external_id_field_name: The name of the external ID field for upsert operation. Default value is 'Id' column. Type: string (or Expression with resultType string). :vartype external_id_field_name: JSON @@ -52106,7 +52528,7 @@ def __init__( sink_retry_wait: Optional[JSON] = None, max_concurrent_connections: Optional[JSON] = None, disable_metrics_collection: Optional[JSON] = None, - write_behavior: Optional[Union[str, "_models.SalesforceSinkWriteBehavior"]] = None, + write_behavior: Optional[Union[str, "_models.SalesforceV2SinkWriteBehavior"]] = None, external_id_field_name: Optional[JSON] = None, ignore_null_values: Optional[JSON] = None, **kwargs: Any @@ -52135,7 +52557,7 @@ def __init__( :paramtype disable_metrics_collection: JSON :keyword write_behavior: The write behavior for the operation. Default is Insert. Known values are: "Insert" and "Upsert". - :paramtype write_behavior: str or ~azure.mgmt.datafactory.models.SalesforceSinkWriteBehavior + :paramtype write_behavior: str or ~azure.mgmt.datafactory.models.SalesforceV2SinkWriteBehavior :keyword external_id_field_name: The name of the external ID field for upsert operation. Default value is 'Id' column. Type: string (or Expression with resultType string). :paramtype external_id_field_name: JSON @@ -52158,14 +52580,14 @@ def __init__( disable_metrics_collection=disable_metrics_collection, **kwargs ) - self.type: str = "SalesforceSink" + self.type: str = "SalesforceServiceCloudV2Sink" self.write_behavior = write_behavior self.external_id_field_name = external_id_field_name self.ignore_null_values = ignore_null_values -class SalesforceSource(TabularSource): - """A copy activity Salesforce source. +class SalesforceServiceCloudV2Source(CopySource): + """A copy activity Salesforce Service Cloud V2 source. All required parameters must be populated in order to send to Azure. @@ -52186,17 +52608,14 @@ class SalesforceSource(TabularSource): :ivar disable_metrics_collection: If true, disable data store metrics collection. Default is false. Type: boolean (or Expression with resultType boolean). :vartype disable_metrics_collection: JSON - :ivar query_timeout: Query timeout. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :vartype query_timeout: JSON + :ivar soql_query: Database query. Type: string (or Expression with resultType string). + :vartype soql_query: JSON + :ivar include_deleted_objects: This property control whether query result contains Deleted + objects. Default is false. Type: boolean (or Expression with resultType boolean). + :vartype include_deleted_objects: JSON :ivar additional_columns: Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). :vartype additional_columns: JSON - :ivar query: Database query. Type: string (or Expression with resultType string). - :vartype query: JSON - :ivar read_behavior: The read behavior for the operation. Default is Query. Allowed values: - Query/QueryAll. Type: string (or Expression with resultType string). - :vartype read_behavior: JSON """ _validation = { @@ -52210,10 +52629,9 @@ class SalesforceSource(TabularSource): "source_retry_wait": {"key": "sourceRetryWait", "type": "object"}, "max_concurrent_connections": {"key": "maxConcurrentConnections", "type": "object"}, "disable_metrics_collection": {"key": "disableMetricsCollection", "type": "object"}, - "query_timeout": {"key": "queryTimeout", "type": "object"}, + "soql_query": {"key": "SOQLQuery", "type": "object"}, + "include_deleted_objects": {"key": "includeDeletedObjects", "type": "object"}, "additional_columns": {"key": "additionalColumns", "type": "object"}, - "query": {"key": "query", "type": "object"}, - "read_behavior": {"key": "readBehavior", "type": "object"}, } def __init__( @@ -52224,10 +52642,9 @@ def __init__( source_retry_wait: Optional[JSON] = None, max_concurrent_connections: Optional[JSON] = None, disable_metrics_collection: Optional[JSON] = None, - query_timeout: Optional[JSON] = None, + soql_query: Optional[JSON] = None, + include_deleted_objects: Optional[JSON] = None, additional_columns: Optional[JSON] = None, - query: Optional[JSON] = None, - read_behavior: Optional[JSON] = None, **kwargs: Any ) -> None: """ @@ -52246,17 +52663,14 @@ def __init__( :keyword disable_metrics_collection: If true, disable data store metrics collection. Default is false. Type: boolean (or Expression with resultType boolean). :paramtype disable_metrics_collection: JSON - :keyword query_timeout: Query timeout. Type: string (or Expression with resultType string), - pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :paramtype query_timeout: JSON + :keyword soql_query: Database query. Type: string (or Expression with resultType string). + :paramtype soql_query: JSON + :keyword include_deleted_objects: This property control whether query result contains Deleted + objects. Default is false. Type: boolean (or Expression with resultType boolean). + :paramtype include_deleted_objects: JSON :keyword additional_columns: Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). :paramtype additional_columns: JSON - :keyword query: Database query. Type: string (or Expression with resultType string). - :paramtype query: JSON - :keyword read_behavior: The read behavior for the operation. Default is Query. Allowed values: - Query/QueryAll. Type: string (or Expression with resultType string). - :paramtype read_behavior: JSON """ super().__init__( additional_properties=additional_properties, @@ -52264,234 +52678,145 @@ def __init__( source_retry_wait=source_retry_wait, max_concurrent_connections=max_concurrent_connections, disable_metrics_collection=disable_metrics_collection, - query_timeout=query_timeout, - additional_columns=additional_columns, - **kwargs - ) - self.type: str = "SalesforceSource" - self.query = query - self.read_behavior = read_behavior - - -class SapBwCubeDataset(Dataset): - """The SAP BW cube dataset. - - All required parameters must be populated in order to send to Azure. - - :ivar additional_properties: Unmatched properties from the message are deserialized to this - collection. - :vartype additional_properties: dict[str, JSON] - :ivar type: Type of dataset. Required. - :vartype type: str - :ivar description: Dataset description. - :vartype description: str - :ivar structure: Columns that define the structure of the dataset. Type: array (or Expression - with resultType array), itemType: DatasetDataElement. - :vartype structure: JSON - :ivar schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :vartype schema: JSON - :ivar linked_service_name: Linked service reference. Required. - :vartype linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference - :ivar parameters: Parameters for dataset. - :vartype parameters: dict[str, ~azure.mgmt.datafactory.models.ParameterSpecification] - :ivar annotations: List of tags that can be used for describing the Dataset. - :vartype annotations: list[JSON] - :ivar folder: The folder that this Dataset is in. If not specified, Dataset will appear at the - root level. - :vartype folder: ~azure.mgmt.datafactory.models.DatasetFolder - """ - - _validation = { - "type": {"required": True}, - "linked_service_name": {"required": True}, - } - - _attribute_map = { - "additional_properties": {"key": "", "type": "{object}"}, - "type": {"key": "type", "type": "str"}, - "description": {"key": "description", "type": "str"}, - "structure": {"key": "structure", "type": "object"}, - "schema": {"key": "schema", "type": "object"}, - "linked_service_name": {"key": "linkedServiceName", "type": "LinkedServiceReference"}, - "parameters": {"key": "parameters", "type": "{ParameterSpecification}"}, - "annotations": {"key": "annotations", "type": "[object]"}, - "folder": {"key": "folder", "type": "DatasetFolder"}, - } - - def __init__( - self, - *, - linked_service_name: "_models.LinkedServiceReference", - additional_properties: Optional[Dict[str, JSON]] = None, - description: Optional[str] = None, - structure: Optional[JSON] = None, - schema: Optional[JSON] = None, - parameters: Optional[Dict[str, "_models.ParameterSpecification"]] = None, - annotations: Optional[List[JSON]] = None, - folder: Optional["_models.DatasetFolder"] = None, - **kwargs: Any - ) -> None: - """ - :keyword additional_properties: Unmatched properties from the message are deserialized to this - collection. - :paramtype additional_properties: dict[str, JSON] - :keyword description: Dataset description. - :paramtype description: str - :keyword structure: Columns that define the structure of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetDataElement. - :paramtype structure: JSON - :keyword schema: Columns that define the physical type schema of the dataset. Type: array (or - Expression with resultType array), itemType: DatasetSchemaDataElement. - :paramtype schema: JSON - :keyword linked_service_name: Linked service reference. Required. - :paramtype linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference - :keyword parameters: Parameters for dataset. - :paramtype parameters: dict[str, ~azure.mgmt.datafactory.models.ParameterSpecification] - :keyword annotations: List of tags that can be used for describing the Dataset. - :paramtype annotations: list[JSON] - :keyword folder: The folder that this Dataset is in. If not specified, Dataset will appear at - the root level. - :paramtype folder: ~azure.mgmt.datafactory.models.DatasetFolder - """ - super().__init__( - additional_properties=additional_properties, - description=description, - structure=structure, - schema=schema, - linked_service_name=linked_service_name, - parameters=parameters, - annotations=annotations, - folder=folder, **kwargs ) - self.type: str = "SapBwCube" + self.type: str = "SalesforceServiceCloudV2Source" + self.soql_query = soql_query + self.include_deleted_objects = include_deleted_objects + self.additional_columns = additional_columns -class SapBWLinkedService(LinkedService): # pylint: disable=too-many-instance-attributes - """SAP Business Warehouse Linked Service. +class SalesforceSink(CopySink): # pylint: disable=too-many-instance-attributes + """A copy activity Salesforce sink. All required parameters must be populated in order to send to Azure. :ivar additional_properties: Unmatched properties from the message are deserialized to this collection. :vartype additional_properties: dict[str, JSON] - :ivar type: Type of linked service. Required. + :ivar type: Copy sink type. Required. :vartype type: str - :ivar connect_via: The integration runtime reference. - :vartype connect_via: ~azure.mgmt.datafactory.models.IntegrationRuntimeReference - :ivar description: Linked service description. - :vartype description: str - :ivar parameters: Parameters for linked service. - :vartype parameters: dict[str, ~azure.mgmt.datafactory.models.ParameterSpecification] - :ivar annotations: List of tags that can be used for describing the linked service. - :vartype annotations: list[JSON] - :ivar server: Host name of the SAP BW instance. Type: string (or Expression with resultType - string). Required. - :vartype server: JSON - :ivar system_number: System number of the BW system. (Usually a two-digit decimal number - represented as a string.) Type: string (or Expression with resultType string). Required. - :vartype system_number: JSON - :ivar client_id: Client ID of the client on the BW system. (Usually a three-digit decimal - number represented as a string) Type: string (or Expression with resultType string). Required. - :vartype client_id: JSON - :ivar user_name: Username to access the SAP BW server. Type: string (or Expression with - resultType string). - :vartype user_name: JSON - :ivar password: Password to access the SAP BW server. - :vartype password: ~azure.mgmt.datafactory.models.SecretBase - :ivar encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Type: string. - :vartype encrypted_credential: str + :ivar write_batch_size: Write batch size. Type: integer (or Expression with resultType + integer), minimum: 0. + :vartype write_batch_size: JSON + :ivar write_batch_timeout: Write batch timeout. Type: string (or Expression with resultType + string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :vartype write_batch_timeout: JSON + :ivar sink_retry_count: Sink retry count. Type: integer (or Expression with resultType + integer). + :vartype sink_retry_count: JSON + :ivar sink_retry_wait: Sink retry wait. Type: string (or Expression with resultType string), + pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :vartype sink_retry_wait: JSON + :ivar max_concurrent_connections: The maximum concurrent connection count for the sink data + store. Type: integer (or Expression with resultType integer). + :vartype max_concurrent_connections: JSON + :ivar disable_metrics_collection: If true, disable data store metrics collection. Default is + false. Type: boolean (or Expression with resultType boolean). + :vartype disable_metrics_collection: JSON + :ivar write_behavior: The write behavior for the operation. Default is Insert. Known values + are: "Insert" and "Upsert". + :vartype write_behavior: str or ~azure.mgmt.datafactory.models.SalesforceSinkWriteBehavior + :ivar external_id_field_name: The name of the external ID field for upsert operation. Default + value is 'Id' column. Type: string (or Expression with resultType string). + :vartype external_id_field_name: JSON + :ivar ignore_null_values: The flag indicating whether or not to ignore null values from input + dataset (except key fields) during write operation. Default value is false. If set it to true, + it means ADF will leave the data in the destination object unchanged when doing upsert/update + operation and insert defined default value when doing insert operation, versus ADF will update + the data in the destination object to NULL when doing upsert/update operation and insert NULL + value when doing insert operation. Type: boolean (or Expression with resultType boolean). + :vartype ignore_null_values: JSON """ _validation = { "type": {"required": True}, - "server": {"required": True}, - "system_number": {"required": True}, - "client_id": {"required": True}, } _attribute_map = { "additional_properties": {"key": "", "type": "{object}"}, "type": {"key": "type", "type": "str"}, - "connect_via": {"key": "connectVia", "type": "IntegrationRuntimeReference"}, - "description": {"key": "description", "type": "str"}, - "parameters": {"key": "parameters", "type": "{ParameterSpecification}"}, - "annotations": {"key": "annotations", "type": "[object]"}, - "server": {"key": "typeProperties.server", "type": "object"}, - "system_number": {"key": "typeProperties.systemNumber", "type": "object"}, - "client_id": {"key": "typeProperties.clientId", "type": "object"}, - "user_name": {"key": "typeProperties.userName", "type": "object"}, - "password": {"key": "typeProperties.password", "type": "SecretBase"}, - "encrypted_credential": {"key": "typeProperties.encryptedCredential", "type": "str"}, + "write_batch_size": {"key": "writeBatchSize", "type": "object"}, + "write_batch_timeout": {"key": "writeBatchTimeout", "type": "object"}, + "sink_retry_count": {"key": "sinkRetryCount", "type": "object"}, + "sink_retry_wait": {"key": "sinkRetryWait", "type": "object"}, + "max_concurrent_connections": {"key": "maxConcurrentConnections", "type": "object"}, + "disable_metrics_collection": {"key": "disableMetricsCollection", "type": "object"}, + "write_behavior": {"key": "writeBehavior", "type": "str"}, + "external_id_field_name": {"key": "externalIdFieldName", "type": "object"}, + "ignore_null_values": {"key": "ignoreNullValues", "type": "object"}, } def __init__( self, *, - server: JSON, - system_number: JSON, - client_id: JSON, additional_properties: Optional[Dict[str, JSON]] = None, - connect_via: Optional["_models.IntegrationRuntimeReference"] = None, - description: Optional[str] = None, - parameters: Optional[Dict[str, "_models.ParameterSpecification"]] = None, - annotations: Optional[List[JSON]] = None, - user_name: Optional[JSON] = None, - password: Optional["_models.SecretBase"] = None, - encrypted_credential: Optional[str] = None, + write_batch_size: Optional[JSON] = None, + write_batch_timeout: Optional[JSON] = None, + sink_retry_count: Optional[JSON] = None, + sink_retry_wait: Optional[JSON] = None, + max_concurrent_connections: Optional[JSON] = None, + disable_metrics_collection: Optional[JSON] = None, + write_behavior: Optional[Union[str, "_models.SalesforceSinkWriteBehavior"]] = None, + external_id_field_name: Optional[JSON] = None, + ignore_null_values: Optional[JSON] = None, **kwargs: Any ) -> None: """ :keyword additional_properties: Unmatched properties from the message are deserialized to this collection. :paramtype additional_properties: dict[str, JSON] - :keyword connect_via: The integration runtime reference. - :paramtype connect_via: ~azure.mgmt.datafactory.models.IntegrationRuntimeReference - :keyword description: Linked service description. - :paramtype description: str - :keyword parameters: Parameters for linked service. - :paramtype parameters: dict[str, ~azure.mgmt.datafactory.models.ParameterSpecification] - :keyword annotations: List of tags that can be used for describing the linked service. - :paramtype annotations: list[JSON] - :keyword server: Host name of the SAP BW instance. Type: string (or Expression with resultType - string). Required. - :paramtype server: JSON - :keyword system_number: System number of the BW system. (Usually a two-digit decimal number - represented as a string.) Type: string (or Expression with resultType string). Required. - :paramtype system_number: JSON - :keyword client_id: Client ID of the client on the BW system. (Usually a three-digit decimal - number represented as a string) Type: string (or Expression with resultType string). Required. - :paramtype client_id: JSON - :keyword user_name: Username to access the SAP BW server. Type: string (or Expression with - resultType string). - :paramtype user_name: JSON - :keyword password: Password to access the SAP BW server. - :paramtype password: ~azure.mgmt.datafactory.models.SecretBase - :keyword encrypted_credential: The encrypted credential used for authentication. Credentials - are encrypted using the integration runtime credential manager. Type: string. - :paramtype encrypted_credential: str + :keyword write_batch_size: Write batch size. Type: integer (or Expression with resultType + integer), minimum: 0. + :paramtype write_batch_size: JSON + :keyword write_batch_timeout: Write batch timeout. Type: string (or Expression with resultType + string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :paramtype write_batch_timeout: JSON + :keyword sink_retry_count: Sink retry count. Type: integer (or Expression with resultType + integer). + :paramtype sink_retry_count: JSON + :keyword sink_retry_wait: Sink retry wait. Type: string (or Expression with resultType string), + pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :paramtype sink_retry_wait: JSON + :keyword max_concurrent_connections: The maximum concurrent connection count for the sink data + store. Type: integer (or Expression with resultType integer). + :paramtype max_concurrent_connections: JSON + :keyword disable_metrics_collection: If true, disable data store metrics collection. Default is + false. Type: boolean (or Expression with resultType boolean). + :paramtype disable_metrics_collection: JSON + :keyword write_behavior: The write behavior for the operation. Default is Insert. Known values + are: "Insert" and "Upsert". + :paramtype write_behavior: str or ~azure.mgmt.datafactory.models.SalesforceSinkWriteBehavior + :keyword external_id_field_name: The name of the external ID field for upsert operation. + Default value is 'Id' column. Type: string (or Expression with resultType string). + :paramtype external_id_field_name: JSON + :keyword ignore_null_values: The flag indicating whether or not to ignore null values from + input dataset (except key fields) during write operation. Default value is false. If set it to + true, it means ADF will leave the data in the destination object unchanged when doing + upsert/update operation and insert defined default value when doing insert operation, versus + ADF will update the data in the destination object to NULL when doing upsert/update operation + and insert NULL value when doing insert operation. Type: boolean (or Expression with resultType + boolean). + :paramtype ignore_null_values: JSON """ super().__init__( additional_properties=additional_properties, - connect_via=connect_via, - description=description, - parameters=parameters, - annotations=annotations, + write_batch_size=write_batch_size, + write_batch_timeout=write_batch_timeout, + sink_retry_count=sink_retry_count, + sink_retry_wait=sink_retry_wait, + max_concurrent_connections=max_concurrent_connections, + disable_metrics_collection=disable_metrics_collection, **kwargs ) - self.type: str = "SapBW" - self.server = server - self.system_number = system_number - self.client_id = client_id - self.user_name = user_name - self.password = password - self.encrypted_credential = encrypted_credential + self.type: str = "SalesforceSink" + self.write_behavior = write_behavior + self.external_id_field_name = external_id_field_name + self.ignore_null_values = ignore_null_values -class SapBwSource(TabularSource): - """A copy activity source for SapBW server via MDX. +class SalesforceSource(TabularSource): + """A copy activity Salesforce source. All required parameters must be populated in order to send to Azure. @@ -52518,8 +52843,11 @@ class SapBwSource(TabularSource): :ivar additional_columns: Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). :vartype additional_columns: JSON - :ivar query: MDX query. Type: string (or Expression with resultType string). + :ivar query: Database query. Type: string (or Expression with resultType string). :vartype query: JSON + :ivar read_behavior: The read behavior for the operation. Default is Query. Allowed values: + Query/QueryAll. Type: string (or Expression with resultType string). + :vartype read_behavior: JSON """ _validation = { @@ -52536,6 +52864,7 @@ class SapBwSource(TabularSource): "query_timeout": {"key": "queryTimeout", "type": "object"}, "additional_columns": {"key": "additionalColumns", "type": "object"}, "query": {"key": "query", "type": "object"}, + "read_behavior": {"key": "readBehavior", "type": "object"}, } def __init__( @@ -52549,6 +52878,7 @@ def __init__( query_timeout: Optional[JSON] = None, additional_columns: Optional[JSON] = None, query: Optional[JSON] = None, + read_behavior: Optional[JSON] = None, **kwargs: Any ) -> None: """ @@ -52573,8 +52903,11 @@ def __init__( :keyword additional_columns: Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). :paramtype additional_columns: JSON - :keyword query: MDX query. Type: string (or Expression with resultType string). + :keyword query: Database query. Type: string (or Expression with resultType string). :paramtype query: JSON + :keyword read_behavior: The read behavior for the operation. Default is Query. Allowed values: + Query/QueryAll. Type: string (or Expression with resultType string). + :paramtype read_behavior: JSON """ super().__init__( additional_properties=additional_properties, @@ -52586,12 +52919,13 @@ def __init__( additional_columns=additional_columns, **kwargs ) - self.type: str = "SapBwSource" + self.type: str = "SalesforceSource" self.query = query + self.read_behavior = read_behavior -class SapCloudForCustomerLinkedService(LinkedService): - """Linked service for SAP Cloud for Customer. +class SalesforceV2LinkedService(LinkedService): # pylint: disable=too-many-instance-attributes + """Linked service for Salesforce V2. All required parameters must be populated in order to send to Azure. @@ -52608,24 +52942,29 @@ class SapCloudForCustomerLinkedService(LinkedService): :vartype parameters: dict[str, ~azure.mgmt.datafactory.models.ParameterSpecification] :ivar annotations: List of tags that can be used for describing the linked service. :vartype annotations: list[JSON] - :ivar url: The URL of SAP Cloud for Customer OData API. For example, - '[https://[tenantname].crm.ondemand.com/sap/c4c/odata/v1]'. Type: string (or Expression with - resultType string). Required. - :vartype url: JSON - :ivar username: The username for Basic authentication. Type: string (or Expression with + :ivar environment_url: The URL of Salesforce instance. For example, + 'https://[domain].my.salesforce.com'. Type: string (or Expression with resultType string). + :vartype environment_url: JSON + :ivar authentication_type: The authentication type to be used to connect to the Salesforce. + Currently, we only support OAuth2ClientCredentials, it is also the default value. + :vartype authentication_type: JSON + :ivar client_id: The client Id for OAuth 2.0 Client Credentials Flow authentication of the + Salesforce instance. Type: string (or Expression with resultType string). + :vartype client_id: JSON + :ivar client_secret: The client secret for OAuth 2.0 Client Credentials Flow authentication of + the Salesforce instance. + :vartype client_secret: ~azure.mgmt.datafactory.models.SecretBase + :ivar api_version: The Salesforce API version used in ADF. The version must be larger than or + equal to 47.0 which is required by Salesforce BULK API 2.0. Type: string (or Expression with resultType string). - :vartype username: JSON - :ivar password: The password for Basic authentication. - :vartype password: ~azure.mgmt.datafactory.models.SecretBase + :vartype api_version: JSON :ivar encrypted_credential: The encrypted credential used for authentication. Credentials are - encrypted using the integration runtime credential manager. Either encryptedCredential or - username/password must be provided. Type: string. + encrypted using the integration runtime credential manager. Type: string. :vartype encrypted_credential: str """ _validation = { "type": {"required": True}, - "url": {"required": True}, } _attribute_map = { @@ -52635,23 +52974,27 @@ class SapCloudForCustomerLinkedService(LinkedService): "description": {"key": "description", "type": "str"}, "parameters": {"key": "parameters", "type": "{ParameterSpecification}"}, "annotations": {"key": "annotations", "type": "[object]"}, - "url": {"key": "typeProperties.url", "type": "object"}, - "username": {"key": "typeProperties.username", "type": "object"}, - "password": {"key": "typeProperties.password", "type": "SecretBase"}, + "environment_url": {"key": "typeProperties.environmentUrl", "type": "object"}, + "authentication_type": {"key": "typeProperties.authenticationType", "type": "object"}, + "client_id": {"key": "typeProperties.clientId", "type": "object"}, + "client_secret": {"key": "typeProperties.clientSecret", "type": "SecretBase"}, + "api_version": {"key": "typeProperties.apiVersion", "type": "object"}, "encrypted_credential": {"key": "typeProperties.encryptedCredential", "type": "str"}, } def __init__( self, *, - url: JSON, additional_properties: Optional[Dict[str, JSON]] = None, connect_via: Optional["_models.IntegrationRuntimeReference"] = None, description: Optional[str] = None, parameters: Optional[Dict[str, "_models.ParameterSpecification"]] = None, annotations: Optional[List[JSON]] = None, - username: Optional[JSON] = None, - password: Optional["_models.SecretBase"] = None, + environment_url: Optional[JSON] = None, + authentication_type: Optional[JSON] = None, + client_id: Optional[JSON] = None, + client_secret: Optional["_models.SecretBase"] = None, + api_version: Optional[JSON] = None, encrypted_credential: Optional[str] = None, **kwargs: Any ) -> None: @@ -52667,18 +53010,24 @@ def __init__( :paramtype parameters: dict[str, ~azure.mgmt.datafactory.models.ParameterSpecification] :keyword annotations: List of tags that can be used for describing the linked service. :paramtype annotations: list[JSON] - :keyword url: The URL of SAP Cloud for Customer OData API. For example, - '[https://[tenantname].crm.ondemand.com/sap/c4c/odata/v1]'. Type: string (or Expression with - resultType string). Required. - :paramtype url: JSON - :keyword username: The username for Basic authentication. Type: string (or Expression with + :keyword environment_url: The URL of Salesforce instance. For example, + 'https://[domain].my.salesforce.com'. Type: string (or Expression with resultType string). + :paramtype environment_url: JSON + :keyword authentication_type: The authentication type to be used to connect to the Salesforce. + Currently, we only support OAuth2ClientCredentials, it is also the default value. + :paramtype authentication_type: JSON + :keyword client_id: The client Id for OAuth 2.0 Client Credentials Flow authentication of the + Salesforce instance. Type: string (or Expression with resultType string). + :paramtype client_id: JSON + :keyword client_secret: The client secret for OAuth 2.0 Client Credentials Flow authentication + of the Salesforce instance. + :paramtype client_secret: ~azure.mgmt.datafactory.models.SecretBase + :keyword api_version: The Salesforce API version used in ADF. The version must be larger than + or equal to 47.0 which is required by Salesforce BULK API 2.0. Type: string (or Expression with resultType string). - :paramtype username: JSON - :keyword password: The password for Basic authentication. - :paramtype password: ~azure.mgmt.datafactory.models.SecretBase + :paramtype api_version: JSON :keyword encrypted_credential: The encrypted credential used for authentication. Credentials - are encrypted using the integration runtime credential manager. Either encryptedCredential or - username/password must be provided. Type: string. + are encrypted using the integration runtime credential manager. Type: string. :paramtype encrypted_credential: str """ super().__init__( @@ -52689,15 +53038,17 @@ def __init__( annotations=annotations, **kwargs ) - self.type: str = "SapCloudForCustomer" - self.url = url - self.username = username - self.password = password + self.type: str = "SalesforceV2" + self.environment_url = environment_url + self.authentication_type = authentication_type + self.client_id = client_id + self.client_secret = client_secret + self.api_version = api_version self.encrypted_credential = encrypted_credential -class SapCloudForCustomerResourceDataset(Dataset): - """The path of the SAP Cloud for Customer OData entity. +class SalesforceV2ObjectDataset(Dataset): # pylint: disable=too-many-instance-attributes + """The Salesforce V2 object dataset. All required parameters must be populated in order to send to Azure. @@ -52723,15 +53074,17 @@ class SapCloudForCustomerResourceDataset(Dataset): :ivar folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. :vartype folder: ~azure.mgmt.datafactory.models.DatasetFolder - :ivar path: The path of the SAP Cloud for Customer OData entity. Type: string (or Expression - with resultType string). Required. - :vartype path: JSON + :ivar object_api_name: The Salesforce V2 object API name. Type: string (or Expression with + resultType string). + :vartype object_api_name: JSON + :ivar report_id: The Salesforce V2 report Id. Type: string (or Expression with resultType + string). + :vartype report_id: JSON """ _validation = { "type": {"required": True}, "linked_service_name": {"required": True}, - "path": {"required": True}, } _attribute_map = { @@ -52744,14 +53097,14 @@ class SapCloudForCustomerResourceDataset(Dataset): "parameters": {"key": "parameters", "type": "{ParameterSpecification}"}, "annotations": {"key": "annotations", "type": "[object]"}, "folder": {"key": "folder", "type": "DatasetFolder"}, - "path": {"key": "typeProperties.path", "type": "object"}, + "object_api_name": {"key": "typeProperties.objectApiName", "type": "object"}, + "report_id": {"key": "typeProperties.reportId", "type": "object"}, } def __init__( self, *, linked_service_name: "_models.LinkedServiceReference", - path: JSON, additional_properties: Optional[Dict[str, JSON]] = None, description: Optional[str] = None, structure: Optional[JSON] = None, @@ -52759,6 +53112,8 @@ def __init__( parameters: Optional[Dict[str, "_models.ParameterSpecification"]] = None, annotations: Optional[List[JSON]] = None, folder: Optional["_models.DatasetFolder"] = None, + object_api_name: Optional[JSON] = None, + report_id: Optional[JSON] = None, **kwargs: Any ) -> None: """ @@ -52782,9 +53137,12 @@ def __init__( :keyword folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. :paramtype folder: ~azure.mgmt.datafactory.models.DatasetFolder - :keyword path: The path of the SAP Cloud for Customer OData entity. Type: string (or Expression - with resultType string). Required. - :paramtype path: JSON + :keyword object_api_name: The Salesforce V2 object API name. Type: string (or Expression with + resultType string). + :paramtype object_api_name: JSON + :keyword report_id: The Salesforce V2 report Id. Type: string (or Expression with resultType + string). + :paramtype report_id: JSON """ super().__init__( additional_properties=additional_properties, @@ -52797,12 +53155,13 @@ def __init__( folder=folder, **kwargs ) - self.type: str = "SapCloudForCustomerResource" - self.path = path + self.type: str = "SalesforceV2Object" + self.object_api_name = object_api_name + self.report_id = report_id -class SapCloudForCustomerSink(CopySink): - """A copy activity SAP Cloud for Customer sink. +class SalesforceV2Sink(CopySink): # pylint: disable=too-many-instance-attributes + """A copy activity Salesforce V2 sink. All required parameters must be populated in order to send to Azure. @@ -52829,15 +53188,19 @@ class SapCloudForCustomerSink(CopySink): :ivar disable_metrics_collection: If true, disable data store metrics collection. Default is false. Type: boolean (or Expression with resultType boolean). :vartype disable_metrics_collection: JSON - :ivar write_behavior: The write behavior for the operation. Default is 'Insert'. Known values - are: "Insert" and "Update". - :vartype write_behavior: str or - ~azure.mgmt.datafactory.models.SapCloudForCustomerSinkWriteBehavior - :ivar http_request_timeout: The timeout (TimeSpan) to get an HTTP response. It is the timeout - to get a response, not the timeout to read response data. Default value: 00:05:00. Type: string - (or Expression with resultType string), pattern: - ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :vartype http_request_timeout: JSON + :ivar write_behavior: The write behavior for the operation. Default is Insert. Known values + are: "Insert" and "Upsert". + :vartype write_behavior: str or ~azure.mgmt.datafactory.models.SalesforceV2SinkWriteBehavior + :ivar external_id_field_name: The name of the external ID field for upsert operation. Default + value is 'Id' column. Type: string (or Expression with resultType string). + :vartype external_id_field_name: JSON + :ivar ignore_null_values: The flag indicating whether or not to ignore null values from input + dataset (except key fields) during write operation. Default value is false. If set it to true, + it means ADF will leave the data in the destination object unchanged when doing upsert/update + operation and insert defined default value when doing insert operation, versus ADF will update + the data in the destination object to NULL when doing upsert/update operation and insert NULL + value when doing insert operation. Type: boolean (or Expression with resultType boolean). + :vartype ignore_null_values: JSON """ _validation = { @@ -52854,7 +53217,8 @@ class SapCloudForCustomerSink(CopySink): "max_concurrent_connections": {"key": "maxConcurrentConnections", "type": "object"}, "disable_metrics_collection": {"key": "disableMetricsCollection", "type": "object"}, "write_behavior": {"key": "writeBehavior", "type": "str"}, - "http_request_timeout": {"key": "httpRequestTimeout", "type": "object"}, + "external_id_field_name": {"key": "externalIdFieldName", "type": "object"}, + "ignore_null_values": {"key": "ignoreNullValues", "type": "object"}, } def __init__( @@ -52867,8 +53231,9 @@ def __init__( sink_retry_wait: Optional[JSON] = None, max_concurrent_connections: Optional[JSON] = None, disable_metrics_collection: Optional[JSON] = None, - write_behavior: Optional[Union[str, "_models.SapCloudForCustomerSinkWriteBehavior"]] = None, - http_request_timeout: Optional[JSON] = None, + write_behavior: Optional[Union[str, "_models.SalesforceV2SinkWriteBehavior"]] = None, + external_id_field_name: Optional[JSON] = None, + ignore_null_values: Optional[JSON] = None, **kwargs: Any ) -> None: """ @@ -52893,15 +53258,20 @@ def __init__( :keyword disable_metrics_collection: If true, disable data store metrics collection. Default is false. Type: boolean (or Expression with resultType boolean). :paramtype disable_metrics_collection: JSON - :keyword write_behavior: The write behavior for the operation. Default is 'Insert'. Known - values are: "Insert" and "Update". - :paramtype write_behavior: str or - ~azure.mgmt.datafactory.models.SapCloudForCustomerSinkWriteBehavior - :keyword http_request_timeout: The timeout (TimeSpan) to get an HTTP response. It is the - timeout to get a response, not the timeout to read response data. Default value: 00:05:00. - Type: string (or Expression with resultType string), pattern: - ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :paramtype http_request_timeout: JSON + :keyword write_behavior: The write behavior for the operation. Default is Insert. Known values + are: "Insert" and "Upsert". + :paramtype write_behavior: str or ~azure.mgmt.datafactory.models.SalesforceV2SinkWriteBehavior + :keyword external_id_field_name: The name of the external ID field for upsert operation. + Default value is 'Id' column. Type: string (or Expression with resultType string). + :paramtype external_id_field_name: JSON + :keyword ignore_null_values: The flag indicating whether or not to ignore null values from + input dataset (except key fields) during write operation. Default value is false. If set it to + true, it means ADF will leave the data in the destination object unchanged when doing + upsert/update operation and insert defined default value when doing insert operation, versus + ADF will update the data in the destination object to NULL when doing upsert/update operation + and insert NULL value when doing insert operation. Type: boolean (or Expression with resultType + boolean). + :paramtype ignore_null_values: JSON """ super().__init__( additional_properties=additional_properties, @@ -52913,13 +53283,14 @@ def __init__( disable_metrics_collection=disable_metrics_collection, **kwargs ) - self.type: str = "SapCloudForCustomerSink" + self.type: str = "SalesforceV2Sink" self.write_behavior = write_behavior - self.http_request_timeout = http_request_timeout + self.external_id_field_name = external_id_field_name + self.ignore_null_values = ignore_null_values -class SapCloudForCustomerSource(TabularSource): - """A copy activity source for SAP Cloud for Customer source. +class SalesforceV2Source(TabularSource): + """A copy activity Salesforce V2 source. All required parameters must be populated in order to send to Azure. @@ -52946,14 +53317,11 @@ class SapCloudForCustomerSource(TabularSource): :ivar additional_columns: Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). :vartype additional_columns: JSON - :ivar query: SAP Cloud for Customer OData query. For example, "$top=1". Type: string (or - Expression with resultType string). - :vartype query: JSON - :ivar http_request_timeout: The timeout (TimeSpan) to get an HTTP response. It is the timeout - to get a response, not the timeout to read response data. Default value: 00:05:00. Type: string - (or Expression with resultType string), pattern: - ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :vartype http_request_timeout: JSON + :ivar soql_query: Database query. Type: string (or Expression with resultType string). + :vartype soql_query: JSON + :ivar include_deleted_objects: This property control whether query result contains Deleted + objects. Default is false. Type: boolean (or Expression with resultType boolean). + :vartype include_deleted_objects: JSON """ _validation = { @@ -52969,8 +53337,8 @@ class SapCloudForCustomerSource(TabularSource): "disable_metrics_collection": {"key": "disableMetricsCollection", "type": "object"}, "query_timeout": {"key": "queryTimeout", "type": "object"}, "additional_columns": {"key": "additionalColumns", "type": "object"}, - "query": {"key": "query", "type": "object"}, - "http_request_timeout": {"key": "httpRequestTimeout", "type": "object"}, + "soql_query": {"key": "SOQLQuery", "type": "object"}, + "include_deleted_objects": {"key": "includeDeletedObjects", "type": "object"}, } def __init__( @@ -52983,8 +53351,8 @@ def __init__( disable_metrics_collection: Optional[JSON] = None, query_timeout: Optional[JSON] = None, additional_columns: Optional[JSON] = None, - query: Optional[JSON] = None, - http_request_timeout: Optional[JSON] = None, + soql_query: Optional[JSON] = None, + include_deleted_objects: Optional[JSON] = None, **kwargs: Any ) -> None: """ @@ -53009,14 +53377,11 @@ def __init__( :keyword additional_columns: Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). :paramtype additional_columns: JSON - :keyword query: SAP Cloud for Customer OData query. For example, "$top=1". Type: string (or - Expression with resultType string). - :paramtype query: JSON - :keyword http_request_timeout: The timeout (TimeSpan) to get an HTTP response. It is the - timeout to get a response, not the timeout to read response data. Default value: 00:05:00. - Type: string (or Expression with resultType string), pattern: - ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :paramtype http_request_timeout: JSON + :keyword soql_query: Database query. Type: string (or Expression with resultType string). + :paramtype soql_query: JSON + :keyword include_deleted_objects: This property control whether query result contains Deleted + objects. Default is false. Type: boolean (or Expression with resultType boolean). + :paramtype include_deleted_objects: JSON """ super().__init__( additional_properties=additional_properties, @@ -53028,33 +53393,793 @@ def __init__( additional_columns=additional_columns, **kwargs ) - self.type: str = "SapCloudForCustomerSource" - self.query = query - self.http_request_timeout = http_request_timeout + self.type: str = "SalesforceV2Source" + self.soql_query = soql_query + self.include_deleted_objects = include_deleted_objects -class SapEccLinkedService(LinkedService): - """Linked service for SAP ERP Central Component(SAP ECC). +class SapBwCubeDataset(Dataset): + """The SAP BW cube dataset. All required parameters must be populated in order to send to Azure. :ivar additional_properties: Unmatched properties from the message are deserialized to this collection. :vartype additional_properties: dict[str, JSON] - :ivar type: Type of linked service. Required. + :ivar type: Type of dataset. Required. :vartype type: str - :ivar connect_via: The integration runtime reference. - :vartype connect_via: ~azure.mgmt.datafactory.models.IntegrationRuntimeReference - :ivar description: Linked service description. + :ivar description: Dataset description. :vartype description: str - :ivar parameters: Parameters for linked service. - :vartype parameters: dict[str, ~azure.mgmt.datafactory.models.ParameterSpecification] - :ivar annotations: List of tags that can be used for describing the linked service. - :vartype annotations: list[JSON] - :ivar url: The URL of SAP ECC OData API. For example, - '[https://hostname:port/sap/opu/odata/sap/servicename/]'. Type: string (or Expression with - resultType string). Required. - :vartype url: JSON + :ivar structure: Columns that define the structure of the dataset. Type: array (or Expression + with resultType array), itemType: DatasetDataElement. + :vartype structure: JSON + :ivar schema: Columns that define the physical type schema of the dataset. Type: array (or + Expression with resultType array), itemType: DatasetSchemaDataElement. + :vartype schema: JSON + :ivar linked_service_name: Linked service reference. Required. + :vartype linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference + :ivar parameters: Parameters for dataset. + :vartype parameters: dict[str, ~azure.mgmt.datafactory.models.ParameterSpecification] + :ivar annotations: List of tags that can be used for describing the Dataset. + :vartype annotations: list[JSON] + :ivar folder: The folder that this Dataset is in. If not specified, Dataset will appear at the + root level. + :vartype folder: ~azure.mgmt.datafactory.models.DatasetFolder + """ + + _validation = { + "type": {"required": True}, + "linked_service_name": {"required": True}, + } + + _attribute_map = { + "additional_properties": {"key": "", "type": "{object}"}, + "type": {"key": "type", "type": "str"}, + "description": {"key": "description", "type": "str"}, + "structure": {"key": "structure", "type": "object"}, + "schema": {"key": "schema", "type": "object"}, + "linked_service_name": {"key": "linkedServiceName", "type": "LinkedServiceReference"}, + "parameters": {"key": "parameters", "type": "{ParameterSpecification}"}, + "annotations": {"key": "annotations", "type": "[object]"}, + "folder": {"key": "folder", "type": "DatasetFolder"}, + } + + def __init__( + self, + *, + linked_service_name: "_models.LinkedServiceReference", + additional_properties: Optional[Dict[str, JSON]] = None, + description: Optional[str] = None, + structure: Optional[JSON] = None, + schema: Optional[JSON] = None, + parameters: Optional[Dict[str, "_models.ParameterSpecification"]] = None, + annotations: Optional[List[JSON]] = None, + folder: Optional["_models.DatasetFolder"] = None, + **kwargs: Any + ) -> None: + """ + :keyword additional_properties: Unmatched properties from the message are deserialized to this + collection. + :paramtype additional_properties: dict[str, JSON] + :keyword description: Dataset description. + :paramtype description: str + :keyword structure: Columns that define the structure of the dataset. Type: array (or + Expression with resultType array), itemType: DatasetDataElement. + :paramtype structure: JSON + :keyword schema: Columns that define the physical type schema of the dataset. Type: array (or + Expression with resultType array), itemType: DatasetSchemaDataElement. + :paramtype schema: JSON + :keyword linked_service_name: Linked service reference. Required. + :paramtype linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference + :keyword parameters: Parameters for dataset. + :paramtype parameters: dict[str, ~azure.mgmt.datafactory.models.ParameterSpecification] + :keyword annotations: List of tags that can be used for describing the Dataset. + :paramtype annotations: list[JSON] + :keyword folder: The folder that this Dataset is in. If not specified, Dataset will appear at + the root level. + :paramtype folder: ~azure.mgmt.datafactory.models.DatasetFolder + """ + super().__init__( + additional_properties=additional_properties, + description=description, + structure=structure, + schema=schema, + linked_service_name=linked_service_name, + parameters=parameters, + annotations=annotations, + folder=folder, + **kwargs + ) + self.type: str = "SapBwCube" + + +class SapBWLinkedService(LinkedService): # pylint: disable=too-many-instance-attributes + """SAP Business Warehouse Linked Service. + + All required parameters must be populated in order to send to Azure. + + :ivar additional_properties: Unmatched properties from the message are deserialized to this + collection. + :vartype additional_properties: dict[str, JSON] + :ivar type: Type of linked service. Required. + :vartype type: str + :ivar connect_via: The integration runtime reference. + :vartype connect_via: ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :ivar description: Linked service description. + :vartype description: str + :ivar parameters: Parameters for linked service. + :vartype parameters: dict[str, ~azure.mgmt.datafactory.models.ParameterSpecification] + :ivar annotations: List of tags that can be used for describing the linked service. + :vartype annotations: list[JSON] + :ivar server: Host name of the SAP BW instance. Type: string (or Expression with resultType + string). Required. + :vartype server: JSON + :ivar system_number: System number of the BW system. (Usually a two-digit decimal number + represented as a string.) Type: string (or Expression with resultType string). Required. + :vartype system_number: JSON + :ivar client_id: Client ID of the client on the BW system. (Usually a three-digit decimal + number represented as a string) Type: string (or Expression with resultType string). Required. + :vartype client_id: JSON + :ivar user_name: Username to access the SAP BW server. Type: string (or Expression with + resultType string). + :vartype user_name: JSON + :ivar password: Password to access the SAP BW server. + :vartype password: ~azure.mgmt.datafactory.models.SecretBase + :ivar encrypted_credential: The encrypted credential used for authentication. Credentials are + encrypted using the integration runtime credential manager. Type: string. + :vartype encrypted_credential: str + """ + + _validation = { + "type": {"required": True}, + "server": {"required": True}, + "system_number": {"required": True}, + "client_id": {"required": True}, + } + + _attribute_map = { + "additional_properties": {"key": "", "type": "{object}"}, + "type": {"key": "type", "type": "str"}, + "connect_via": {"key": "connectVia", "type": "IntegrationRuntimeReference"}, + "description": {"key": "description", "type": "str"}, + "parameters": {"key": "parameters", "type": "{ParameterSpecification}"}, + "annotations": {"key": "annotations", "type": "[object]"}, + "server": {"key": "typeProperties.server", "type": "object"}, + "system_number": {"key": "typeProperties.systemNumber", "type": "object"}, + "client_id": {"key": "typeProperties.clientId", "type": "object"}, + "user_name": {"key": "typeProperties.userName", "type": "object"}, + "password": {"key": "typeProperties.password", "type": "SecretBase"}, + "encrypted_credential": {"key": "typeProperties.encryptedCredential", "type": "str"}, + } + + def __init__( + self, + *, + server: JSON, + system_number: JSON, + client_id: JSON, + additional_properties: Optional[Dict[str, JSON]] = None, + connect_via: Optional["_models.IntegrationRuntimeReference"] = None, + description: Optional[str] = None, + parameters: Optional[Dict[str, "_models.ParameterSpecification"]] = None, + annotations: Optional[List[JSON]] = None, + user_name: Optional[JSON] = None, + password: Optional["_models.SecretBase"] = None, + encrypted_credential: Optional[str] = None, + **kwargs: Any + ) -> None: + """ + :keyword additional_properties: Unmatched properties from the message are deserialized to this + collection. + :paramtype additional_properties: dict[str, JSON] + :keyword connect_via: The integration runtime reference. + :paramtype connect_via: ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :keyword description: Linked service description. + :paramtype description: str + :keyword parameters: Parameters for linked service. + :paramtype parameters: dict[str, ~azure.mgmt.datafactory.models.ParameterSpecification] + :keyword annotations: List of tags that can be used for describing the linked service. + :paramtype annotations: list[JSON] + :keyword server: Host name of the SAP BW instance. Type: string (or Expression with resultType + string). Required. + :paramtype server: JSON + :keyword system_number: System number of the BW system. (Usually a two-digit decimal number + represented as a string.) Type: string (or Expression with resultType string). Required. + :paramtype system_number: JSON + :keyword client_id: Client ID of the client on the BW system. (Usually a three-digit decimal + number represented as a string) Type: string (or Expression with resultType string). Required. + :paramtype client_id: JSON + :keyword user_name: Username to access the SAP BW server. Type: string (or Expression with + resultType string). + :paramtype user_name: JSON + :keyword password: Password to access the SAP BW server. + :paramtype password: ~azure.mgmt.datafactory.models.SecretBase + :keyword encrypted_credential: The encrypted credential used for authentication. Credentials + are encrypted using the integration runtime credential manager. Type: string. + :paramtype encrypted_credential: str + """ + super().__init__( + additional_properties=additional_properties, + connect_via=connect_via, + description=description, + parameters=parameters, + annotations=annotations, + **kwargs + ) + self.type: str = "SapBW" + self.server = server + self.system_number = system_number + self.client_id = client_id + self.user_name = user_name + self.password = password + self.encrypted_credential = encrypted_credential + + +class SapBwSource(TabularSource): + """A copy activity source for SapBW server via MDX. + + All required parameters must be populated in order to send to Azure. + + :ivar additional_properties: Unmatched properties from the message are deserialized to this + collection. + :vartype additional_properties: dict[str, JSON] + :ivar type: Copy source type. Required. + :vartype type: str + :ivar source_retry_count: Source retry count. Type: integer (or Expression with resultType + integer). + :vartype source_retry_count: JSON + :ivar source_retry_wait: Source retry wait. Type: string (or Expression with resultType + string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :vartype source_retry_wait: JSON + :ivar max_concurrent_connections: The maximum concurrent connection count for the source data + store. Type: integer (or Expression with resultType integer). + :vartype max_concurrent_connections: JSON + :ivar disable_metrics_collection: If true, disable data store metrics collection. Default is + false. Type: boolean (or Expression with resultType boolean). + :vartype disable_metrics_collection: JSON + :ivar query_timeout: Query timeout. Type: string (or Expression with resultType string), + pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :vartype query_timeout: JSON + :ivar additional_columns: Specifies the additional columns to be added to source data. Type: + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :vartype additional_columns: JSON + :ivar query: MDX query. Type: string (or Expression with resultType string). + :vartype query: JSON + """ + + _validation = { + "type": {"required": True}, + } + + _attribute_map = { + "additional_properties": {"key": "", "type": "{object}"}, + "type": {"key": "type", "type": "str"}, + "source_retry_count": {"key": "sourceRetryCount", "type": "object"}, + "source_retry_wait": {"key": "sourceRetryWait", "type": "object"}, + "max_concurrent_connections": {"key": "maxConcurrentConnections", "type": "object"}, + "disable_metrics_collection": {"key": "disableMetricsCollection", "type": "object"}, + "query_timeout": {"key": "queryTimeout", "type": "object"}, + "additional_columns": {"key": "additionalColumns", "type": "object"}, + "query": {"key": "query", "type": "object"}, + } + + def __init__( + self, + *, + additional_properties: Optional[Dict[str, JSON]] = None, + source_retry_count: Optional[JSON] = None, + source_retry_wait: Optional[JSON] = None, + max_concurrent_connections: Optional[JSON] = None, + disable_metrics_collection: Optional[JSON] = None, + query_timeout: Optional[JSON] = None, + additional_columns: Optional[JSON] = None, + query: Optional[JSON] = None, + **kwargs: Any + ) -> None: + """ + :keyword additional_properties: Unmatched properties from the message are deserialized to this + collection. + :paramtype additional_properties: dict[str, JSON] + :keyword source_retry_count: Source retry count. Type: integer (or Expression with resultType + integer). + :paramtype source_retry_count: JSON + :keyword source_retry_wait: Source retry wait. Type: string (or Expression with resultType + string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :paramtype source_retry_wait: JSON + :keyword max_concurrent_connections: The maximum concurrent connection count for the source + data store. Type: integer (or Expression with resultType integer). + :paramtype max_concurrent_connections: JSON + :keyword disable_metrics_collection: If true, disable data store metrics collection. Default is + false. Type: boolean (or Expression with resultType boolean). + :paramtype disable_metrics_collection: JSON + :keyword query_timeout: Query timeout. Type: string (or Expression with resultType string), + pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :paramtype query_timeout: JSON + :keyword additional_columns: Specifies the additional columns to be added to source data. Type: + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :paramtype additional_columns: JSON + :keyword query: MDX query. Type: string (or Expression with resultType string). + :paramtype query: JSON + """ + super().__init__( + additional_properties=additional_properties, + source_retry_count=source_retry_count, + source_retry_wait=source_retry_wait, + max_concurrent_connections=max_concurrent_connections, + disable_metrics_collection=disable_metrics_collection, + query_timeout=query_timeout, + additional_columns=additional_columns, + **kwargs + ) + self.type: str = "SapBwSource" + self.query = query + + +class SapCloudForCustomerLinkedService(LinkedService): + """Linked service for SAP Cloud for Customer. + + All required parameters must be populated in order to send to Azure. + + :ivar additional_properties: Unmatched properties from the message are deserialized to this + collection. + :vartype additional_properties: dict[str, JSON] + :ivar type: Type of linked service. Required. + :vartype type: str + :ivar connect_via: The integration runtime reference. + :vartype connect_via: ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :ivar description: Linked service description. + :vartype description: str + :ivar parameters: Parameters for linked service. + :vartype parameters: dict[str, ~azure.mgmt.datafactory.models.ParameterSpecification] + :ivar annotations: List of tags that can be used for describing the linked service. + :vartype annotations: list[JSON] + :ivar url: The URL of SAP Cloud for Customer OData API. For example, + '[https://[tenantname].crm.ondemand.com/sap/c4c/odata/v1]'. Type: string (or Expression with + resultType string). Required. + :vartype url: JSON + :ivar username: The username for Basic authentication. Type: string (or Expression with + resultType string). + :vartype username: JSON + :ivar password: The password for Basic authentication. + :vartype password: ~azure.mgmt.datafactory.models.SecretBase + :ivar encrypted_credential: The encrypted credential used for authentication. Credentials are + encrypted using the integration runtime credential manager. Either encryptedCredential or + username/password must be provided. Type: string. + :vartype encrypted_credential: str + """ + + _validation = { + "type": {"required": True}, + "url": {"required": True}, + } + + _attribute_map = { + "additional_properties": {"key": "", "type": "{object}"}, + "type": {"key": "type", "type": "str"}, + "connect_via": {"key": "connectVia", "type": "IntegrationRuntimeReference"}, + "description": {"key": "description", "type": "str"}, + "parameters": {"key": "parameters", "type": "{ParameterSpecification}"}, + "annotations": {"key": "annotations", "type": "[object]"}, + "url": {"key": "typeProperties.url", "type": "object"}, + "username": {"key": "typeProperties.username", "type": "object"}, + "password": {"key": "typeProperties.password", "type": "SecretBase"}, + "encrypted_credential": {"key": "typeProperties.encryptedCredential", "type": "str"}, + } + + def __init__( + self, + *, + url: JSON, + additional_properties: Optional[Dict[str, JSON]] = None, + connect_via: Optional["_models.IntegrationRuntimeReference"] = None, + description: Optional[str] = None, + parameters: Optional[Dict[str, "_models.ParameterSpecification"]] = None, + annotations: Optional[List[JSON]] = None, + username: Optional[JSON] = None, + password: Optional["_models.SecretBase"] = None, + encrypted_credential: Optional[str] = None, + **kwargs: Any + ) -> None: + """ + :keyword additional_properties: Unmatched properties from the message are deserialized to this + collection. + :paramtype additional_properties: dict[str, JSON] + :keyword connect_via: The integration runtime reference. + :paramtype connect_via: ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :keyword description: Linked service description. + :paramtype description: str + :keyword parameters: Parameters for linked service. + :paramtype parameters: dict[str, ~azure.mgmt.datafactory.models.ParameterSpecification] + :keyword annotations: List of tags that can be used for describing the linked service. + :paramtype annotations: list[JSON] + :keyword url: The URL of SAP Cloud for Customer OData API. For example, + '[https://[tenantname].crm.ondemand.com/sap/c4c/odata/v1]'. Type: string (or Expression with + resultType string). Required. + :paramtype url: JSON + :keyword username: The username for Basic authentication. Type: string (or Expression with + resultType string). + :paramtype username: JSON + :keyword password: The password for Basic authentication. + :paramtype password: ~azure.mgmt.datafactory.models.SecretBase + :keyword encrypted_credential: The encrypted credential used for authentication. Credentials + are encrypted using the integration runtime credential manager. Either encryptedCredential or + username/password must be provided. Type: string. + :paramtype encrypted_credential: str + """ + super().__init__( + additional_properties=additional_properties, + connect_via=connect_via, + description=description, + parameters=parameters, + annotations=annotations, + **kwargs + ) + self.type: str = "SapCloudForCustomer" + self.url = url + self.username = username + self.password = password + self.encrypted_credential = encrypted_credential + + +class SapCloudForCustomerResourceDataset(Dataset): + """The path of the SAP Cloud for Customer OData entity. + + All required parameters must be populated in order to send to Azure. + + :ivar additional_properties: Unmatched properties from the message are deserialized to this + collection. + :vartype additional_properties: dict[str, JSON] + :ivar type: Type of dataset. Required. + :vartype type: str + :ivar description: Dataset description. + :vartype description: str + :ivar structure: Columns that define the structure of the dataset. Type: array (or Expression + with resultType array), itemType: DatasetDataElement. + :vartype structure: JSON + :ivar schema: Columns that define the physical type schema of the dataset. Type: array (or + Expression with resultType array), itemType: DatasetSchemaDataElement. + :vartype schema: JSON + :ivar linked_service_name: Linked service reference. Required. + :vartype linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference + :ivar parameters: Parameters for dataset. + :vartype parameters: dict[str, ~azure.mgmt.datafactory.models.ParameterSpecification] + :ivar annotations: List of tags that can be used for describing the Dataset. + :vartype annotations: list[JSON] + :ivar folder: The folder that this Dataset is in. If not specified, Dataset will appear at the + root level. + :vartype folder: ~azure.mgmt.datafactory.models.DatasetFolder + :ivar path: The path of the SAP Cloud for Customer OData entity. Type: string (or Expression + with resultType string). Required. + :vartype path: JSON + """ + + _validation = { + "type": {"required": True}, + "linked_service_name": {"required": True}, + "path": {"required": True}, + } + + _attribute_map = { + "additional_properties": {"key": "", "type": "{object}"}, + "type": {"key": "type", "type": "str"}, + "description": {"key": "description", "type": "str"}, + "structure": {"key": "structure", "type": "object"}, + "schema": {"key": "schema", "type": "object"}, + "linked_service_name": {"key": "linkedServiceName", "type": "LinkedServiceReference"}, + "parameters": {"key": "parameters", "type": "{ParameterSpecification}"}, + "annotations": {"key": "annotations", "type": "[object]"}, + "folder": {"key": "folder", "type": "DatasetFolder"}, + "path": {"key": "typeProperties.path", "type": "object"}, + } + + def __init__( + self, + *, + linked_service_name: "_models.LinkedServiceReference", + path: JSON, + additional_properties: Optional[Dict[str, JSON]] = None, + description: Optional[str] = None, + structure: Optional[JSON] = None, + schema: Optional[JSON] = None, + parameters: Optional[Dict[str, "_models.ParameterSpecification"]] = None, + annotations: Optional[List[JSON]] = None, + folder: Optional["_models.DatasetFolder"] = None, + **kwargs: Any + ) -> None: + """ + :keyword additional_properties: Unmatched properties from the message are deserialized to this + collection. + :paramtype additional_properties: dict[str, JSON] + :keyword description: Dataset description. + :paramtype description: str + :keyword structure: Columns that define the structure of the dataset. Type: array (or + Expression with resultType array), itemType: DatasetDataElement. + :paramtype structure: JSON + :keyword schema: Columns that define the physical type schema of the dataset. Type: array (or + Expression with resultType array), itemType: DatasetSchemaDataElement. + :paramtype schema: JSON + :keyword linked_service_name: Linked service reference. Required. + :paramtype linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference + :keyword parameters: Parameters for dataset. + :paramtype parameters: dict[str, ~azure.mgmt.datafactory.models.ParameterSpecification] + :keyword annotations: List of tags that can be used for describing the Dataset. + :paramtype annotations: list[JSON] + :keyword folder: The folder that this Dataset is in. If not specified, Dataset will appear at + the root level. + :paramtype folder: ~azure.mgmt.datafactory.models.DatasetFolder + :keyword path: The path of the SAP Cloud for Customer OData entity. Type: string (or Expression + with resultType string). Required. + :paramtype path: JSON + """ + super().__init__( + additional_properties=additional_properties, + description=description, + structure=structure, + schema=schema, + linked_service_name=linked_service_name, + parameters=parameters, + annotations=annotations, + folder=folder, + **kwargs + ) + self.type: str = "SapCloudForCustomerResource" + self.path = path + + +class SapCloudForCustomerSink(CopySink): + """A copy activity SAP Cloud for Customer sink. + + All required parameters must be populated in order to send to Azure. + + :ivar additional_properties: Unmatched properties from the message are deserialized to this + collection. + :vartype additional_properties: dict[str, JSON] + :ivar type: Copy sink type. Required. + :vartype type: str + :ivar write_batch_size: Write batch size. Type: integer (or Expression with resultType + integer), minimum: 0. + :vartype write_batch_size: JSON + :ivar write_batch_timeout: Write batch timeout. Type: string (or Expression with resultType + string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :vartype write_batch_timeout: JSON + :ivar sink_retry_count: Sink retry count. Type: integer (or Expression with resultType + integer). + :vartype sink_retry_count: JSON + :ivar sink_retry_wait: Sink retry wait. Type: string (or Expression with resultType string), + pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :vartype sink_retry_wait: JSON + :ivar max_concurrent_connections: The maximum concurrent connection count for the sink data + store. Type: integer (or Expression with resultType integer). + :vartype max_concurrent_connections: JSON + :ivar disable_metrics_collection: If true, disable data store metrics collection. Default is + false. Type: boolean (or Expression with resultType boolean). + :vartype disable_metrics_collection: JSON + :ivar write_behavior: The write behavior for the operation. Default is 'Insert'. Known values + are: "Insert" and "Update". + :vartype write_behavior: str or + ~azure.mgmt.datafactory.models.SapCloudForCustomerSinkWriteBehavior + :ivar http_request_timeout: The timeout (TimeSpan) to get an HTTP response. It is the timeout + to get a response, not the timeout to read response data. Default value: 00:05:00. Type: string + (or Expression with resultType string), pattern: + ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :vartype http_request_timeout: JSON + """ + + _validation = { + "type": {"required": True}, + } + + _attribute_map = { + "additional_properties": {"key": "", "type": "{object}"}, + "type": {"key": "type", "type": "str"}, + "write_batch_size": {"key": "writeBatchSize", "type": "object"}, + "write_batch_timeout": {"key": "writeBatchTimeout", "type": "object"}, + "sink_retry_count": {"key": "sinkRetryCount", "type": "object"}, + "sink_retry_wait": {"key": "sinkRetryWait", "type": "object"}, + "max_concurrent_connections": {"key": "maxConcurrentConnections", "type": "object"}, + "disable_metrics_collection": {"key": "disableMetricsCollection", "type": "object"}, + "write_behavior": {"key": "writeBehavior", "type": "str"}, + "http_request_timeout": {"key": "httpRequestTimeout", "type": "object"}, + } + + def __init__( + self, + *, + additional_properties: Optional[Dict[str, JSON]] = None, + write_batch_size: Optional[JSON] = None, + write_batch_timeout: Optional[JSON] = None, + sink_retry_count: Optional[JSON] = None, + sink_retry_wait: Optional[JSON] = None, + max_concurrent_connections: Optional[JSON] = None, + disable_metrics_collection: Optional[JSON] = None, + write_behavior: Optional[Union[str, "_models.SapCloudForCustomerSinkWriteBehavior"]] = None, + http_request_timeout: Optional[JSON] = None, + **kwargs: Any + ) -> None: + """ + :keyword additional_properties: Unmatched properties from the message are deserialized to this + collection. + :paramtype additional_properties: dict[str, JSON] + :keyword write_batch_size: Write batch size. Type: integer (or Expression with resultType + integer), minimum: 0. + :paramtype write_batch_size: JSON + :keyword write_batch_timeout: Write batch timeout. Type: string (or Expression with resultType + string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :paramtype write_batch_timeout: JSON + :keyword sink_retry_count: Sink retry count. Type: integer (or Expression with resultType + integer). + :paramtype sink_retry_count: JSON + :keyword sink_retry_wait: Sink retry wait. Type: string (or Expression with resultType string), + pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :paramtype sink_retry_wait: JSON + :keyword max_concurrent_connections: The maximum concurrent connection count for the sink data + store. Type: integer (or Expression with resultType integer). + :paramtype max_concurrent_connections: JSON + :keyword disable_metrics_collection: If true, disable data store metrics collection. Default is + false. Type: boolean (or Expression with resultType boolean). + :paramtype disable_metrics_collection: JSON + :keyword write_behavior: The write behavior for the operation. Default is 'Insert'. Known + values are: "Insert" and "Update". + :paramtype write_behavior: str or + ~azure.mgmt.datafactory.models.SapCloudForCustomerSinkWriteBehavior + :keyword http_request_timeout: The timeout (TimeSpan) to get an HTTP response. It is the + timeout to get a response, not the timeout to read response data. Default value: 00:05:00. + Type: string (or Expression with resultType string), pattern: + ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :paramtype http_request_timeout: JSON + """ + super().__init__( + additional_properties=additional_properties, + write_batch_size=write_batch_size, + write_batch_timeout=write_batch_timeout, + sink_retry_count=sink_retry_count, + sink_retry_wait=sink_retry_wait, + max_concurrent_connections=max_concurrent_connections, + disable_metrics_collection=disable_metrics_collection, + **kwargs + ) + self.type: str = "SapCloudForCustomerSink" + self.write_behavior = write_behavior + self.http_request_timeout = http_request_timeout + + +class SapCloudForCustomerSource(TabularSource): + """A copy activity source for SAP Cloud for Customer source. + + All required parameters must be populated in order to send to Azure. + + :ivar additional_properties: Unmatched properties from the message are deserialized to this + collection. + :vartype additional_properties: dict[str, JSON] + :ivar type: Copy source type. Required. + :vartype type: str + :ivar source_retry_count: Source retry count. Type: integer (or Expression with resultType + integer). + :vartype source_retry_count: JSON + :ivar source_retry_wait: Source retry wait. Type: string (or Expression with resultType + string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :vartype source_retry_wait: JSON + :ivar max_concurrent_connections: The maximum concurrent connection count for the source data + store. Type: integer (or Expression with resultType integer). + :vartype max_concurrent_connections: JSON + :ivar disable_metrics_collection: If true, disable data store metrics collection. Default is + false. Type: boolean (or Expression with resultType boolean). + :vartype disable_metrics_collection: JSON + :ivar query_timeout: Query timeout. Type: string (or Expression with resultType string), + pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :vartype query_timeout: JSON + :ivar additional_columns: Specifies the additional columns to be added to source data. Type: + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :vartype additional_columns: JSON + :ivar query: SAP Cloud for Customer OData query. For example, "$top=1". Type: string (or + Expression with resultType string). + :vartype query: JSON + :ivar http_request_timeout: The timeout (TimeSpan) to get an HTTP response. It is the timeout + to get a response, not the timeout to read response data. Default value: 00:05:00. Type: string + (or Expression with resultType string), pattern: + ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :vartype http_request_timeout: JSON + """ + + _validation = { + "type": {"required": True}, + } + + _attribute_map = { + "additional_properties": {"key": "", "type": "{object}"}, + "type": {"key": "type", "type": "str"}, + "source_retry_count": {"key": "sourceRetryCount", "type": "object"}, + "source_retry_wait": {"key": "sourceRetryWait", "type": "object"}, + "max_concurrent_connections": {"key": "maxConcurrentConnections", "type": "object"}, + "disable_metrics_collection": {"key": "disableMetricsCollection", "type": "object"}, + "query_timeout": {"key": "queryTimeout", "type": "object"}, + "additional_columns": {"key": "additionalColumns", "type": "object"}, + "query": {"key": "query", "type": "object"}, + "http_request_timeout": {"key": "httpRequestTimeout", "type": "object"}, + } + + def __init__( + self, + *, + additional_properties: Optional[Dict[str, JSON]] = None, + source_retry_count: Optional[JSON] = None, + source_retry_wait: Optional[JSON] = None, + max_concurrent_connections: Optional[JSON] = None, + disable_metrics_collection: Optional[JSON] = None, + query_timeout: Optional[JSON] = None, + additional_columns: Optional[JSON] = None, + query: Optional[JSON] = None, + http_request_timeout: Optional[JSON] = None, + **kwargs: Any + ) -> None: + """ + :keyword additional_properties: Unmatched properties from the message are deserialized to this + collection. + :paramtype additional_properties: dict[str, JSON] + :keyword source_retry_count: Source retry count. Type: integer (or Expression with resultType + integer). + :paramtype source_retry_count: JSON + :keyword source_retry_wait: Source retry wait. Type: string (or Expression with resultType + string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :paramtype source_retry_wait: JSON + :keyword max_concurrent_connections: The maximum concurrent connection count for the source + data store. Type: integer (or Expression with resultType integer). + :paramtype max_concurrent_connections: JSON + :keyword disable_metrics_collection: If true, disable data store metrics collection. Default is + false. Type: boolean (or Expression with resultType boolean). + :paramtype disable_metrics_collection: JSON + :keyword query_timeout: Query timeout. Type: string (or Expression with resultType string), + pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :paramtype query_timeout: JSON + :keyword additional_columns: Specifies the additional columns to be added to source data. Type: + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :paramtype additional_columns: JSON + :keyword query: SAP Cloud for Customer OData query. For example, "$top=1". Type: string (or + Expression with resultType string). + :paramtype query: JSON + :keyword http_request_timeout: The timeout (TimeSpan) to get an HTTP response. It is the + timeout to get a response, not the timeout to read response data. Default value: 00:05:00. + Type: string (or Expression with resultType string), pattern: + ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :paramtype http_request_timeout: JSON + """ + super().__init__( + additional_properties=additional_properties, + source_retry_count=source_retry_count, + source_retry_wait=source_retry_wait, + max_concurrent_connections=max_concurrent_connections, + disable_metrics_collection=disable_metrics_collection, + query_timeout=query_timeout, + additional_columns=additional_columns, + **kwargs + ) + self.type: str = "SapCloudForCustomerSource" + self.query = query + self.http_request_timeout = http_request_timeout + + +class SapEccLinkedService(LinkedService): + """Linked service for SAP ERP Central Component(SAP ECC). + + All required parameters must be populated in order to send to Azure. + + :ivar additional_properties: Unmatched properties from the message are deserialized to this + collection. + :vartype additional_properties: dict[str, JSON] + :ivar type: Type of linked service. Required. + :vartype type: str + :ivar connect_via: The integration runtime reference. + :vartype connect_via: ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :ivar description: Linked service description. + :vartype description: str + :ivar parameters: Parameters for linked service. + :vartype parameters: dict[str, ~azure.mgmt.datafactory.models.ParameterSpecification] + :ivar annotations: List of tags that can be used for describing the linked service. + :vartype annotations: list[JSON] + :ivar url: The URL of SAP ECC OData API. For example, + '[https://hostname:port/sap/opu/odata/sap/servicename/]'. Type: string (or Expression with + resultType string). Required. + :vartype url: JSON :ivar username: The username for Basic authentication. Type: string (or Expression with resultType string). :vartype username: JSON @@ -57056,6 +58181,9 @@ class SftpWriteSettings(StoreWriteSettings): :vartype disable_metrics_collection: JSON :ivar copy_behavior: The type of copy behavior for copy sink. :vartype copy_behavior: JSON + :ivar metadata: Specify the custom metadata to be added to sink data. Type: array of objects + (or Expression with resultType array of objects). + :vartype metadata: list[~azure.mgmt.datafactory.models.MetadataItem] :ivar operation_timeout: Specifies the timeout for writing each chunk to SFTP server. Default value: 01:00:00 (one hour). Type: string (or Expression with resultType string). :vartype operation_timeout: JSON @@ -57075,6 +58203,7 @@ class SftpWriteSettings(StoreWriteSettings): "max_concurrent_connections": {"key": "maxConcurrentConnections", "type": "object"}, "disable_metrics_collection": {"key": "disableMetricsCollection", "type": "object"}, "copy_behavior": {"key": "copyBehavior", "type": "object"}, + "metadata": {"key": "metadata", "type": "[MetadataItem]"}, "operation_timeout": {"key": "operationTimeout", "type": "object"}, "use_temp_file_rename": {"key": "useTempFileRename", "type": "object"}, } @@ -57086,6 +58215,7 @@ def __init__( max_concurrent_connections: Optional[JSON] = None, disable_metrics_collection: Optional[JSON] = None, copy_behavior: Optional[JSON] = None, + metadata: Optional[List["_models.MetadataItem"]] = None, operation_timeout: Optional[JSON] = None, use_temp_file_rename: Optional[JSON] = None, **kwargs: Any @@ -57102,6 +58232,9 @@ def __init__( :paramtype disable_metrics_collection: JSON :keyword copy_behavior: The type of copy behavior for copy sink. :paramtype copy_behavior: JSON + :keyword metadata: Specify the custom metadata to be added to sink data. Type: array of objects + (or Expression with resultType array of objects). + :paramtype metadata: list[~azure.mgmt.datafactory.models.MetadataItem] :keyword operation_timeout: Specifies the timeout for writing each chunk to SFTP server. Default value: 01:00:00 (one hour). Type: string (or Expression with resultType string). :paramtype operation_timeout: JSON @@ -57115,6 +58248,7 @@ def __init__( max_concurrent_connections=max_concurrent_connections, disable_metrics_collection=disable_metrics_collection, copy_behavior=copy_behavior, + metadata=metadata, **kwargs ) self.type: str = "SftpWriteSettings" @@ -58414,6 +59548,496 @@ def __init__( self.export_settings = export_settings +class SnowflakeV2Dataset(Dataset): # pylint: disable=too-many-instance-attributes + """The snowflake dataset. + + All required parameters must be populated in order to send to Azure. + + :ivar additional_properties: Unmatched properties from the message are deserialized to this + collection. + :vartype additional_properties: dict[str, JSON] + :ivar type: Type of dataset. Required. + :vartype type: str + :ivar description: Dataset description. + :vartype description: str + :ivar structure: Columns that define the structure of the dataset. Type: array (or Expression + with resultType array), itemType: DatasetDataElement. + :vartype structure: JSON + :ivar schema: Columns that define the physical type schema of the dataset. Type: array (or + Expression with resultType array), itemType: DatasetSchemaDataElement. + :vartype schema: JSON + :ivar linked_service_name: Linked service reference. Required. + :vartype linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference + :ivar parameters: Parameters for dataset. + :vartype parameters: dict[str, ~azure.mgmt.datafactory.models.ParameterSpecification] + :ivar annotations: List of tags that can be used for describing the Dataset. + :vartype annotations: list[JSON] + :ivar folder: The folder that this Dataset is in. If not specified, Dataset will appear at the + root level. + :vartype folder: ~azure.mgmt.datafactory.models.DatasetFolder + :ivar schema_type_properties_schema: The schema name of the Snowflake database. Type: string + (or Expression with resultType string). + :vartype schema_type_properties_schema: JSON + :ivar table: The table name of the Snowflake database. Type: string (or Expression with + resultType string). + :vartype table: JSON + """ + + _validation = { + "type": {"required": True}, + "linked_service_name": {"required": True}, + } + + _attribute_map = { + "additional_properties": {"key": "", "type": "{object}"}, + "type": {"key": "type", "type": "str"}, + "description": {"key": "description", "type": "str"}, + "structure": {"key": "structure", "type": "object"}, + "schema": {"key": "schema", "type": "object"}, + "linked_service_name": {"key": "linkedServiceName", "type": "LinkedServiceReference"}, + "parameters": {"key": "parameters", "type": "{ParameterSpecification}"}, + "annotations": {"key": "annotations", "type": "[object]"}, + "folder": {"key": "folder", "type": "DatasetFolder"}, + "schema_type_properties_schema": {"key": "typeProperties.schema", "type": "object"}, + "table": {"key": "typeProperties.table", "type": "object"}, + } + + def __init__( + self, + *, + linked_service_name: "_models.LinkedServiceReference", + additional_properties: Optional[Dict[str, JSON]] = None, + description: Optional[str] = None, + structure: Optional[JSON] = None, + schema: Optional[JSON] = None, + parameters: Optional[Dict[str, "_models.ParameterSpecification"]] = None, + annotations: Optional[List[JSON]] = None, + folder: Optional["_models.DatasetFolder"] = None, + schema_type_properties_schema: Optional[JSON] = None, + table: Optional[JSON] = None, + **kwargs: Any + ) -> None: + """ + :keyword additional_properties: Unmatched properties from the message are deserialized to this + collection. + :paramtype additional_properties: dict[str, JSON] + :keyword description: Dataset description. + :paramtype description: str + :keyword structure: Columns that define the structure of the dataset. Type: array (or + Expression with resultType array), itemType: DatasetDataElement. + :paramtype structure: JSON + :keyword schema: Columns that define the physical type schema of the dataset. Type: array (or + Expression with resultType array), itemType: DatasetSchemaDataElement. + :paramtype schema: JSON + :keyword linked_service_name: Linked service reference. Required. + :paramtype linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference + :keyword parameters: Parameters for dataset. + :paramtype parameters: dict[str, ~azure.mgmt.datafactory.models.ParameterSpecification] + :keyword annotations: List of tags that can be used for describing the Dataset. + :paramtype annotations: list[JSON] + :keyword folder: The folder that this Dataset is in. If not specified, Dataset will appear at + the root level. + :paramtype folder: ~azure.mgmt.datafactory.models.DatasetFolder + :keyword schema_type_properties_schema: The schema name of the Snowflake database. Type: string + (or Expression with resultType string). + :paramtype schema_type_properties_schema: JSON + :keyword table: The table name of the Snowflake database. Type: string (or Expression with + resultType string). + :paramtype table: JSON + """ + super().__init__( + additional_properties=additional_properties, + description=description, + structure=structure, + schema=schema, + linked_service_name=linked_service_name, + parameters=parameters, + annotations=annotations, + folder=folder, + **kwargs + ) + self.type: str = "SnowflakeV2Table" + self.schema_type_properties_schema = schema_type_properties_schema + self.table = table + + +class SnowflakeV2LinkedService(LinkedService): # pylint: disable=too-many-instance-attributes + """Snowflake linked service. + + All required parameters must be populated in order to send to Azure. + + :ivar additional_properties: Unmatched properties from the message are deserialized to this + collection. + :vartype additional_properties: dict[str, JSON] + :ivar type: Type of linked service. Required. + :vartype type: str + :ivar connect_via: The integration runtime reference. + :vartype connect_via: ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :ivar description: Linked service description. + :vartype description: str + :ivar parameters: Parameters for linked service. + :vartype parameters: dict[str, ~azure.mgmt.datafactory.models.ParameterSpecification] + :ivar annotations: List of tags that can be used for describing the linked service. + :vartype annotations: list[JSON] + :ivar account_identifier: The account identifier of your Snowflake account, e.g. + xy12345.east-us-2.azure. Required. + :vartype account_identifier: JSON + :ivar user: The name of the Snowflake user. + :vartype user: JSON + :ivar password: The Azure key vault secret reference of password in connection string. + :vartype password: ~azure.mgmt.datafactory.models.SecretBase + :ivar database: The name of the Snowflake database. Required. + :vartype database: JSON + :ivar warehouse: The name of the Snowflake warehouse. Required. + :vartype warehouse: JSON + :ivar authentication_type: The type used for authentication. Type: string. Known values are: + "Basic", "KeyPair", and "AADServicePrincipal". + :vartype authentication_type: str or ~azure.mgmt.datafactory.models.SnowflakeAuthenticationType + :ivar client_id: The client ID of the application registered in Azure Active Directory for + AADServicePrincipal authentication. + :vartype client_id: JSON + :ivar client_secret: The Azure key vault secret reference of client secret for + AADServicePrincipal authentication. + :vartype client_secret: ~azure.mgmt.datafactory.models.SecretBase + :ivar tenant_id: The tenant ID of the application registered in Azure Active Directory for + AADServicePrincipal authentication. + :vartype tenant_id: JSON + :ivar scope: The scope of the application registered in Azure Active Directory for + AADServicePrincipal authentication. + :vartype scope: JSON + :ivar private_key: The Azure key vault secret reference of privateKey for KeyPair auth. + :vartype private_key: ~azure.mgmt.datafactory.models.SecretBase + :ivar private_key_passphrase: The Azure key vault secret reference of private key password for + KeyPair auth with encrypted private key. + :vartype private_key_passphrase: ~azure.mgmt.datafactory.models.SecretBase + :ivar encrypted_credential: The encrypted credential used for authentication. Credentials are + encrypted using the integration runtime credential manager. Type: string. + :vartype encrypted_credential: str + """ + + _validation = { + "type": {"required": True}, + "account_identifier": {"required": True}, + "database": {"required": True}, + "warehouse": {"required": True}, + } + + _attribute_map = { + "additional_properties": {"key": "", "type": "{object}"}, + "type": {"key": "type", "type": "str"}, + "connect_via": {"key": "connectVia", "type": "IntegrationRuntimeReference"}, + "description": {"key": "description", "type": "str"}, + "parameters": {"key": "parameters", "type": "{ParameterSpecification}"}, + "annotations": {"key": "annotations", "type": "[object]"}, + "account_identifier": {"key": "typeProperties.accountIdentifier", "type": "object"}, + "user": {"key": "typeProperties.user", "type": "object"}, + "password": {"key": "typeProperties.password", "type": "SecretBase"}, + "database": {"key": "typeProperties.database", "type": "object"}, + "warehouse": {"key": "typeProperties.warehouse", "type": "object"}, + "authentication_type": {"key": "typeProperties.authenticationType", "type": "str"}, + "client_id": {"key": "typeProperties.clientId", "type": "object"}, + "client_secret": {"key": "typeProperties.clientSecret", "type": "SecretBase"}, + "tenant_id": {"key": "typeProperties.tenantId", "type": "object"}, + "scope": {"key": "typeProperties.scope", "type": "object"}, + "private_key": {"key": "typeProperties.privateKey", "type": "SecretBase"}, + "private_key_passphrase": {"key": "typeProperties.privateKeyPassphrase", "type": "SecretBase"}, + "encrypted_credential": {"key": "typeProperties.encryptedCredential", "type": "str"}, + } + + def __init__( + self, + *, + account_identifier: JSON, + database: JSON, + warehouse: JSON, + additional_properties: Optional[Dict[str, JSON]] = None, + connect_via: Optional["_models.IntegrationRuntimeReference"] = None, + description: Optional[str] = None, + parameters: Optional[Dict[str, "_models.ParameterSpecification"]] = None, + annotations: Optional[List[JSON]] = None, + user: Optional[JSON] = None, + password: Optional["_models.SecretBase"] = None, + authentication_type: Union[str, "_models.SnowflakeAuthenticationType"] = "Basic", + client_id: Optional[JSON] = None, + client_secret: Optional["_models.SecretBase"] = None, + tenant_id: Optional[JSON] = None, + scope: Optional[JSON] = None, + private_key: Optional["_models.SecretBase"] = None, + private_key_passphrase: Optional["_models.SecretBase"] = None, + encrypted_credential: Optional[str] = None, + **kwargs: Any + ) -> None: + """ + :keyword additional_properties: Unmatched properties from the message are deserialized to this + collection. + :paramtype additional_properties: dict[str, JSON] + :keyword connect_via: The integration runtime reference. + :paramtype connect_via: ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :keyword description: Linked service description. + :paramtype description: str + :keyword parameters: Parameters for linked service. + :paramtype parameters: dict[str, ~azure.mgmt.datafactory.models.ParameterSpecification] + :keyword annotations: List of tags that can be used for describing the linked service. + :paramtype annotations: list[JSON] + :keyword account_identifier: The account identifier of your Snowflake account, e.g. + xy12345.east-us-2.azure. Required. + :paramtype account_identifier: JSON + :keyword user: The name of the Snowflake user. + :paramtype user: JSON + :keyword password: The Azure key vault secret reference of password in connection string. + :paramtype password: ~azure.mgmt.datafactory.models.SecretBase + :keyword database: The name of the Snowflake database. Required. + :paramtype database: JSON + :keyword warehouse: The name of the Snowflake warehouse. Required. + :paramtype warehouse: JSON + :keyword authentication_type: The type used for authentication. Type: string. Known values are: + "Basic", "KeyPair", and "AADServicePrincipal". + :paramtype authentication_type: str or + ~azure.mgmt.datafactory.models.SnowflakeAuthenticationType + :keyword client_id: The client ID of the application registered in Azure Active Directory for + AADServicePrincipal authentication. + :paramtype client_id: JSON + :keyword client_secret: The Azure key vault secret reference of client secret for + AADServicePrincipal authentication. + :paramtype client_secret: ~azure.mgmt.datafactory.models.SecretBase + :keyword tenant_id: The tenant ID of the application registered in Azure Active Directory for + AADServicePrincipal authentication. + :paramtype tenant_id: JSON + :keyword scope: The scope of the application registered in Azure Active Directory for + AADServicePrincipal authentication. + :paramtype scope: JSON + :keyword private_key: The Azure key vault secret reference of privateKey for KeyPair auth. + :paramtype private_key: ~azure.mgmt.datafactory.models.SecretBase + :keyword private_key_passphrase: The Azure key vault secret reference of private key password + for KeyPair auth with encrypted private key. + :paramtype private_key_passphrase: ~azure.mgmt.datafactory.models.SecretBase + :keyword encrypted_credential: The encrypted credential used for authentication. Credentials + are encrypted using the integration runtime credential manager. Type: string. + :paramtype encrypted_credential: str + """ + super().__init__( + additional_properties=additional_properties, + connect_via=connect_via, + description=description, + parameters=parameters, + annotations=annotations, + **kwargs + ) + self.type: str = "SnowflakeV2" + self.account_identifier = account_identifier + self.user = user + self.password = password + self.database = database + self.warehouse = warehouse + self.authentication_type = authentication_type + self.client_id = client_id + self.client_secret = client_secret + self.tenant_id = tenant_id + self.scope = scope + self.private_key = private_key + self.private_key_passphrase = private_key_passphrase + self.encrypted_credential = encrypted_credential + + +class SnowflakeV2Sink(CopySink): + """A copy activity snowflake sink. + + All required parameters must be populated in order to send to Azure. + + :ivar additional_properties: Unmatched properties from the message are deserialized to this + collection. + :vartype additional_properties: dict[str, JSON] + :ivar type: Copy sink type. Required. + :vartype type: str + :ivar write_batch_size: Write batch size. Type: integer (or Expression with resultType + integer), minimum: 0. + :vartype write_batch_size: JSON + :ivar write_batch_timeout: Write batch timeout. Type: string (or Expression with resultType + string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :vartype write_batch_timeout: JSON + :ivar sink_retry_count: Sink retry count. Type: integer (or Expression with resultType + integer). + :vartype sink_retry_count: JSON + :ivar sink_retry_wait: Sink retry wait. Type: string (or Expression with resultType string), + pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :vartype sink_retry_wait: JSON + :ivar max_concurrent_connections: The maximum concurrent connection count for the sink data + store. Type: integer (or Expression with resultType integer). + :vartype max_concurrent_connections: JSON + :ivar disable_metrics_collection: If true, disable data store metrics collection. Default is + false. Type: boolean (or Expression with resultType boolean). + :vartype disable_metrics_collection: JSON + :ivar pre_copy_script: SQL pre-copy script. Type: string (or Expression with resultType + string). + :vartype pre_copy_script: JSON + :ivar import_settings: Snowflake import settings. + :vartype import_settings: ~azure.mgmt.datafactory.models.SnowflakeImportCopyCommand + """ + + _validation = { + "type": {"required": True}, + } + + _attribute_map = { + "additional_properties": {"key": "", "type": "{object}"}, + "type": {"key": "type", "type": "str"}, + "write_batch_size": {"key": "writeBatchSize", "type": "object"}, + "write_batch_timeout": {"key": "writeBatchTimeout", "type": "object"}, + "sink_retry_count": {"key": "sinkRetryCount", "type": "object"}, + "sink_retry_wait": {"key": "sinkRetryWait", "type": "object"}, + "max_concurrent_connections": {"key": "maxConcurrentConnections", "type": "object"}, + "disable_metrics_collection": {"key": "disableMetricsCollection", "type": "object"}, + "pre_copy_script": {"key": "preCopyScript", "type": "object"}, + "import_settings": {"key": "importSettings", "type": "SnowflakeImportCopyCommand"}, + } + + def __init__( + self, + *, + additional_properties: Optional[Dict[str, JSON]] = None, + write_batch_size: Optional[JSON] = None, + write_batch_timeout: Optional[JSON] = None, + sink_retry_count: Optional[JSON] = None, + sink_retry_wait: Optional[JSON] = None, + max_concurrent_connections: Optional[JSON] = None, + disable_metrics_collection: Optional[JSON] = None, + pre_copy_script: Optional[JSON] = None, + import_settings: Optional["_models.SnowflakeImportCopyCommand"] = None, + **kwargs: Any + ) -> None: + """ + :keyword additional_properties: Unmatched properties from the message are deserialized to this + collection. + :paramtype additional_properties: dict[str, JSON] + :keyword write_batch_size: Write batch size. Type: integer (or Expression with resultType + integer), minimum: 0. + :paramtype write_batch_size: JSON + :keyword write_batch_timeout: Write batch timeout. Type: string (or Expression with resultType + string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :paramtype write_batch_timeout: JSON + :keyword sink_retry_count: Sink retry count. Type: integer (or Expression with resultType + integer). + :paramtype sink_retry_count: JSON + :keyword sink_retry_wait: Sink retry wait. Type: string (or Expression with resultType string), + pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :paramtype sink_retry_wait: JSON + :keyword max_concurrent_connections: The maximum concurrent connection count for the sink data + store. Type: integer (or Expression with resultType integer). + :paramtype max_concurrent_connections: JSON + :keyword disable_metrics_collection: If true, disable data store metrics collection. Default is + false. Type: boolean (or Expression with resultType boolean). + :paramtype disable_metrics_collection: JSON + :keyword pre_copy_script: SQL pre-copy script. Type: string (or Expression with resultType + string). + :paramtype pre_copy_script: JSON + :keyword import_settings: Snowflake import settings. + :paramtype import_settings: ~azure.mgmt.datafactory.models.SnowflakeImportCopyCommand + """ + super().__init__( + additional_properties=additional_properties, + write_batch_size=write_batch_size, + write_batch_timeout=write_batch_timeout, + sink_retry_count=sink_retry_count, + sink_retry_wait=sink_retry_wait, + max_concurrent_connections=max_concurrent_connections, + disable_metrics_collection=disable_metrics_collection, + **kwargs + ) + self.type: str = "SnowflakeV2Sink" + self.pre_copy_script = pre_copy_script + self.import_settings = import_settings + + +class SnowflakeV2Source(CopySource): + """A copy activity snowflake source. + + All required parameters must be populated in order to send to Azure. + + :ivar additional_properties: Unmatched properties from the message are deserialized to this + collection. + :vartype additional_properties: dict[str, JSON] + :ivar type: Copy source type. Required. + :vartype type: str + :ivar source_retry_count: Source retry count. Type: integer (or Expression with resultType + integer). + :vartype source_retry_count: JSON + :ivar source_retry_wait: Source retry wait. Type: string (or Expression with resultType + string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :vartype source_retry_wait: JSON + :ivar max_concurrent_connections: The maximum concurrent connection count for the source data + store. Type: integer (or Expression with resultType integer). + :vartype max_concurrent_connections: JSON + :ivar disable_metrics_collection: If true, disable data store metrics collection. Default is + false. Type: boolean (or Expression with resultType boolean). + :vartype disable_metrics_collection: JSON + :ivar query: Snowflake Sql query. Type: string (or Expression with resultType string). + :vartype query: JSON + :ivar export_settings: Snowflake export settings. Required. + :vartype export_settings: ~azure.mgmt.datafactory.models.SnowflakeExportCopyCommand + """ + + _validation = { + "type": {"required": True}, + "export_settings": {"required": True}, + } + + _attribute_map = { + "additional_properties": {"key": "", "type": "{object}"}, + "type": {"key": "type", "type": "str"}, + "source_retry_count": {"key": "sourceRetryCount", "type": "object"}, + "source_retry_wait": {"key": "sourceRetryWait", "type": "object"}, + "max_concurrent_connections": {"key": "maxConcurrentConnections", "type": "object"}, + "disable_metrics_collection": {"key": "disableMetricsCollection", "type": "object"}, + "query": {"key": "query", "type": "object"}, + "export_settings": {"key": "exportSettings", "type": "SnowflakeExportCopyCommand"}, + } + + def __init__( + self, + *, + export_settings: "_models.SnowflakeExportCopyCommand", + additional_properties: Optional[Dict[str, JSON]] = None, + source_retry_count: Optional[JSON] = None, + source_retry_wait: Optional[JSON] = None, + max_concurrent_connections: Optional[JSON] = None, + disable_metrics_collection: Optional[JSON] = None, + query: Optional[JSON] = None, + **kwargs: Any + ) -> None: + """ + :keyword additional_properties: Unmatched properties from the message are deserialized to this + collection. + :paramtype additional_properties: dict[str, JSON] + :keyword source_retry_count: Source retry count. Type: integer (or Expression with resultType + integer). + :paramtype source_retry_count: JSON + :keyword source_retry_wait: Source retry wait. Type: string (or Expression with resultType + string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :paramtype source_retry_wait: JSON + :keyword max_concurrent_connections: The maximum concurrent connection count for the source + data store. Type: integer (or Expression with resultType integer). + :paramtype max_concurrent_connections: JSON + :keyword disable_metrics_collection: If true, disable data store metrics collection. Default is + false. Type: boolean (or Expression with resultType boolean). + :paramtype disable_metrics_collection: JSON + :keyword query: Snowflake Sql query. Type: string (or Expression with resultType string). + :paramtype query: JSON + :keyword export_settings: Snowflake export settings. Required. + :paramtype export_settings: ~azure.mgmt.datafactory.models.SnowflakeExportCopyCommand + """ + super().__init__( + additional_properties=additional_properties, + source_retry_count=source_retry_count, + source_retry_wait=source_retry_wait, + max_concurrent_connections=max_concurrent_connections, + disable_metrics_collection=disable_metrics_collection, + **kwargs + ) + self.type: str = "SnowflakeV2Source" + self.query = query + self.export_settings = export_settings + + class SparkConfigurationParametrizationReference(_serialization.Model): """Spark configuration reference. @@ -59145,7 +60769,8 @@ class SqlDWSource(TabularSource): # pylint: disable=too-many-instance-attribute is ReadCommitted. Type: string (or Expression with resultType string). :vartype isolation_level: JSON :ivar partition_option: The partition mechanism that will be used for Sql read in parallel. - Possible values include: "None", "PhysicalPartitionsOfTable", "DynamicRange". + Possible values include: "None", "PhysicalPartitionsOfTable", "DynamicRange". Type: string (or + Expression with resultType string). :vartype partition_option: JSON :ivar partition_settings: The settings that will be leveraged for Sql source partitioning. :vartype partition_settings: ~azure.mgmt.datafactory.models.SqlPartitionSettings @@ -59228,7 +60853,8 @@ def __init__( value is ReadCommitted. Type: string (or Expression with resultType string). :paramtype isolation_level: JSON :keyword partition_option: The partition mechanism that will be used for Sql read in parallel. - Possible values include: "None", "PhysicalPartitionsOfTable", "DynamicRange". + Possible values include: "None", "PhysicalPartitionsOfTable", "DynamicRange". Type: string (or + Expression with resultType string). :paramtype partition_option: JSON :keyword partition_settings: The settings that will be leveraged for Sql source partitioning. :paramtype partition_settings: ~azure.mgmt.datafactory.models.SqlPartitionSettings @@ -59332,8 +60958,8 @@ class SqlMISink(CopySink): # pylint: disable=too-many-instance-attributes :ivar sql_writer_use_table_lock: Whether to use table lock during bulk copy. Type: boolean (or Expression with resultType boolean). :vartype sql_writer_use_table_lock: JSON - :ivar write_behavior: White behavior when copying data into azure SQL MI. Type: - SqlWriteBehaviorEnum (or Expression with resultType SqlWriteBehaviorEnum). + :ivar write_behavior: White behavior when copying data into azure SQL MI. Type: string (or + Expression with resultType string). :vartype write_behavior: JSON :ivar upsert_settings: SQL upsert settings. :vartype upsert_settings: ~azure.mgmt.datafactory.models.SqlUpsertSettings @@ -59429,8 +61055,8 @@ def __init__( :keyword sql_writer_use_table_lock: Whether to use table lock during bulk copy. Type: boolean (or Expression with resultType boolean). :paramtype sql_writer_use_table_lock: JSON - :keyword write_behavior: White behavior when copying data into azure SQL MI. Type: - SqlWriteBehaviorEnum (or Expression with resultType SqlWriteBehaviorEnum). + :keyword write_behavior: White behavior when copying data into azure SQL MI. Type: string (or + Expression with resultType string). :paramtype write_behavior: JSON :keyword upsert_settings: SQL upsert settings. :paramtype upsert_settings: ~azure.mgmt.datafactory.models.SqlUpsertSettings @@ -59501,7 +61127,8 @@ class SqlMISource(TabularSource): # pylint: disable=too-many-instance-attribute :ivar produce_additional_types: Which additional types to produce. :vartype produce_additional_types: JSON :ivar partition_option: The partition mechanism that will be used for Sql read in parallel. - Possible values include: "None", "PhysicalPartitionsOfTable", "DynamicRange". + Possible values include: "None", "PhysicalPartitionsOfTable", "DynamicRange". Type: string (or + Expression with resultType string). :vartype partition_option: JSON :ivar partition_settings: The settings that will be leveraged for Sql source partitioning. :vartype partition_settings: ~azure.mgmt.datafactory.models.SqlPartitionSettings @@ -59587,7 +61214,8 @@ def __init__( :keyword produce_additional_types: Which additional types to produce. :paramtype produce_additional_types: JSON :keyword partition_option: The partition mechanism that will be used for Sql read in parallel. - Possible values include: "None", "PhysicalPartitionsOfTable", "DynamicRange". + Possible values include: "None", "PhysicalPartitionsOfTable", "DynamicRange". Type: string (or + Expression with resultType string). :paramtype partition_option: JSON :keyword partition_settings: The settings that will be leveraged for Sql source partitioning. :paramtype partition_settings: ~azure.mgmt.datafactory.models.SqlPartitionSettings @@ -59830,8 +61458,8 @@ class SqlServerSink(CopySink): # pylint: disable=too-many-instance-attributes :ivar sql_writer_use_table_lock: Whether to use table lock during bulk copy. Type: boolean (or Expression with resultType boolean). :vartype sql_writer_use_table_lock: JSON - :ivar write_behavior: Write behavior when copying data into sql server. Type: - SqlWriteBehaviorEnum (or Expression with resultType SqlWriteBehaviorEnum). + :ivar write_behavior: Write behavior when copying data into sql server. Type: string (or + Expression with resultType string). :vartype write_behavior: JSON :ivar upsert_settings: SQL upsert settings. :vartype upsert_settings: ~azure.mgmt.datafactory.models.SqlUpsertSettings @@ -59927,8 +61555,8 @@ def __init__( :keyword sql_writer_use_table_lock: Whether to use table lock during bulk copy. Type: boolean (or Expression with resultType boolean). :paramtype sql_writer_use_table_lock: JSON - :keyword write_behavior: Write behavior when copying data into sql server. Type: - SqlWriteBehaviorEnum (or Expression with resultType SqlWriteBehaviorEnum). + :keyword write_behavior: Write behavior when copying data into sql server. Type: string (or + Expression with resultType string). :paramtype write_behavior: JSON :keyword upsert_settings: SQL upsert settings. :paramtype upsert_settings: ~azure.mgmt.datafactory.models.SqlUpsertSettings @@ -59999,7 +61627,8 @@ class SqlServerSource(TabularSource): # pylint: disable=too-many-instance-attri :ivar produce_additional_types: Which additional types to produce. :vartype produce_additional_types: JSON :ivar partition_option: The partition mechanism that will be used for Sql read in parallel. - Possible values include: "None", "PhysicalPartitionsOfTable", "DynamicRange". + Possible values include: "None", "PhysicalPartitionsOfTable", "DynamicRange". Type: string (or + Expression with resultType string). :vartype partition_option: JSON :ivar partition_settings: The settings that will be leveraged for Sql source partitioning. :vartype partition_settings: ~azure.mgmt.datafactory.models.SqlPartitionSettings @@ -60085,7 +61714,8 @@ def __init__( :keyword produce_additional_types: Which additional types to produce. :paramtype produce_additional_types: JSON :keyword partition_option: The partition mechanism that will be used for Sql read in parallel. - Possible values include: "None", "PhysicalPartitionsOfTable", "DynamicRange". + Possible values include: "None", "PhysicalPartitionsOfTable", "DynamicRange". Type: string (or + Expression with resultType string). :paramtype partition_option: JSON :keyword partition_settings: The settings that will be leveraged for Sql source partitioning. :paramtype partition_settings: ~azure.mgmt.datafactory.models.SqlPartitionSettings @@ -60401,8 +62031,8 @@ class SqlSink(CopySink): # pylint: disable=too-many-instance-attributes :ivar sql_writer_use_table_lock: Whether to use table lock during bulk copy. Type: boolean (or Expression with resultType boolean). :vartype sql_writer_use_table_lock: JSON - :ivar write_behavior: Write behavior when copying data into sql. Type: SqlWriteBehaviorEnum (or - Expression with resultType SqlWriteBehaviorEnum). + :ivar write_behavior: Write behavior when copying data into sql. Type: string (or Expression + with resultType string). :vartype write_behavior: JSON :ivar upsert_settings: SQL upsert settings. :vartype upsert_settings: ~azure.mgmt.datafactory.models.SqlUpsertSettings @@ -60498,8 +62128,8 @@ def __init__( :keyword sql_writer_use_table_lock: Whether to use table lock during bulk copy. Type: boolean (or Expression with resultType boolean). :paramtype sql_writer_use_table_lock: JSON - :keyword write_behavior: Write behavior when copying data into sql. Type: SqlWriteBehaviorEnum - (or Expression with resultType SqlWriteBehaviorEnum). + :keyword write_behavior: Write behavior when copying data into sql. Type: string (or Expression + with resultType string). :paramtype write_behavior: JSON :keyword upsert_settings: SQL upsert settings. :paramtype upsert_settings: ~azure.mgmt.datafactory.models.SqlUpsertSettings @@ -60568,7 +62198,8 @@ class SqlSource(TabularSource): # pylint: disable=too-many-instance-attributes is ReadCommitted. Type: string (or Expression with resultType string). :vartype isolation_level: JSON :ivar partition_option: The partition mechanism that will be used for Sql read in parallel. - Possible values include: "None", "PhysicalPartitionsOfTable", "DynamicRange". + Possible values include: "None", "PhysicalPartitionsOfTable", "DynamicRange". Type: string (or + Expression with resultType string). :vartype partition_option: JSON :ivar partition_settings: The settings that will be leveraged for Sql source partitioning. :vartype partition_settings: ~azure.mgmt.datafactory.models.SqlPartitionSettings @@ -60650,7 +62281,8 @@ def __init__( value is ReadCommitted. Type: string (or Expression with resultType string). :paramtype isolation_level: JSON :keyword partition_option: The partition mechanism that will be used for Sql read in parallel. - Possible values include: "None", "PhysicalPartitionsOfTable", "DynamicRange". + Possible values include: "None", "PhysicalPartitionsOfTable", "DynamicRange". Type: string (or + Expression with resultType string). :paramtype partition_option: JSON :keyword partition_settings: The settings that will be leveraged for Sql source partitioning. :paramtype partition_settings: ~azure.mgmt.datafactory.models.SqlPartitionSettings @@ -65081,7 +66713,481 @@ class ValidationActivity(ControlActivity): # pylint: disable=too-many-instance- _validation = { "name": {"required": True}, "type": {"required": True}, - "dataset": {"required": True}, + "dataset": {"required": True}, + } + + _attribute_map = { + "additional_properties": {"key": "", "type": "{object}"}, + "name": {"key": "name", "type": "str"}, + "type": {"key": "type", "type": "str"}, + "description": {"key": "description", "type": "str"}, + "state": {"key": "state", "type": "str"}, + "on_inactive_mark_as": {"key": "onInactiveMarkAs", "type": "str"}, + "depends_on": {"key": "dependsOn", "type": "[ActivityDependency]"}, + "user_properties": {"key": "userProperties", "type": "[UserProperty]"}, + "timeout": {"key": "typeProperties.timeout", "type": "object"}, + "sleep": {"key": "typeProperties.sleep", "type": "object"}, + "minimum_size": {"key": "typeProperties.minimumSize", "type": "object"}, + "child_items": {"key": "typeProperties.childItems", "type": "object"}, + "dataset": {"key": "typeProperties.dataset", "type": "DatasetReference"}, + } + + def __init__( + self, + *, + name: str, + dataset: "_models.DatasetReference", + additional_properties: Optional[Dict[str, JSON]] = None, + description: Optional[str] = None, + state: Optional[Union[str, "_models.ActivityState"]] = None, + on_inactive_mark_as: Optional[Union[str, "_models.ActivityOnInactiveMarkAs"]] = None, + depends_on: Optional[List["_models.ActivityDependency"]] = None, + user_properties: Optional[List["_models.UserProperty"]] = None, + timeout: Optional[JSON] = None, + sleep: Optional[JSON] = None, + minimum_size: Optional[JSON] = None, + child_items: Optional[JSON] = None, + **kwargs: Any + ) -> None: + """ + :keyword additional_properties: Unmatched properties from the message are deserialized to this + collection. + :paramtype additional_properties: dict[str, JSON] + :keyword name: Activity name. Required. + :paramtype name: str + :keyword description: Activity description. + :paramtype description: str + :keyword state: Activity state. This is an optional property and if not provided, the state + will be Active by default. Known values are: "Active" and "Inactive". + :paramtype state: str or ~azure.mgmt.datafactory.models.ActivityState + :keyword on_inactive_mark_as: Status result of the activity when the state is set to Inactive. + This is an optional property and if not provided when the activity is inactive, the status will + be Succeeded by default. Known values are: "Succeeded", "Failed", and "Skipped". + :paramtype on_inactive_mark_as: str or ~azure.mgmt.datafactory.models.ActivityOnInactiveMarkAs + :keyword depends_on: Activity depends on condition. + :paramtype depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] + :keyword user_properties: Activity user properties. + :paramtype user_properties: list[~azure.mgmt.datafactory.models.UserProperty] + :keyword timeout: Specifies the timeout for the activity to run. If there is no value + specified, it takes the value of TimeSpan.FromDays(7) which is 1 week as default. Type: string + (or Expression with resultType string), pattern: + ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :paramtype timeout: JSON + :keyword sleep: A delay in seconds between validation attempts. If no value is specified, 10 + seconds will be used as the default. Type: integer (or Expression with resultType integer). + :paramtype sleep: JSON + :keyword minimum_size: Can be used if dataset points to a file. The file must be greater than + or equal in size to the value specified. Type: integer (or Expression with resultType integer). + :paramtype minimum_size: JSON + :keyword child_items: Can be used if dataset points to a folder. If set to true, the folder + must have at least one file. If set to false, the folder must be empty. Type: boolean (or + Expression with resultType boolean). + :paramtype child_items: JSON + :keyword dataset: Validation activity dataset reference. Required. + :paramtype dataset: ~azure.mgmt.datafactory.models.DatasetReference + """ + super().__init__( + additional_properties=additional_properties, + name=name, + description=description, + state=state, + on_inactive_mark_as=on_inactive_mark_as, + depends_on=depends_on, + user_properties=user_properties, + **kwargs + ) + self.type: str = "Validation" + self.timeout = timeout + self.sleep = sleep + self.minimum_size = minimum_size + self.child_items = child_items + self.dataset = dataset + + +class VariableSpecification(_serialization.Model): + """Definition of a single variable for a Pipeline. + + All required parameters must be populated in order to send to Azure. + + :ivar type: Variable type. Required. Known values are: "String", "Bool", and "Array". + :vartype type: str or ~azure.mgmt.datafactory.models.VariableType + :ivar default_value: Default value of variable. + :vartype default_value: JSON + """ + + _validation = { + "type": {"required": True}, + } + + _attribute_map = { + "type": {"key": "type", "type": "str"}, + "default_value": {"key": "defaultValue", "type": "object"}, + } + + def __init__( + self, *, type: Union[str, "_models.VariableType"], default_value: Optional[JSON] = None, **kwargs: Any + ) -> None: + """ + :keyword type: Variable type. Required. Known values are: "String", "Bool", and "Array". + :paramtype type: str or ~azure.mgmt.datafactory.models.VariableType + :keyword default_value: Default value of variable. + :paramtype default_value: JSON + """ + super().__init__(**kwargs) + self.type = type + self.default_value = default_value + + +class VerticaLinkedService(LinkedService): + """Vertica linked service. + + All required parameters must be populated in order to send to Azure. + + :ivar additional_properties: Unmatched properties from the message are deserialized to this + collection. + :vartype additional_properties: dict[str, JSON] + :ivar type: Type of linked service. Required. + :vartype type: str + :ivar connect_via: The integration runtime reference. + :vartype connect_via: ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :ivar description: Linked service description. + :vartype description: str + :ivar parameters: Parameters for linked service. + :vartype parameters: dict[str, ~azure.mgmt.datafactory.models.ParameterSpecification] + :ivar annotations: List of tags that can be used for describing the linked service. + :vartype annotations: list[JSON] + :ivar connection_string: An ODBC connection string. Type: string, SecureString or + AzureKeyVaultSecretReference. + :vartype connection_string: JSON + :ivar pwd: The Azure key vault secret reference of password in connection string. + :vartype pwd: ~azure.mgmt.datafactory.models.AzureKeyVaultSecretReference + :ivar encrypted_credential: The encrypted credential used for authentication. Credentials are + encrypted using the integration runtime credential manager. Type: string. + :vartype encrypted_credential: str + """ + + _validation = { + "type": {"required": True}, + } + + _attribute_map = { + "additional_properties": {"key": "", "type": "{object}"}, + "type": {"key": "type", "type": "str"}, + "connect_via": {"key": "connectVia", "type": "IntegrationRuntimeReference"}, + "description": {"key": "description", "type": "str"}, + "parameters": {"key": "parameters", "type": "{ParameterSpecification}"}, + "annotations": {"key": "annotations", "type": "[object]"}, + "connection_string": {"key": "typeProperties.connectionString", "type": "object"}, + "pwd": {"key": "typeProperties.pwd", "type": "AzureKeyVaultSecretReference"}, + "encrypted_credential": {"key": "typeProperties.encryptedCredential", "type": "str"}, + } + + def __init__( + self, + *, + additional_properties: Optional[Dict[str, JSON]] = None, + connect_via: Optional["_models.IntegrationRuntimeReference"] = None, + description: Optional[str] = None, + parameters: Optional[Dict[str, "_models.ParameterSpecification"]] = None, + annotations: Optional[List[JSON]] = None, + connection_string: Optional[JSON] = None, + pwd: Optional["_models.AzureKeyVaultSecretReference"] = None, + encrypted_credential: Optional[str] = None, + **kwargs: Any + ) -> None: + """ + :keyword additional_properties: Unmatched properties from the message are deserialized to this + collection. + :paramtype additional_properties: dict[str, JSON] + :keyword connect_via: The integration runtime reference. + :paramtype connect_via: ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :keyword description: Linked service description. + :paramtype description: str + :keyword parameters: Parameters for linked service. + :paramtype parameters: dict[str, ~azure.mgmt.datafactory.models.ParameterSpecification] + :keyword annotations: List of tags that can be used for describing the linked service. + :paramtype annotations: list[JSON] + :keyword connection_string: An ODBC connection string. Type: string, SecureString or + AzureKeyVaultSecretReference. + :paramtype connection_string: JSON + :keyword pwd: The Azure key vault secret reference of password in connection string. + :paramtype pwd: ~azure.mgmt.datafactory.models.AzureKeyVaultSecretReference + :keyword encrypted_credential: The encrypted credential used for authentication. Credentials + are encrypted using the integration runtime credential manager. Type: string. + :paramtype encrypted_credential: str + """ + super().__init__( + additional_properties=additional_properties, + connect_via=connect_via, + description=description, + parameters=parameters, + annotations=annotations, + **kwargs + ) + self.type: str = "Vertica" + self.connection_string = connection_string + self.pwd = pwd + self.encrypted_credential = encrypted_credential + + +class VerticaSource(TabularSource): + """A copy activity Vertica source. + + All required parameters must be populated in order to send to Azure. + + :ivar additional_properties: Unmatched properties from the message are deserialized to this + collection. + :vartype additional_properties: dict[str, JSON] + :ivar type: Copy source type. Required. + :vartype type: str + :ivar source_retry_count: Source retry count. Type: integer (or Expression with resultType + integer). + :vartype source_retry_count: JSON + :ivar source_retry_wait: Source retry wait. Type: string (or Expression with resultType + string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :vartype source_retry_wait: JSON + :ivar max_concurrent_connections: The maximum concurrent connection count for the source data + store. Type: integer (or Expression with resultType integer). + :vartype max_concurrent_connections: JSON + :ivar disable_metrics_collection: If true, disable data store metrics collection. Default is + false. Type: boolean (or Expression with resultType boolean). + :vartype disable_metrics_collection: JSON + :ivar query_timeout: Query timeout. Type: string (or Expression with resultType string), + pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :vartype query_timeout: JSON + :ivar additional_columns: Specifies the additional columns to be added to source data. Type: + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :vartype additional_columns: JSON + :ivar query: A query to retrieve data from source. Type: string (or Expression with resultType + string). + :vartype query: JSON + """ + + _validation = { + "type": {"required": True}, + } + + _attribute_map = { + "additional_properties": {"key": "", "type": "{object}"}, + "type": {"key": "type", "type": "str"}, + "source_retry_count": {"key": "sourceRetryCount", "type": "object"}, + "source_retry_wait": {"key": "sourceRetryWait", "type": "object"}, + "max_concurrent_connections": {"key": "maxConcurrentConnections", "type": "object"}, + "disable_metrics_collection": {"key": "disableMetricsCollection", "type": "object"}, + "query_timeout": {"key": "queryTimeout", "type": "object"}, + "additional_columns": {"key": "additionalColumns", "type": "object"}, + "query": {"key": "query", "type": "object"}, + } + + def __init__( + self, + *, + additional_properties: Optional[Dict[str, JSON]] = None, + source_retry_count: Optional[JSON] = None, + source_retry_wait: Optional[JSON] = None, + max_concurrent_connections: Optional[JSON] = None, + disable_metrics_collection: Optional[JSON] = None, + query_timeout: Optional[JSON] = None, + additional_columns: Optional[JSON] = None, + query: Optional[JSON] = None, + **kwargs: Any + ) -> None: + """ + :keyword additional_properties: Unmatched properties from the message are deserialized to this + collection. + :paramtype additional_properties: dict[str, JSON] + :keyword source_retry_count: Source retry count. Type: integer (or Expression with resultType + integer). + :paramtype source_retry_count: JSON + :keyword source_retry_wait: Source retry wait. Type: string (or Expression with resultType + string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :paramtype source_retry_wait: JSON + :keyword max_concurrent_connections: The maximum concurrent connection count for the source + data store. Type: integer (or Expression with resultType integer). + :paramtype max_concurrent_connections: JSON + :keyword disable_metrics_collection: If true, disable data store metrics collection. Default is + false. Type: boolean (or Expression with resultType boolean). + :paramtype disable_metrics_collection: JSON + :keyword query_timeout: Query timeout. Type: string (or Expression with resultType string), + pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :paramtype query_timeout: JSON + :keyword additional_columns: Specifies the additional columns to be added to source data. Type: + array of objects(AdditionalColumns) (or Expression with resultType array of objects). + :paramtype additional_columns: JSON + :keyword query: A query to retrieve data from source. Type: string (or Expression with + resultType string). + :paramtype query: JSON + """ + super().__init__( + additional_properties=additional_properties, + source_retry_count=source_retry_count, + source_retry_wait=source_retry_wait, + max_concurrent_connections=max_concurrent_connections, + disable_metrics_collection=disable_metrics_collection, + query_timeout=query_timeout, + additional_columns=additional_columns, + **kwargs + ) + self.type: str = "VerticaSource" + self.query = query + + +class VerticaTableDataset(Dataset): # pylint: disable=too-many-instance-attributes + """Vertica dataset. + + All required parameters must be populated in order to send to Azure. + + :ivar additional_properties: Unmatched properties from the message are deserialized to this + collection. + :vartype additional_properties: dict[str, JSON] + :ivar type: Type of dataset. Required. + :vartype type: str + :ivar description: Dataset description. + :vartype description: str + :ivar structure: Columns that define the structure of the dataset. Type: array (or Expression + with resultType array), itemType: DatasetDataElement. + :vartype structure: JSON + :ivar schema: Columns that define the physical type schema of the dataset. Type: array (or + Expression with resultType array), itemType: DatasetSchemaDataElement. + :vartype schema: JSON + :ivar linked_service_name: Linked service reference. Required. + :vartype linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference + :ivar parameters: Parameters for dataset. + :vartype parameters: dict[str, ~azure.mgmt.datafactory.models.ParameterSpecification] + :ivar annotations: List of tags that can be used for describing the Dataset. + :vartype annotations: list[JSON] + :ivar folder: The folder that this Dataset is in. If not specified, Dataset will appear at the + root level. + :vartype folder: ~azure.mgmt.datafactory.models.DatasetFolder + :ivar table_name: This property will be retired. Please consider using schema + table + properties instead. + :vartype table_name: JSON + :ivar table: The table name of the Vertica. Type: string (or Expression with resultType + string). + :vartype table: JSON + :ivar schema_type_properties_schema: The schema name of the Vertica. Type: string (or + Expression with resultType string). + :vartype schema_type_properties_schema: JSON + """ + + _validation = { + "type": {"required": True}, + "linked_service_name": {"required": True}, + } + + _attribute_map = { + "additional_properties": {"key": "", "type": "{object}"}, + "type": {"key": "type", "type": "str"}, + "description": {"key": "description", "type": "str"}, + "structure": {"key": "structure", "type": "object"}, + "schema": {"key": "schema", "type": "object"}, + "linked_service_name": {"key": "linkedServiceName", "type": "LinkedServiceReference"}, + "parameters": {"key": "parameters", "type": "{ParameterSpecification}"}, + "annotations": {"key": "annotations", "type": "[object]"}, + "folder": {"key": "folder", "type": "DatasetFolder"}, + "table_name": {"key": "typeProperties.tableName", "type": "object"}, + "table": {"key": "typeProperties.table", "type": "object"}, + "schema_type_properties_schema": {"key": "typeProperties.schema", "type": "object"}, + } + + def __init__( + self, + *, + linked_service_name: "_models.LinkedServiceReference", + additional_properties: Optional[Dict[str, JSON]] = None, + description: Optional[str] = None, + structure: Optional[JSON] = None, + schema: Optional[JSON] = None, + parameters: Optional[Dict[str, "_models.ParameterSpecification"]] = None, + annotations: Optional[List[JSON]] = None, + folder: Optional["_models.DatasetFolder"] = None, + table_name: Optional[JSON] = None, + table: Optional[JSON] = None, + schema_type_properties_schema: Optional[JSON] = None, + **kwargs: Any + ) -> None: + """ + :keyword additional_properties: Unmatched properties from the message are deserialized to this + collection. + :paramtype additional_properties: dict[str, JSON] + :keyword description: Dataset description. + :paramtype description: str + :keyword structure: Columns that define the structure of the dataset. Type: array (or + Expression with resultType array), itemType: DatasetDataElement. + :paramtype structure: JSON + :keyword schema: Columns that define the physical type schema of the dataset. Type: array (or + Expression with resultType array), itemType: DatasetSchemaDataElement. + :paramtype schema: JSON + :keyword linked_service_name: Linked service reference. Required. + :paramtype linked_service_name: ~azure.mgmt.datafactory.models.LinkedServiceReference + :keyword parameters: Parameters for dataset. + :paramtype parameters: dict[str, ~azure.mgmt.datafactory.models.ParameterSpecification] + :keyword annotations: List of tags that can be used for describing the Dataset. + :paramtype annotations: list[JSON] + :keyword folder: The folder that this Dataset is in. If not specified, Dataset will appear at + the root level. + :paramtype folder: ~azure.mgmt.datafactory.models.DatasetFolder + :keyword table_name: This property will be retired. Please consider using schema + table + properties instead. + :paramtype table_name: JSON + :keyword table: The table name of the Vertica. Type: string (or Expression with resultType + string). + :paramtype table: JSON + :keyword schema_type_properties_schema: The schema name of the Vertica. Type: string (or + Expression with resultType string). + :paramtype schema_type_properties_schema: JSON + """ + super().__init__( + additional_properties=additional_properties, + description=description, + structure=structure, + schema=schema, + linked_service_name=linked_service_name, + parameters=parameters, + annotations=annotations, + folder=folder, + **kwargs + ) + self.type: str = "VerticaTable" + self.table_name = table_name + self.table = table + self.schema_type_properties_schema = schema_type_properties_schema + + +class WaitActivity(ControlActivity): + """This activity suspends pipeline execution for the specified interval. + + All required parameters must be populated in order to send to Azure. + + :ivar additional_properties: Unmatched properties from the message are deserialized to this + collection. + :vartype additional_properties: dict[str, JSON] + :ivar name: Activity name. Required. + :vartype name: str + :ivar type: Type of activity. Required. + :vartype type: str + :ivar description: Activity description. + :vartype description: str + :ivar state: Activity state. This is an optional property and if not provided, the state will + be Active by default. Known values are: "Active" and "Inactive". + :vartype state: str or ~azure.mgmt.datafactory.models.ActivityState + :ivar on_inactive_mark_as: Status result of the activity when the state is set to Inactive. + This is an optional property and if not provided when the activity is inactive, the status will + be Succeeded by default. Known values are: "Succeeded", "Failed", and "Skipped". + :vartype on_inactive_mark_as: str or ~azure.mgmt.datafactory.models.ActivityOnInactiveMarkAs + :ivar depends_on: Activity depends on condition. + :vartype depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] + :ivar user_properties: Activity user properties. + :vartype user_properties: list[~azure.mgmt.datafactory.models.UserProperty] + :ivar wait_time_in_seconds: Duration in seconds. Type: integer (or Expression with resultType + integer). Required. + :vartype wait_time_in_seconds: JSON + """ + + _validation = { + "name": {"required": True}, + "type": {"required": True}, + "wait_time_in_seconds": {"required": True}, } _attribute_map = { @@ -65093,28 +67199,20 @@ class ValidationActivity(ControlActivity): # pylint: disable=too-many-instance- "on_inactive_mark_as": {"key": "onInactiveMarkAs", "type": "str"}, "depends_on": {"key": "dependsOn", "type": "[ActivityDependency]"}, "user_properties": {"key": "userProperties", "type": "[UserProperty]"}, - "timeout": {"key": "typeProperties.timeout", "type": "object"}, - "sleep": {"key": "typeProperties.sleep", "type": "object"}, - "minimum_size": {"key": "typeProperties.minimumSize", "type": "object"}, - "child_items": {"key": "typeProperties.childItems", "type": "object"}, - "dataset": {"key": "typeProperties.dataset", "type": "DatasetReference"}, + "wait_time_in_seconds": {"key": "typeProperties.waitTimeInSeconds", "type": "object"}, } def __init__( self, *, name: str, - dataset: "_models.DatasetReference", + wait_time_in_seconds: JSON, additional_properties: Optional[Dict[str, JSON]] = None, description: Optional[str] = None, state: Optional[Union[str, "_models.ActivityState"]] = None, on_inactive_mark_as: Optional[Union[str, "_models.ActivityOnInactiveMarkAs"]] = None, depends_on: Optional[List["_models.ActivityDependency"]] = None, user_properties: Optional[List["_models.UserProperty"]] = None, - timeout: Optional[JSON] = None, - sleep: Optional[JSON] = None, - minimum_size: Optional[JSON] = None, - child_items: Optional[JSON] = None, **kwargs: Any ) -> None: """ @@ -65136,23 +67234,9 @@ def __init__( :paramtype depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] :keyword user_properties: Activity user properties. :paramtype user_properties: list[~azure.mgmt.datafactory.models.UserProperty] - :keyword timeout: Specifies the timeout for the activity to run. If there is no value - specified, it takes the value of TimeSpan.FromDays(7) which is 1 week as default. Type: string - (or Expression with resultType string), pattern: - ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - :paramtype timeout: JSON - :keyword sleep: A delay in seconds between validation attempts. If no value is specified, 10 - seconds will be used as the default. Type: integer (or Expression with resultType integer). - :paramtype sleep: JSON - :keyword minimum_size: Can be used if dataset points to a file. The file must be greater than - or equal in size to the value specified. Type: integer (or Expression with resultType integer). - :paramtype minimum_size: JSON - :keyword child_items: Can be used if dataset points to a folder. If set to true, the folder - must have at least one file. If set to false, the folder must be empty. Type: boolean (or - Expression with resultType boolean). - :paramtype child_items: JSON - :keyword dataset: Validation activity dataset reference. Required. - :paramtype dataset: ~azure.mgmt.datafactory.models.DatasetReference + :keyword wait_time_in_seconds: Duration in seconds. Type: integer (or Expression with + resultType integer). Required. + :paramtype wait_time_in_seconds: JSON """ super().__init__( additional_properties=additional_properties, @@ -65164,50 +67248,12 @@ def __init__( user_properties=user_properties, **kwargs ) - self.type: str = "Validation" - self.timeout = timeout - self.sleep = sleep - self.minimum_size = minimum_size - self.child_items = child_items - self.dataset = dataset - - -class VariableSpecification(_serialization.Model): - """Definition of a single variable for a Pipeline. - - All required parameters must be populated in order to send to Azure. - - :ivar type: Variable type. Required. Known values are: "String", "Bool", and "Array". - :vartype type: str or ~azure.mgmt.datafactory.models.VariableType - :ivar default_value: Default value of variable. - :vartype default_value: JSON - """ - - _validation = { - "type": {"required": True}, - } - - _attribute_map = { - "type": {"key": "type", "type": "str"}, - "default_value": {"key": "defaultValue", "type": "object"}, - } - - def __init__( - self, *, type: Union[str, "_models.VariableType"], default_value: Optional[JSON] = None, **kwargs: Any - ) -> None: - """ - :keyword type: Variable type. Required. Known values are: "String", "Bool", and "Array". - :paramtype type: str or ~azure.mgmt.datafactory.models.VariableType - :keyword default_value: Default value of variable. - :paramtype default_value: JSON - """ - super().__init__(**kwargs) - self.type = type - self.default_value = default_value + self.type: str = "Wait" + self.wait_time_in_seconds = wait_time_in_seconds -class VerticaLinkedService(LinkedService): - """Vertica linked service. +class WarehouseLinkedService(LinkedService): # pylint: disable=too-many-instance-attributes + """Microsoft Fabric Warehouse linked service. All required parameters must be populated in order to send to Azure. @@ -65224,18 +67270,43 @@ class VerticaLinkedService(LinkedService): :vartype parameters: dict[str, ~azure.mgmt.datafactory.models.ParameterSpecification] :ivar annotations: List of tags that can be used for describing the linked service. :vartype annotations: list[JSON] - :ivar connection_string: An ODBC connection string. Type: string, SecureString or - AzureKeyVaultSecretReference. - :vartype connection_string: JSON - :ivar pwd: The Azure key vault secret reference of password in connection string. - :vartype pwd: ~azure.mgmt.datafactory.models.AzureKeyVaultSecretReference + :ivar artifact_id: The ID of Microsoft Fabric Warehouse artifact. Type: string (or Expression + with resultType string). Required. + :vartype artifact_id: JSON + :ivar endpoint: The endpoint of Microsoft Fabric Warehouse server. Type: string (or Expression + with resultType string). Required. + :vartype endpoint: JSON + :ivar workspace_id: The ID of Microsoft Fabric workspace. Type: string (or Expression with + resultType string). + :vartype workspace_id: JSON + :ivar service_principal_id: The ID of the application used to authenticate against Microsoft + Fabric Warehouse. Type: string (or Expression with resultType string). + :vartype service_principal_id: JSON + :ivar service_principal_key: The Key of the application used to authenticate against Microsoft + Fabric Warehouse. + :vartype service_principal_key: ~azure.mgmt.datafactory.models.SecretBase + :ivar tenant: The name or ID of the tenant to which the service principal belongs. Type: string + (or Expression with resultType string). + :vartype tenant: JSON :ivar encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string. :vartype encrypted_credential: str + :ivar service_principal_credential_type: The service principal credential type to use in + Server-To-Server authentication. 'ServicePrincipalKey' for key/secret, 'ServicePrincipalCert' + for certificate. Type: string (or Expression with resultType string). + :vartype service_principal_credential_type: JSON + :ivar service_principal_credential: The credential of the service principal object in Azure + Active Directory. If servicePrincipalCredentialType is 'ServicePrincipalKey', + servicePrincipalCredential can be SecureString or AzureKeyVaultSecretReference. If + servicePrincipalCredentialType is 'ServicePrincipalCert', servicePrincipalCredential can only + be AzureKeyVaultSecretReference. + :vartype service_principal_credential: ~azure.mgmt.datafactory.models.SecretBase """ _validation = { "type": {"required": True}, + "artifact_id": {"required": True}, + "endpoint": {"required": True}, } _attribute_map = { @@ -65245,22 +67316,34 @@ class VerticaLinkedService(LinkedService): "description": {"key": "description", "type": "str"}, "parameters": {"key": "parameters", "type": "{ParameterSpecification}"}, "annotations": {"key": "annotations", "type": "[object]"}, - "connection_string": {"key": "typeProperties.connectionString", "type": "object"}, - "pwd": {"key": "typeProperties.pwd", "type": "AzureKeyVaultSecretReference"}, + "artifact_id": {"key": "typeProperties.artifactId", "type": "object"}, + "endpoint": {"key": "typeProperties.endpoint", "type": "object"}, + "workspace_id": {"key": "typeProperties.workspaceId", "type": "object"}, + "service_principal_id": {"key": "typeProperties.servicePrincipalId", "type": "object"}, + "service_principal_key": {"key": "typeProperties.servicePrincipalKey", "type": "SecretBase"}, + "tenant": {"key": "typeProperties.tenant", "type": "object"}, "encrypted_credential": {"key": "typeProperties.encryptedCredential", "type": "str"}, + "service_principal_credential_type": {"key": "typeProperties.servicePrincipalCredentialType", "type": "object"}, + "service_principal_credential": {"key": "typeProperties.servicePrincipalCredential", "type": "SecretBase"}, } def __init__( self, *, + artifact_id: JSON, + endpoint: JSON, additional_properties: Optional[Dict[str, JSON]] = None, connect_via: Optional["_models.IntegrationRuntimeReference"] = None, description: Optional[str] = None, parameters: Optional[Dict[str, "_models.ParameterSpecification"]] = None, annotations: Optional[List[JSON]] = None, - connection_string: Optional[JSON] = None, - pwd: Optional["_models.AzureKeyVaultSecretReference"] = None, + workspace_id: Optional[JSON] = None, + service_principal_id: Optional[JSON] = None, + service_principal_key: Optional["_models.SecretBase"] = None, + tenant: Optional[JSON] = None, encrypted_credential: Optional[str] = None, + service_principal_credential_type: Optional[JSON] = None, + service_principal_credential: Optional["_models.SecretBase"] = None, **kwargs: Any ) -> None: """ @@ -65275,14 +67358,37 @@ def __init__( :paramtype parameters: dict[str, ~azure.mgmt.datafactory.models.ParameterSpecification] :keyword annotations: List of tags that can be used for describing the linked service. :paramtype annotations: list[JSON] - :keyword connection_string: An ODBC connection string. Type: string, SecureString or - AzureKeyVaultSecretReference. - :paramtype connection_string: JSON - :keyword pwd: The Azure key vault secret reference of password in connection string. - :paramtype pwd: ~azure.mgmt.datafactory.models.AzureKeyVaultSecretReference + :keyword artifact_id: The ID of Microsoft Fabric Warehouse artifact. Type: string (or + Expression with resultType string). Required. + :paramtype artifact_id: JSON + :keyword endpoint: The endpoint of Microsoft Fabric Warehouse server. Type: string (or + Expression with resultType string). Required. + :paramtype endpoint: JSON + :keyword workspace_id: The ID of Microsoft Fabric workspace. Type: string (or Expression with + resultType string). + :paramtype workspace_id: JSON + :keyword service_principal_id: The ID of the application used to authenticate against Microsoft + Fabric Warehouse. Type: string (or Expression with resultType string). + :paramtype service_principal_id: JSON + :keyword service_principal_key: The Key of the application used to authenticate against + Microsoft Fabric Warehouse. + :paramtype service_principal_key: ~azure.mgmt.datafactory.models.SecretBase + :keyword tenant: The name or ID of the tenant to which the service principal belongs. Type: + string (or Expression with resultType string). + :paramtype tenant: JSON :keyword encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string. :paramtype encrypted_credential: str + :keyword service_principal_credential_type: The service principal credential type to use in + Server-To-Server authentication. 'ServicePrincipalKey' for key/secret, 'ServicePrincipalCert' + for certificate. Type: string (or Expression with resultType string). + :paramtype service_principal_credential_type: JSON + :keyword service_principal_credential: The credential of the service principal object in Azure + Active Directory. If servicePrincipalCredentialType is 'ServicePrincipalKey', + servicePrincipalCredential can be SecureString or AzureKeyVaultSecretReference. If + servicePrincipalCredentialType is 'ServicePrincipalCert', servicePrincipalCredential can only + be AzureKeyVaultSecretReference. + :paramtype service_principal_credential: ~azure.mgmt.datafactory.models.SecretBase """ super().__init__( additional_properties=additional_properties, @@ -65292,14 +67398,158 @@ def __init__( annotations=annotations, **kwargs ) - self.type: str = "Vertica" - self.connection_string = connection_string - self.pwd = pwd + self.type: str = "Warehouse" + self.artifact_id = artifact_id + self.endpoint = endpoint + self.workspace_id = workspace_id + self.service_principal_id = service_principal_id + self.service_principal_key = service_principal_key + self.tenant = tenant self.encrypted_credential = encrypted_credential + self.service_principal_credential_type = service_principal_credential_type + self.service_principal_credential = service_principal_credential -class VerticaSource(TabularSource): - """A copy activity Vertica source. +class WarehouseSink(CopySink): # pylint: disable=too-many-instance-attributes + """A copy activity Microsoft Fabric Warehouse sink. + + All required parameters must be populated in order to send to Azure. + + :ivar additional_properties: Unmatched properties from the message are deserialized to this + collection. + :vartype additional_properties: dict[str, JSON] + :ivar type: Copy sink type. Required. + :vartype type: str + :ivar write_batch_size: Write batch size. Type: integer (or Expression with resultType + integer), minimum: 0. + :vartype write_batch_size: JSON + :ivar write_batch_timeout: Write batch timeout. Type: string (or Expression with resultType + string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :vartype write_batch_timeout: JSON + :ivar sink_retry_count: Sink retry count. Type: integer (or Expression with resultType + integer). + :vartype sink_retry_count: JSON + :ivar sink_retry_wait: Sink retry wait. Type: string (or Expression with resultType string), + pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :vartype sink_retry_wait: JSON + :ivar max_concurrent_connections: The maximum concurrent connection count for the sink data + store. Type: integer (or Expression with resultType integer). + :vartype max_concurrent_connections: JSON + :ivar disable_metrics_collection: If true, disable data store metrics collection. Default is + false. Type: boolean (or Expression with resultType boolean). + :vartype disable_metrics_collection: JSON + :ivar pre_copy_script: SQL pre-copy script. Type: string (or Expression with resultType + string). + :vartype pre_copy_script: JSON + :ivar allow_copy_command: Indicates to use Copy Command to copy data into SQL Data Warehouse. + Type: boolean (or Expression with resultType boolean). + :vartype allow_copy_command: JSON + :ivar copy_command_settings: Specifies Copy Command related settings when allowCopyCommand is + true. + :vartype copy_command_settings: ~azure.mgmt.datafactory.models.DWCopyCommandSettings + :ivar table_option: The option to handle sink table, such as autoCreate. For now only + 'autoCreate' value is supported. Type: string (or Expression with resultType string). + :vartype table_option: JSON + :ivar write_behavior: Write behavior when copying data into azure Microsoft Fabric Data + Warehouse. Type: DWWriteBehaviorEnum (or Expression with resultType DWWriteBehaviorEnum). + :vartype write_behavior: JSON + """ + + _validation = { + "type": {"required": True}, + } + + _attribute_map = { + "additional_properties": {"key": "", "type": "{object}"}, + "type": {"key": "type", "type": "str"}, + "write_batch_size": {"key": "writeBatchSize", "type": "object"}, + "write_batch_timeout": {"key": "writeBatchTimeout", "type": "object"}, + "sink_retry_count": {"key": "sinkRetryCount", "type": "object"}, + "sink_retry_wait": {"key": "sinkRetryWait", "type": "object"}, + "max_concurrent_connections": {"key": "maxConcurrentConnections", "type": "object"}, + "disable_metrics_collection": {"key": "disableMetricsCollection", "type": "object"}, + "pre_copy_script": {"key": "preCopyScript", "type": "object"}, + "allow_copy_command": {"key": "allowCopyCommand", "type": "object"}, + "copy_command_settings": {"key": "copyCommandSettings", "type": "DWCopyCommandSettings"}, + "table_option": {"key": "tableOption", "type": "object"}, + "write_behavior": {"key": "writeBehavior", "type": "object"}, + } + + def __init__( + self, + *, + additional_properties: Optional[Dict[str, JSON]] = None, + write_batch_size: Optional[JSON] = None, + write_batch_timeout: Optional[JSON] = None, + sink_retry_count: Optional[JSON] = None, + sink_retry_wait: Optional[JSON] = None, + max_concurrent_connections: Optional[JSON] = None, + disable_metrics_collection: Optional[JSON] = None, + pre_copy_script: Optional[JSON] = None, + allow_copy_command: Optional[JSON] = None, + copy_command_settings: Optional["_models.DWCopyCommandSettings"] = None, + table_option: Optional[JSON] = None, + write_behavior: Optional[JSON] = None, + **kwargs: Any + ) -> None: + """ + :keyword additional_properties: Unmatched properties from the message are deserialized to this + collection. + :paramtype additional_properties: dict[str, JSON] + :keyword write_batch_size: Write batch size. Type: integer (or Expression with resultType + integer), minimum: 0. + :paramtype write_batch_size: JSON + :keyword write_batch_timeout: Write batch timeout. Type: string (or Expression with resultType + string), pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :paramtype write_batch_timeout: JSON + :keyword sink_retry_count: Sink retry count. Type: integer (or Expression with resultType + integer). + :paramtype sink_retry_count: JSON + :keyword sink_retry_wait: Sink retry wait. Type: string (or Expression with resultType string), + pattern: ((\d+).)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :paramtype sink_retry_wait: JSON + :keyword max_concurrent_connections: The maximum concurrent connection count for the sink data + store. Type: integer (or Expression with resultType integer). + :paramtype max_concurrent_connections: JSON + :keyword disable_metrics_collection: If true, disable data store metrics collection. Default is + false. Type: boolean (or Expression with resultType boolean). + :paramtype disable_metrics_collection: JSON + :keyword pre_copy_script: SQL pre-copy script. Type: string (or Expression with resultType + string). + :paramtype pre_copy_script: JSON + :keyword allow_copy_command: Indicates to use Copy Command to copy data into SQL Data + Warehouse. Type: boolean (or Expression with resultType boolean). + :paramtype allow_copy_command: JSON + :keyword copy_command_settings: Specifies Copy Command related settings when allowCopyCommand + is true. + :paramtype copy_command_settings: ~azure.mgmt.datafactory.models.DWCopyCommandSettings + :keyword table_option: The option to handle sink table, such as autoCreate. For now only + 'autoCreate' value is supported. Type: string (or Expression with resultType string). + :paramtype table_option: JSON + :keyword write_behavior: Write behavior when copying data into azure Microsoft Fabric Data + Warehouse. Type: DWWriteBehaviorEnum (or Expression with resultType DWWriteBehaviorEnum). + :paramtype write_behavior: JSON + """ + super().__init__( + additional_properties=additional_properties, + write_batch_size=write_batch_size, + write_batch_timeout=write_batch_timeout, + sink_retry_count=sink_retry_count, + sink_retry_wait=sink_retry_wait, + max_concurrent_connections=max_concurrent_connections, + disable_metrics_collection=disable_metrics_collection, + **kwargs + ) + self.type: str = "WarehouseSink" + self.pre_copy_script = pre_copy_script + self.allow_copy_command = allow_copy_command + self.copy_command_settings = copy_command_settings + self.table_option = table_option + self.write_behavior = write_behavior + + +class WarehouseSource(TabularSource): # pylint: disable=too-many-instance-attributes + """A copy activity Microsoft Fabric Warehouse source. All required parameters must be populated in order to send to Azure. @@ -65326,9 +67576,27 @@ class VerticaSource(TabularSource): :ivar additional_columns: Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). :vartype additional_columns: JSON - :ivar query: A query to retrieve data from source. Type: string (or Expression with resultType - string). - :vartype query: JSON + :ivar sql_reader_query: Microsoft Fabric Warehouse reader query. Type: string (or Expression + with resultType string). + :vartype sql_reader_query: JSON + :ivar sql_reader_stored_procedure_name: Name of the stored procedure for a Microsoft Fabric + Warehouse source. This cannot be used at the same time as SqlReaderQuery. Type: string (or + Expression with resultType string). + :vartype sql_reader_stored_procedure_name: JSON + :ivar stored_procedure_parameters: Value and type setting for stored procedure parameters. + Example: "{Parameter1: {value: "1", type: "int"}}". Type: object (or Expression with resultType + object), itemType: StoredProcedureParameter. + :vartype stored_procedure_parameters: JSON + :ivar isolation_level: Specifies the transaction locking behavior for the Microsoft Fabric + Warehouse source. Allowed values: + ReadCommitted/ReadUncommitted/RepeatableRead/Serializable/Snapshot. The default value is + ReadCommitted. Type: string (or Expression with resultType string). + :vartype isolation_level: JSON + :ivar partition_option: The partition mechanism that will be used for Sql read in parallel. + Possible values include: "None", "PhysicalPartitionsOfTable", "DynamicRange". + :vartype partition_option: JSON + :ivar partition_settings: The settings that will be leveraged for Sql source partitioning. + :vartype partition_settings: ~azure.mgmt.datafactory.models.SqlPartitionSettings """ _validation = { @@ -65344,7 +67612,12 @@ class VerticaSource(TabularSource): "disable_metrics_collection": {"key": "disableMetricsCollection", "type": "object"}, "query_timeout": {"key": "queryTimeout", "type": "object"}, "additional_columns": {"key": "additionalColumns", "type": "object"}, - "query": {"key": "query", "type": "object"}, + "sql_reader_query": {"key": "sqlReaderQuery", "type": "object"}, + "sql_reader_stored_procedure_name": {"key": "sqlReaderStoredProcedureName", "type": "object"}, + "stored_procedure_parameters": {"key": "storedProcedureParameters", "type": "object"}, + "isolation_level": {"key": "isolationLevel", "type": "object"}, + "partition_option": {"key": "partitionOption", "type": "object"}, + "partition_settings": {"key": "partitionSettings", "type": "SqlPartitionSettings"}, } def __init__( @@ -65357,7 +67630,12 @@ def __init__( disable_metrics_collection: Optional[JSON] = None, query_timeout: Optional[JSON] = None, additional_columns: Optional[JSON] = None, - query: Optional[JSON] = None, + sql_reader_query: Optional[JSON] = None, + sql_reader_stored_procedure_name: Optional[JSON] = None, + stored_procedure_parameters: Optional[JSON] = None, + isolation_level: Optional[JSON] = None, + partition_option: Optional[JSON] = None, + partition_settings: Optional["_models.SqlPartitionSettings"] = None, **kwargs: Any ) -> None: """ @@ -65382,9 +67660,27 @@ def __init__( :keyword additional_columns: Specifies the additional columns to be added to source data. Type: array of objects(AdditionalColumns) (or Expression with resultType array of objects). :paramtype additional_columns: JSON - :keyword query: A query to retrieve data from source. Type: string (or Expression with - resultType string). - :paramtype query: JSON + :keyword sql_reader_query: Microsoft Fabric Warehouse reader query. Type: string (or Expression + with resultType string). + :paramtype sql_reader_query: JSON + :keyword sql_reader_stored_procedure_name: Name of the stored procedure for a Microsoft Fabric + Warehouse source. This cannot be used at the same time as SqlReaderQuery. Type: string (or + Expression with resultType string). + :paramtype sql_reader_stored_procedure_name: JSON + :keyword stored_procedure_parameters: Value and type setting for stored procedure parameters. + Example: "{Parameter1: {value: "1", type: "int"}}". Type: object (or Expression with resultType + object), itemType: StoredProcedureParameter. + :paramtype stored_procedure_parameters: JSON + :keyword isolation_level: Specifies the transaction locking behavior for the Microsoft Fabric + Warehouse source. Allowed values: + ReadCommitted/ReadUncommitted/RepeatableRead/Serializable/Snapshot. The default value is + ReadCommitted. Type: string (or Expression with resultType string). + :paramtype isolation_level: JSON + :keyword partition_option: The partition mechanism that will be used for Sql read in parallel. + Possible values include: "None", "PhysicalPartitionsOfTable", "DynamicRange". + :paramtype partition_option: JSON + :keyword partition_settings: The settings that will be leveraged for Sql source partitioning. + :paramtype partition_settings: ~azure.mgmt.datafactory.models.SqlPartitionSettings """ super().__init__( additional_properties=additional_properties, @@ -65396,12 +67692,17 @@ def __init__( additional_columns=additional_columns, **kwargs ) - self.type: str = "VerticaSource" - self.query = query + self.type: str = "WarehouseSource" + self.sql_reader_query = sql_reader_query + self.sql_reader_stored_procedure_name = sql_reader_stored_procedure_name + self.stored_procedure_parameters = stored_procedure_parameters + self.isolation_level = isolation_level + self.partition_option = partition_option + self.partition_settings = partition_settings -class VerticaTableDataset(Dataset): # pylint: disable=too-many-instance-attributes - """Vertica dataset. +class WarehouseTableDataset(Dataset): # pylint: disable=too-many-instance-attributes + """Microsoft Fabric Warehouse dataset. All required parameters must be populated in order to send to Azure. @@ -65427,15 +67728,12 @@ class VerticaTableDataset(Dataset): # pylint: disable=too-many-instance-attribu :ivar folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. :vartype folder: ~azure.mgmt.datafactory.models.DatasetFolder - :ivar table_name: This property will be retired. Please consider using schema + table - properties instead. - :vartype table_name: JSON - :ivar table: The table name of the Vertica. Type: string (or Expression with resultType - string). - :vartype table: JSON - :ivar schema_type_properties_schema: The schema name of the Vertica. Type: string (or - Expression with resultType string). + :ivar schema_type_properties_schema: The schema name of the Microsoft Fabric Warehouse. Type: + string (or Expression with resultType string). :vartype schema_type_properties_schema: JSON + :ivar table: The table name of the Microsoft Fabric Warehouse. Type: string (or Expression with + resultType string). + :vartype table: JSON """ _validation = { @@ -65453,9 +67751,8 @@ class VerticaTableDataset(Dataset): # pylint: disable=too-many-instance-attribu "parameters": {"key": "parameters", "type": "{ParameterSpecification}"}, "annotations": {"key": "annotations", "type": "[object]"}, "folder": {"key": "folder", "type": "DatasetFolder"}, - "table_name": {"key": "typeProperties.tableName", "type": "object"}, - "table": {"key": "typeProperties.table", "type": "object"}, "schema_type_properties_schema": {"key": "typeProperties.schema", "type": "object"}, + "table": {"key": "typeProperties.table", "type": "object"}, } def __init__( @@ -65469,9 +67766,8 @@ def __init__( parameters: Optional[Dict[str, "_models.ParameterSpecification"]] = None, annotations: Optional[List[JSON]] = None, folder: Optional["_models.DatasetFolder"] = None, - table_name: Optional[JSON] = None, - table: Optional[JSON] = None, schema_type_properties_schema: Optional[JSON] = None, + table: Optional[JSON] = None, **kwargs: Any ) -> None: """ @@ -65495,15 +67791,12 @@ def __init__( :keyword folder: The folder that this Dataset is in. If not specified, Dataset will appear at the root level. :paramtype folder: ~azure.mgmt.datafactory.models.DatasetFolder - :keyword table_name: This property will be retired. Please consider using schema + table - properties instead. - :paramtype table_name: JSON - :keyword table: The table name of the Vertica. Type: string (or Expression with resultType - string). - :paramtype table: JSON - :keyword schema_type_properties_schema: The schema name of the Vertica. Type: string (or - Expression with resultType string). + :keyword schema_type_properties_schema: The schema name of the Microsoft Fabric Warehouse. + Type: string (or Expression with resultType string). :paramtype schema_type_properties_schema: JSON + :keyword table: The table name of the Microsoft Fabric Warehouse. Type: string (or Expression + with resultType string). + :paramtype table: JSON """ super().__init__( additional_properties=additional_properties, @@ -65516,108 +67809,9 @@ def __init__( folder=folder, **kwargs ) - self.type: str = "VerticaTable" - self.table_name = table_name - self.table = table + self.type: str = "WarehouseTable" self.schema_type_properties_schema = schema_type_properties_schema - - -class WaitActivity(ControlActivity): - """This activity suspends pipeline execution for the specified interval. - - All required parameters must be populated in order to send to Azure. - - :ivar additional_properties: Unmatched properties from the message are deserialized to this - collection. - :vartype additional_properties: dict[str, JSON] - :ivar name: Activity name. Required. - :vartype name: str - :ivar type: Type of activity. Required. - :vartype type: str - :ivar description: Activity description. - :vartype description: str - :ivar state: Activity state. This is an optional property and if not provided, the state will - be Active by default. Known values are: "Active" and "Inactive". - :vartype state: str or ~azure.mgmt.datafactory.models.ActivityState - :ivar on_inactive_mark_as: Status result of the activity when the state is set to Inactive. - This is an optional property and if not provided when the activity is inactive, the status will - be Succeeded by default. Known values are: "Succeeded", "Failed", and "Skipped". - :vartype on_inactive_mark_as: str or ~azure.mgmt.datafactory.models.ActivityOnInactiveMarkAs - :ivar depends_on: Activity depends on condition. - :vartype depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] - :ivar user_properties: Activity user properties. - :vartype user_properties: list[~azure.mgmt.datafactory.models.UserProperty] - :ivar wait_time_in_seconds: Duration in seconds. Type: integer (or Expression with resultType - integer). Required. - :vartype wait_time_in_seconds: JSON - """ - - _validation = { - "name": {"required": True}, - "type": {"required": True}, - "wait_time_in_seconds": {"required": True}, - } - - _attribute_map = { - "additional_properties": {"key": "", "type": "{object}"}, - "name": {"key": "name", "type": "str"}, - "type": {"key": "type", "type": "str"}, - "description": {"key": "description", "type": "str"}, - "state": {"key": "state", "type": "str"}, - "on_inactive_mark_as": {"key": "onInactiveMarkAs", "type": "str"}, - "depends_on": {"key": "dependsOn", "type": "[ActivityDependency]"}, - "user_properties": {"key": "userProperties", "type": "[UserProperty]"}, - "wait_time_in_seconds": {"key": "typeProperties.waitTimeInSeconds", "type": "object"}, - } - - def __init__( - self, - *, - name: str, - wait_time_in_seconds: JSON, - additional_properties: Optional[Dict[str, JSON]] = None, - description: Optional[str] = None, - state: Optional[Union[str, "_models.ActivityState"]] = None, - on_inactive_mark_as: Optional[Union[str, "_models.ActivityOnInactiveMarkAs"]] = None, - depends_on: Optional[List["_models.ActivityDependency"]] = None, - user_properties: Optional[List["_models.UserProperty"]] = None, - **kwargs: Any - ) -> None: - """ - :keyword additional_properties: Unmatched properties from the message are deserialized to this - collection. - :paramtype additional_properties: dict[str, JSON] - :keyword name: Activity name. Required. - :paramtype name: str - :keyword description: Activity description. - :paramtype description: str - :keyword state: Activity state. This is an optional property and if not provided, the state - will be Active by default. Known values are: "Active" and "Inactive". - :paramtype state: str or ~azure.mgmt.datafactory.models.ActivityState - :keyword on_inactive_mark_as: Status result of the activity when the state is set to Inactive. - This is an optional property and if not provided when the activity is inactive, the status will - be Succeeded by default. Known values are: "Succeeded", "Failed", and "Skipped". - :paramtype on_inactive_mark_as: str or ~azure.mgmt.datafactory.models.ActivityOnInactiveMarkAs - :keyword depends_on: Activity depends on condition. - :paramtype depends_on: list[~azure.mgmt.datafactory.models.ActivityDependency] - :keyword user_properties: Activity user properties. - :paramtype user_properties: list[~azure.mgmt.datafactory.models.UserProperty] - :keyword wait_time_in_seconds: Duration in seconds. Type: integer (or Expression with - resultType integer). Required. - :paramtype wait_time_in_seconds: JSON - """ - super().__init__( - additional_properties=additional_properties, - name=name, - description=description, - state=state, - on_inactive_mark_as=on_inactive_mark_as, - depends_on=depends_on, - user_properties=user_properties, - **kwargs - ) - self.type: str = "Wait" - self.wait_time_in_seconds = wait_time_in_seconds + self.table = table class WebActivity(ExecutionActivity): # pylint: disable=too-many-instance-attributes @@ -65657,7 +67851,7 @@ class WebActivity(ExecutionActivity): # pylint: disable=too-many-instance-attri :vartype url: JSON :ivar headers: Represents the headers that will be sent to the request. For example, to set the language and type on a request: "headers" : { "Accept-Language": "en-us", "Content-Type": - "application/json" }. Type: string (or Expression with resultType string). + "application/json" }. Type: dictionary (or Expression with resultType dictionary). :vartype headers: JSON :ivar body: Represents the payload that will be sent to the endpoint. Required for POST/PUT method, not allowed for GET method Type: string (or Expression with resultType string). @@ -65666,6 +67860,15 @@ class WebActivity(ExecutionActivity): # pylint: disable=too-many-instance-attri :vartype authentication: ~azure.mgmt.datafactory.models.WebActivityAuthentication :ivar disable_cert_validation: When set to true, Certificate validation will be disabled. :vartype disable_cert_validation: bool + :ivar http_request_timeout: Timeout for the HTTP request to get a response. Format is in + TimeSpan (hh:mm:ss). This value is the timeout to get a response, not the activity timeout. The + default value is 00:01:00 (1 minute). The range is from 1 to 10 minutes. + :vartype http_request_timeout: JSON + :ivar turn_off_async: Option to disable invoking HTTP GET on location given in response header + of a HTTP 202 Response. If set true, it stops invoking HTTP GET on http location given in + response header. If set false then continues to invoke HTTP GET call on location given in http + response headers. + :vartype turn_off_async: bool :ivar datasets: List of datasets passed to web endpoint. :vartype datasets: list[~azure.mgmt.datafactory.models.DatasetReference] :ivar linked_services: List of linked services passed to web endpoint. @@ -65698,6 +67901,8 @@ class WebActivity(ExecutionActivity): # pylint: disable=too-many-instance-attri "body": {"key": "typeProperties.body", "type": "object"}, "authentication": {"key": "typeProperties.authentication", "type": "WebActivityAuthentication"}, "disable_cert_validation": {"key": "typeProperties.disableCertValidation", "type": "bool"}, + "http_request_timeout": {"key": "typeProperties.httpRequestTimeout", "type": "object"}, + "turn_off_async": {"key": "typeProperties.turnOffAsync", "type": "bool"}, "datasets": {"key": "typeProperties.datasets", "type": "[DatasetReference]"}, "linked_services": {"key": "typeProperties.linkedServices", "type": "[LinkedServiceReference]"}, "connect_via": {"key": "typeProperties.connectVia", "type": "IntegrationRuntimeReference"}, @@ -65721,6 +67926,8 @@ def __init__( body: Optional[JSON] = None, authentication: Optional["_models.WebActivityAuthentication"] = None, disable_cert_validation: Optional[bool] = None, + http_request_timeout: Optional[JSON] = None, + turn_off_async: Optional[bool] = None, datasets: Optional[List["_models.DatasetReference"]] = None, linked_services: Optional[List["_models.LinkedServiceReference"]] = None, connect_via: Optional["_models.IntegrationRuntimeReference"] = None, @@ -65757,7 +67964,7 @@ def __init__( :paramtype url: JSON :keyword headers: Represents the headers that will be sent to the request. For example, to set the language and type on a request: "headers" : { "Accept-Language": "en-us", "Content-Type": - "application/json" }. Type: string (or Expression with resultType string). + "application/json" }. Type: dictionary (or Expression with resultType dictionary). :paramtype headers: JSON :keyword body: Represents the payload that will be sent to the endpoint. Required for POST/PUT method, not allowed for GET method Type: string (or Expression with resultType string). @@ -65766,6 +67973,15 @@ def __init__( :paramtype authentication: ~azure.mgmt.datafactory.models.WebActivityAuthentication :keyword disable_cert_validation: When set to true, Certificate validation will be disabled. :paramtype disable_cert_validation: bool + :keyword http_request_timeout: Timeout for the HTTP request to get a response. Format is in + TimeSpan (hh:mm:ss). This value is the timeout to get a response, not the activity timeout. The + default value is 00:01:00 (1 minute). The range is from 1 to 10 minutes. + :paramtype http_request_timeout: JSON + :keyword turn_off_async: Option to disable invoking HTTP GET on location given in response + header of a HTTP 202 Response. If set true, it stops invoking HTTP GET on http location given + in response header. If set false then continues to invoke HTTP GET call on location given in + http response headers. + :paramtype turn_off_async: bool :keyword datasets: List of datasets passed to web endpoint. :paramtype datasets: list[~azure.mgmt.datafactory.models.DatasetReference] :keyword linked_services: List of linked services passed to web endpoint. @@ -65792,6 +68008,8 @@ def __init__( self.body = body self.authentication = authentication self.disable_cert_validation = disable_cert_validation + self.http_request_timeout = http_request_timeout + self.turn_off_async = turn_off_async self.datasets = datasets self.linked_services = linked_services self.connect_via = connect_via @@ -66088,7 +68306,7 @@ class WebHookActivity(ControlActivity): # pylint: disable=too-many-instance-att :vartype timeout: str :ivar headers: Represents the headers that will be sent to the request. For example, to set the language and type on a request: "headers" : { "Accept-Language": "en-us", "Content-Type": - "application/json" }. Type: string (or Expression with resultType string). + "application/json" }. Type: dictionary (or Expression with resultType dictionary). :vartype headers: JSON :ivar body: Represents the payload that will be sent to the endpoint. Required for POST/PUT method, not allowed for GET method Type: string (or Expression with resultType string). @@ -66180,7 +68398,7 @@ def __init__( :paramtype timeout: str :keyword headers: Represents the headers that will be sent to the request. For example, to set the language and type on a request: "headers" : { "Accept-Language": "en-us", "Content-Type": - "application/json" }. Type: string (or Expression with resultType string). + "application/json" }. Type: dictionary (or Expression with resultType dictionary). :paramtype headers: JSON :keyword body: Represents the payload that will be sent to the endpoint. Required for POST/PUT method, not allowed for GET method Type: string (or Expression with resultType string).