From 523fb34310d85d35637f583c969e051899bf5bc7 Mon Sep 17 00:00:00 2001 From: SDKAuto Date: Wed, 11 May 2022 05:53:08 +0000 Subject: [PATCH] CodeGen from PR 18956 in Azure/azure-rest-api-specs Merge 3b478489713c1ee63461752bc987d8b2e1ad9f23 into ac286e05bedd69d301e1dc94ec70f66e8536686b --- .../azure-mgmt-datafactory/_meta.json | 4 +- .../azure/mgmt/datafactory/_version.py | 2 +- .../azure/mgmt/datafactory/models/__init__.py | 4 + .../mgmt/datafactory/models/_models_py3.py | 116 +++++++++++++++++- 4 files changed, 121 insertions(+), 5 deletions(-) diff --git a/sdk/datafactory/azure-mgmt-datafactory/_meta.json b/sdk/datafactory/azure-mgmt-datafactory/_meta.json index c3b35e851fab..c40234979ed4 100644 --- a/sdk/datafactory/azure-mgmt-datafactory/_meta.json +++ b/sdk/datafactory/azure-mgmt-datafactory/_meta.json @@ -4,8 +4,8 @@ "@autorest/python@5.13.0", "@autorest/modelerfour@4.19.3" ], - "commit": "256b8ec7d045dbe2daf91030b0d6b7f09c8e42d9", + "commit": "b6bd290cd8c34e33387bd0bcf6cdaf9b03fecddd", "repository_url": "https://github.com/Azure/azure-rest-api-specs", - "autorest_command": "autorest specification/datafactory/resource-manager/readme.md --multiapi --python --python-sdks-folder=/home/vsts/work/1/azure-sdk-for-python/sdk --python3-only --use=@autorest/python@5.13.0 --use=@autorest/modelerfour@4.19.3 --version=3.7.2", + "autorest_command": "autorest specification/datafactory/resource-manager/readme.md --multiapi --python --python-sdks-folder=/mnt/vss/_work/1/s/azure-sdk-for-python/sdk --python3-only --use=@autorest/python@5.13.0 --use=@autorest/modelerfour@4.19.3 --version=3.7.2", "readme": "specification/datafactory/resource-manager/readme.md" } \ No newline at end of file diff --git a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/_version.py b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/_version.py index 1d9186e2318f..c47f66669f1b 100644 --- a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/_version.py +++ b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/_version.py @@ -6,4 +6,4 @@ # Changes may cause incorrect behavior and will be lost if the code is regenerated. # -------------------------------------------------------------------------- -VERSION = "2.4.0" +VERSION = "1.0.0" diff --git a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py index 2ed5dafb742a..8ad758e292ed 100644 --- a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py +++ b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py @@ -500,6 +500,7 @@ from ._models_py3 import PrestoLinkedService from ._models_py3 import PrestoObjectDataset from ._models_py3 import PrestoSource +from ._models_py3 import PrivateEndpoint from ._models_py3 import PrivateEndpointConnectionListResponse from ._models_py3 import PrivateEndpointConnectionResource from ._models_py3 import PrivateLinkConnectionApprovalRequest @@ -508,6 +509,7 @@ from ._models_py3 import PrivateLinkResource from ._models_py3 import PrivateLinkResourceProperties from ._models_py3 import PrivateLinkResourcesWrapper +from ._models_py3 import PurviewConfiguration from ._models_py3 import QueryDataFlowDebugSessionsResponse from ._models_py3 import QuickBooksLinkedService from ._models_py3 import QuickBooksObjectDataset @@ -1314,6 +1316,7 @@ 'PrestoLinkedService', 'PrestoObjectDataset', 'PrestoSource', + 'PrivateEndpoint', 'PrivateEndpointConnectionListResponse', 'PrivateEndpointConnectionResource', 'PrivateLinkConnectionApprovalRequest', @@ -1322,6 +1325,7 @@ 'PrivateLinkResource', 'PrivateLinkResourceProperties', 'PrivateLinkResourcesWrapper', + 'PurviewConfiguration', 'QueryDataFlowDebugSessionsResponse', 'QuickBooksLinkedService', 'QuickBooksObjectDataset', diff --git a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/_models_py3.py b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/_models_py3.py index 1f3267a1c40f..722e0124d61c 100644 --- a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/_models_py3.py +++ b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/_models_py3.py @@ -17986,6 +17986,8 @@ class DataFlowSink(Transformation): :vartype flowlet: ~azure.mgmt.datafactory.models.DataFlowReference :ivar schema_linked_service: Schema linked service reference. :vartype schema_linked_service: ~azure.mgmt.datafactory.models.LinkedServiceReference + :ivar rejected_data_linked_service: Rejected data linked service reference. + :vartype rejected_data_linked_service: ~azure.mgmt.datafactory.models.LinkedServiceReference """ _validation = { @@ -17999,6 +18001,7 @@ class DataFlowSink(Transformation): 'linked_service': {'key': 'linkedService', 'type': 'LinkedServiceReference'}, 'flowlet': {'key': 'flowlet', 'type': 'DataFlowReference'}, 'schema_linked_service': {'key': 'schemaLinkedService', 'type': 'LinkedServiceReference'}, + 'rejected_data_linked_service': {'key': 'rejectedDataLinkedService', 'type': 'LinkedServiceReference'}, } def __init__( @@ -18010,6 +18013,7 @@ def __init__( linked_service: Optional["LinkedServiceReference"] = None, flowlet: Optional["DataFlowReference"] = None, schema_linked_service: Optional["LinkedServiceReference"] = None, + rejected_data_linked_service: Optional["LinkedServiceReference"] = None, **kwargs ): """ @@ -18025,9 +18029,12 @@ def __init__( :paramtype flowlet: ~azure.mgmt.datafactory.models.DataFlowReference :keyword schema_linked_service: Schema linked service reference. :paramtype schema_linked_service: ~azure.mgmt.datafactory.models.LinkedServiceReference + :keyword rejected_data_linked_service: Rejected data linked service reference. + :paramtype rejected_data_linked_service: ~azure.mgmt.datafactory.models.LinkedServiceReference """ super(DataFlowSink, self).__init__(name=name, description=description, dataset=dataset, linked_service=linked_service, flowlet=flowlet, **kwargs) self.schema_linked_service = schema_linked_service + self.rejected_data_linked_service = rejected_data_linked_service class DataFlowSource(Transformation): @@ -22453,6 +22460,9 @@ class ExecuteDataFlowActivity(ExecutionActivity): the same save order to be processed concurrently. Type: boolean (or Expression with resultType boolean). :vartype run_concurrently: any + :ivar source_staging_concurrency: Specify number of parallel staging for sources applicable to + the sink. Type: integer (or Expression with resultType integer). + :vartype source_staging_concurrency: any """ _validation = { @@ -22477,6 +22487,7 @@ class ExecuteDataFlowActivity(ExecutionActivity): 'trace_level': {'key': 'typeProperties.traceLevel', 'type': 'object'}, 'continue_on_error': {'key': 'typeProperties.continueOnError', 'type': 'object'}, 'run_concurrently': {'key': 'typeProperties.runConcurrently', 'type': 'object'}, + 'source_staging_concurrency': {'key': 'typeProperties.sourceStagingConcurrency', 'type': 'object'}, } def __init__( @@ -22496,6 +22507,7 @@ def __init__( trace_level: Optional[Any] = None, continue_on_error: Optional[Any] = None, run_concurrently: Optional[Any] = None, + source_staging_concurrency: Optional[Any] = None, **kwargs ): """ @@ -22532,6 +22544,9 @@ def __init__( with the same save order to be processed concurrently. Type: boolean (or Expression with resultType boolean). :paramtype run_concurrently: any + :keyword source_staging_concurrency: Specify number of parallel staging for sources applicable + to the sink. Type: integer (or Expression with resultType integer). + :paramtype source_staging_concurrency: any """ super(ExecuteDataFlowActivity, self).__init__(additional_properties=additional_properties, name=name, description=description, depends_on=depends_on, user_properties=user_properties, linked_service_name=linked_service_name, policy=policy, **kwargs) self.type = 'ExecuteDataFlow' # type: str @@ -22542,6 +22557,7 @@ def __init__( self.trace_level = trace_level self.continue_on_error = continue_on_error self.run_concurrently = run_concurrently + self.source_staging_concurrency = source_staging_concurrency class ExecuteDataFlowActivityTypeProperties(msrest.serialization.Model): @@ -22567,6 +22583,9 @@ class ExecuteDataFlowActivityTypeProperties(msrest.serialization.Model): the same save order to be processed concurrently. Type: boolean (or Expression with resultType boolean). :vartype run_concurrently: any + :ivar source_staging_concurrency: Specify number of parallel staging for sources applicable to + the sink. Type: integer (or Expression with resultType integer). + :vartype source_staging_concurrency: any """ _validation = { @@ -22581,6 +22600,7 @@ class ExecuteDataFlowActivityTypeProperties(msrest.serialization.Model): 'trace_level': {'key': 'traceLevel', 'type': 'object'}, 'continue_on_error': {'key': 'continueOnError', 'type': 'object'}, 'run_concurrently': {'key': 'runConcurrently', 'type': 'object'}, + 'source_staging_concurrency': {'key': 'sourceStagingConcurrency', 'type': 'object'}, } def __init__( @@ -22593,6 +22613,7 @@ def __init__( trace_level: Optional[Any] = None, continue_on_error: Optional[Any] = None, run_concurrently: Optional[Any] = None, + source_staging_concurrency: Optional[Any] = None, **kwargs ): """ @@ -22614,6 +22635,9 @@ def __init__( with the same save order to be processed concurrently. Type: boolean (or Expression with resultType boolean). :paramtype run_concurrently: any + :keyword source_staging_concurrency: Specify number of parallel staging for sources applicable + to the sink. Type: integer (or Expression with resultType integer). + :paramtype source_staging_concurrency: any """ super(ExecuteDataFlowActivityTypeProperties, self).__init__(**kwargs) self.data_flow = data_flow @@ -22623,6 +22647,7 @@ def __init__( self.trace_level = trace_level self.continue_on_error = continue_on_error self.run_concurrently = run_concurrently + self.source_staging_concurrency = source_staging_concurrency class ExecuteDataFlowActivityTypePropertiesCompute(msrest.serialization.Model): @@ -22815,6 +22840,9 @@ class ExecutePowerQueryActivityTypeProperties(ExecuteDataFlowActivityTypePropert the same save order to be processed concurrently. Type: boolean (or Expression with resultType boolean). :vartype run_concurrently: any + :ivar source_staging_concurrency: Specify number of parallel staging for sources applicable to + the sink. Type: integer (or Expression with resultType integer). + :vartype source_staging_concurrency: any :ivar sinks: (Deprecated. Please use Queries). List of Power Query activity sinks mapped to a queryName. :vartype sinks: dict[str, ~azure.mgmt.datafactory.models.PowerQuerySink] @@ -22834,6 +22862,7 @@ class ExecutePowerQueryActivityTypeProperties(ExecuteDataFlowActivityTypePropert 'trace_level': {'key': 'traceLevel', 'type': 'object'}, 'continue_on_error': {'key': 'continueOnError', 'type': 'object'}, 'run_concurrently': {'key': 'runConcurrently', 'type': 'object'}, + 'source_staging_concurrency': {'key': 'sourceStagingConcurrency', 'type': 'object'}, 'sinks': {'key': 'sinks', 'type': '{PowerQuerySink}'}, 'queries': {'key': 'queries', 'type': '[PowerQuerySinkMapping]'}, } @@ -22848,6 +22877,7 @@ def __init__( trace_level: Optional[Any] = None, continue_on_error: Optional[Any] = None, run_concurrently: Optional[Any] = None, + source_staging_concurrency: Optional[Any] = None, sinks: Optional[Dict[str, "PowerQuerySink"]] = None, queries: Optional[List["PowerQuerySinkMapping"]] = None, **kwargs @@ -22871,13 +22901,16 @@ def __init__( with the same save order to be processed concurrently. Type: boolean (or Expression with resultType boolean). :paramtype run_concurrently: any + :keyword source_staging_concurrency: Specify number of parallel staging for sources applicable + to the sink. Type: integer (or Expression with resultType integer). + :paramtype source_staging_concurrency: any :keyword sinks: (Deprecated. Please use Queries). List of Power Query activity sinks mapped to a queryName. :paramtype sinks: dict[str, ~azure.mgmt.datafactory.models.PowerQuerySink] :keyword queries: List of mapping for Power Query mashup query to sink dataset(s). :paramtype queries: list[~azure.mgmt.datafactory.models.PowerQuerySinkMapping] """ - super(ExecutePowerQueryActivityTypeProperties, self).__init__(data_flow=data_flow, staging=staging, integration_runtime=integration_runtime, compute=compute, trace_level=trace_level, continue_on_error=continue_on_error, run_concurrently=run_concurrently, **kwargs) + super(ExecutePowerQueryActivityTypeProperties, self).__init__(data_flow=data_flow, staging=staging, integration_runtime=integration_runtime, compute=compute, trace_level=trace_level, continue_on_error=continue_on_error, run_concurrently=run_concurrently, source_staging_concurrency=source_staging_concurrency, **kwargs) self.sinks = sinks self.queries = queries @@ -23093,6 +23126,9 @@ class ExecuteWranglingDataflowActivity(Activity): the same save order to be processed concurrently. Type: boolean (or Expression with resultType boolean). :vartype run_concurrently: any + :ivar source_staging_concurrency: Specify number of parallel staging for sources applicable to + the sink. Type: integer (or Expression with resultType integer). + :vartype source_staging_concurrency: any :ivar sinks: (Deprecated. Please use Queries). List of Power Query activity sinks mapped to a queryName. :vartype sinks: dict[str, ~azure.mgmt.datafactory.models.PowerQuerySink] @@ -23121,6 +23157,7 @@ class ExecuteWranglingDataflowActivity(Activity): 'trace_level': {'key': 'typeProperties.traceLevel', 'type': 'object'}, 'continue_on_error': {'key': 'typeProperties.continueOnError', 'type': 'object'}, 'run_concurrently': {'key': 'typeProperties.runConcurrently', 'type': 'object'}, + 'source_staging_concurrency': {'key': 'typeProperties.sourceStagingConcurrency', 'type': 'object'}, 'sinks': {'key': 'typeProperties.sinks', 'type': '{PowerQuerySink}'}, 'queries': {'key': 'typeProperties.queries', 'type': '[PowerQuerySinkMapping]'}, } @@ -23141,6 +23178,7 @@ def __init__( trace_level: Optional[Any] = None, continue_on_error: Optional[Any] = None, run_concurrently: Optional[Any] = None, + source_staging_concurrency: Optional[Any] = None, sinks: Optional[Dict[str, "PowerQuerySink"]] = None, queries: Optional[List["PowerQuerySinkMapping"]] = None, **kwargs @@ -23177,6 +23215,9 @@ def __init__( with the same save order to be processed concurrently. Type: boolean (or Expression with resultType boolean). :paramtype run_concurrently: any + :keyword source_staging_concurrency: Specify number of parallel staging for sources applicable + to the sink. Type: integer (or Expression with resultType integer). + :paramtype source_staging_concurrency: any :keyword sinks: (Deprecated. Please use Queries). List of Power Query activity sinks mapped to a queryName. :paramtype sinks: dict[str, ~azure.mgmt.datafactory.models.PowerQuerySink] @@ -23193,6 +23234,7 @@ def __init__( self.trace_level = trace_level self.continue_on_error = continue_on_error self.run_concurrently = run_concurrently + self.source_staging_concurrency = source_staging_concurrency self.sinks = sinks self.queries = queries @@ -23450,6 +23492,8 @@ class Factory(Resource): :vartype create_time: ~datetime.datetime :ivar version: Version of the factory. :vartype version: str + :ivar purview_configuration: Purview information of the factory. + :vartype purview_configuration: ~azure.mgmt.datafactory.models.PurviewConfiguration :ivar repo_configuration: Git repo information of the factory. :vartype repo_configuration: ~azure.mgmt.datafactory.models.FactoryRepoConfiguration :ivar global_parameters: List of parameters for factory. @@ -23484,6 +23528,7 @@ class Factory(Resource): 'provisioning_state': {'key': 'properties.provisioningState', 'type': 'str'}, 'create_time': {'key': 'properties.createTime', 'type': 'iso-8601'}, 'version': {'key': 'properties.version', 'type': 'str'}, + 'purview_configuration': {'key': 'properties.purviewConfiguration', 'type': 'PurviewConfiguration'}, 'repo_configuration': {'key': 'properties.repoConfiguration', 'type': 'FactoryRepoConfiguration'}, 'global_parameters': {'key': 'properties.globalParameters', 'type': '{GlobalParameterSpecification}'}, 'encryption': {'key': 'properties.encryption', 'type': 'EncryptionConfiguration'}, @@ -23497,6 +23542,7 @@ def __init__( tags: Optional[Dict[str, str]] = None, additional_properties: Optional[Dict[str, Any]] = None, identity: Optional["FactoryIdentity"] = None, + purview_configuration: Optional["PurviewConfiguration"] = None, repo_configuration: Optional["FactoryRepoConfiguration"] = None, global_parameters: Optional[Dict[str, "GlobalParameterSpecification"]] = None, encryption: Optional["EncryptionConfiguration"] = None, @@ -23513,6 +23559,8 @@ def __init__( :paramtype additional_properties: dict[str, any] :keyword identity: Managed service identity of the factory. :paramtype identity: ~azure.mgmt.datafactory.models.FactoryIdentity + :keyword purview_configuration: Purview information of the factory. + :paramtype purview_configuration: ~azure.mgmt.datafactory.models.PurviewConfiguration :keyword repo_configuration: Git repo information of the factory. :paramtype repo_configuration: ~azure.mgmt.datafactory.models.FactoryRepoConfiguration :keyword global_parameters: List of parameters for factory. @@ -23530,6 +23578,7 @@ def __init__( self.provisioning_state = None self.create_time = None self.version = None + self.purview_configuration = purview_configuration self.repo_configuration = repo_configuration self.global_parameters = global_parameters self.encryption = encryption @@ -42452,6 +42501,8 @@ class PowerQuerySink(DataFlowSink): :vartype flowlet: ~azure.mgmt.datafactory.models.DataFlowReference :ivar schema_linked_service: Schema linked service reference. :vartype schema_linked_service: ~azure.mgmt.datafactory.models.LinkedServiceReference + :ivar rejected_data_linked_service: Rejected data linked service reference. + :vartype rejected_data_linked_service: ~azure.mgmt.datafactory.models.LinkedServiceReference :ivar script: sink script. :vartype script: str """ @@ -42467,6 +42518,7 @@ class PowerQuerySink(DataFlowSink): 'linked_service': {'key': 'linkedService', 'type': 'LinkedServiceReference'}, 'flowlet': {'key': 'flowlet', 'type': 'DataFlowReference'}, 'schema_linked_service': {'key': 'schemaLinkedService', 'type': 'LinkedServiceReference'}, + 'rejected_data_linked_service': {'key': 'rejectedDataLinkedService', 'type': 'LinkedServiceReference'}, 'script': {'key': 'script', 'type': 'str'}, } @@ -42479,6 +42531,7 @@ def __init__( linked_service: Optional["LinkedServiceReference"] = None, flowlet: Optional["DataFlowReference"] = None, schema_linked_service: Optional["LinkedServiceReference"] = None, + rejected_data_linked_service: Optional["LinkedServiceReference"] = None, script: Optional[str] = None, **kwargs ): @@ -42495,10 +42548,12 @@ def __init__( :paramtype flowlet: ~azure.mgmt.datafactory.models.DataFlowReference :keyword schema_linked_service: Schema linked service reference. :paramtype schema_linked_service: ~azure.mgmt.datafactory.models.LinkedServiceReference + :keyword rejected_data_linked_service: Rejected data linked service reference. + :paramtype rejected_data_linked_service: ~azure.mgmt.datafactory.models.LinkedServiceReference :keyword script: sink script. :paramtype script: str """ - super(PowerQuerySink, self).__init__(name=name, description=description, dataset=dataset, linked_service=linked_service, flowlet=flowlet, schema_linked_service=schema_linked_service, **kwargs) + super(PowerQuerySink, self).__init__(name=name, description=description, dataset=dataset, linked_service=linked_service, flowlet=flowlet, schema_linked_service=schema_linked_service, rejected_data_linked_service=rejected_data_linked_service, **kwargs) self.script = script @@ -42990,6 +43045,31 @@ def __init__( self.query = query +class PrivateEndpoint(msrest.serialization.Model): + """Private endpoint which a connection belongs to. + + :ivar id: The resource Id for private endpoint. + :vartype id: str + """ + + _attribute_map = { + 'id': {'key': 'id', 'type': 'str'}, + } + + def __init__( + self, + *, + id: Optional[str] = None, + **kwargs + ): + """ + :keyword id: The resource Id for private endpoint. + :paramtype id: str + """ + super(PrivateEndpoint, self).__init__(**kwargs) + self.id = id + + class PrivateEndpointConnectionListResponse(msrest.serialization.Model): """A list of linked service resources. @@ -43080,25 +43160,32 @@ class PrivateLinkConnectionApprovalRequest(msrest.serialization.Model): :ivar private_link_service_connection_state: The state of a private link connection. :vartype private_link_service_connection_state: ~azure.mgmt.datafactory.models.PrivateLinkConnectionState + :ivar private_endpoint: The resource of private endpoint. + :vartype private_endpoint: ~azure.mgmt.datafactory.models.PrivateEndpoint """ _attribute_map = { 'private_link_service_connection_state': {'key': 'privateLinkServiceConnectionState', 'type': 'PrivateLinkConnectionState'}, + 'private_endpoint': {'key': 'privateEndpoint', 'type': 'PrivateEndpoint'}, } def __init__( self, *, private_link_service_connection_state: Optional["PrivateLinkConnectionState"] = None, + private_endpoint: Optional["PrivateEndpoint"] = None, **kwargs ): """ :keyword private_link_service_connection_state: The state of a private link connection. :paramtype private_link_service_connection_state: ~azure.mgmt.datafactory.models.PrivateLinkConnectionState + :keyword private_endpoint: The resource of private endpoint. + :paramtype private_endpoint: ~azure.mgmt.datafactory.models.PrivateEndpoint """ super(PrivateLinkConnectionApprovalRequest, self).__init__(**kwargs) self.private_link_service_connection_state = private_link_service_connection_state + self.private_endpoint = private_endpoint class PrivateLinkConnectionApprovalRequestResource(SubResource): @@ -43300,6 +43387,31 @@ def __init__( self.value = value +class PurviewConfiguration(msrest.serialization.Model): + """Purview configuration. + + :ivar purview_resource_id: Purview resource id. + :vartype purview_resource_id: str + """ + + _attribute_map = { + 'purview_resource_id': {'key': 'purviewResourceId', 'type': 'str'}, + } + + def __init__( + self, + *, + purview_resource_id: Optional[str] = None, + **kwargs + ): + """ + :keyword purview_resource_id: Purview resource id. + :paramtype purview_resource_id: str + """ + super(PurviewConfiguration, self).__init__(**kwargs) + self.purview_resource_id = purview_resource_id + + class QueryDataFlowDebugSessionsResponse(msrest.serialization.Model): """A list of active debug sessions.