From fad0b31e7c2ced30edb5579844d878ebb370ec93 Mon Sep 17 00:00:00 2001 From: SDKAuto Date: Thu, 14 Oct 2021 23:50:38 +0000 Subject: [PATCH] CodeGen from PR 16394 in Azure/azure-rest-api-specs Merge f05ce21e79151e76917a804b59fd87dc400d96f1 into 7e6be1c9e6c04f932795cb114599e9b5b0c90d0a --- .../azure-mgmt-datafactory/_meta.json | 2 +- .../azure/mgmt/datafactory/_version.py | 2 +- .../azure/mgmt/datafactory/models/__init__.py | 3 + .../azure/mgmt/datafactory/models/_models.py | 104 +++++++++++++- .../mgmt/datafactory/models/_models_py3.py | 132 ++++++++++++++++-- 5 files changed, 225 insertions(+), 18 deletions(-) diff --git a/sdk/datafactory/azure-mgmt-datafactory/_meta.json b/sdk/datafactory/azure-mgmt-datafactory/_meta.json index 9d25a2b50221..40bfe0ff5552 100644 --- a/sdk/datafactory/azure-mgmt-datafactory/_meta.json +++ b/sdk/datafactory/azure-mgmt-datafactory/_meta.json @@ -4,7 +4,7 @@ "@autorest/python@5.8.4", "@autorest/modelerfour@4.19.2" ], - "commit": "2ce915398bfadd5333820487595a9623187dcb59", + "commit": "0a73c60843e9066b1c85630e921d191b8f9d183e", "repository_url": "https://github.com/Azure/azure-rest-api-specs", "autorest_command": "autorest specification/datafactory/resource-manager/readme.md --multiapi --python --python-mode=update --python-sdks-folder=/home/vsts/work/1/s/azure-sdk-for-python/sdk --track2 --use=@autorest/python@5.8.4 --use=@autorest/modelerfour@4.19.2 --version=3.4.5", "readme": "specification/datafactory/resource-manager/readme.md" diff --git a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/_version.py b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/_version.py index 48944bf3938a..c47f66669f1b 100644 --- a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/_version.py +++ b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/_version.py @@ -6,4 +6,4 @@ # Changes may cause incorrect behavior and will be lost if the code is regenerated. # -------------------------------------------------------------------------- -VERSION = "2.0.0" +VERSION = "1.0.0" diff --git a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py index d05e600bf0b6..9a93a161658f 100644 --- a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py +++ b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py @@ -279,6 +279,7 @@ from ._models_py3 import FileSystemSink from ._models_py3 import FileSystemSource from ._models_py3 import FilterActivity + from ._models_py3 import Flowlet from ._models_py3 import ForEachActivity from ._models_py3 import FormatReadSettings from ._models_py3 import FormatWriteSettings @@ -974,6 +975,7 @@ from ._models import FileSystemSink # type: ignore from ._models import FileSystemSource # type: ignore from ._models import FilterActivity # type: ignore + from ._models import Flowlet # type: ignore from ._models import ForEachActivity # type: ignore from ._models import FormatReadSettings # type: ignore from ._models import FormatWriteSettings # type: ignore @@ -1768,6 +1770,7 @@ 'FileSystemSink', 'FileSystemSource', 'FilterActivity', + 'Flowlet', 'ForEachActivity', 'FormatReadSettings', 'FormatWriteSettings', diff --git a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/_models.py b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/_models.py index 5f4f0c460837..c0c1fb23dddb 100644 --- a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/_models.py +++ b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/_models.py @@ -11051,7 +11051,7 @@ class DataFlow(msrest.serialization.Model): """Azure Data Factory nested object which contains a flow with data movements and transformations. You probably want to use the sub-classes and not this class directly. Known - sub-classes are: MappingDataFlow, WranglingDataFlow. + sub-classes are: Flowlet, MappingDataFlow, WranglingDataFlow. All required parameters must be populated in order to send to Azure. @@ -11078,7 +11078,7 @@ class DataFlow(msrest.serialization.Model): } _subtype_map = { - 'type': {'MappingDataFlow': 'MappingDataFlow', 'WranglingDataFlow': 'WranglingDataFlow'} + 'type': {'Flowlet': 'Flowlet', 'MappingDataFlow': 'MappingDataFlow', 'WranglingDataFlow': 'WranglingDataFlow'} } def __init__( @@ -11190,6 +11190,8 @@ class DataFlowDebugPackage(msrest.serialization.Model): :type session_id: str :param data_flow: Data flow instance. :type data_flow: ~azure.mgmt.datafactory.models.DataFlowDebugResource + :param data_flows: List of Data flows. + :type data_flows: list[~azure.mgmt.datafactory.models.DataFlowDebugResource] :param datasets: List of datasets. :type datasets: list[~azure.mgmt.datafactory.models.DatasetDebugResource] :param linked_services: List of linked services. @@ -11204,6 +11206,7 @@ class DataFlowDebugPackage(msrest.serialization.Model): 'additional_properties': {'key': '', 'type': '{object}'}, 'session_id': {'key': 'sessionId', 'type': 'str'}, 'data_flow': {'key': 'dataFlow', 'type': 'DataFlowDebugResource'}, + 'data_flows': {'key': 'dataFlows', 'type': '[DataFlowDebugResource]'}, 'datasets': {'key': 'datasets', 'type': '[DatasetDebugResource]'}, 'linked_services': {'key': 'linkedServices', 'type': '[LinkedServiceDebugResource]'}, 'staging': {'key': 'staging', 'type': 'DataFlowStagingInfo'}, @@ -11218,6 +11221,7 @@ def __init__( self.additional_properties = kwargs.get('additional_properties', None) self.session_id = kwargs.get('session_id', None) self.data_flow = kwargs.get('data_flow', None) + self.data_flows = kwargs.get('data_flows', None) self.datasets = kwargs.get('datasets', None) self.linked_services = kwargs.get('linked_services', None) self.staging = kwargs.get('staging', None) @@ -11496,6 +11500,12 @@ class Transformation(msrest.serialization.Model): :type name: str :param description: Transformation description. :type description: str + :param dataset: Dataset reference. + :type dataset: ~azure.mgmt.datafactory.models.DatasetReference + :param linked_service: Linked service reference. + :type linked_service: ~azure.mgmt.datafactory.models.LinkedServiceReference + :param flowlet: Flowlet Reference. + :type flowlet: ~azure.mgmt.datafactory.models.DataFlowReference """ _validation = { @@ -11505,6 +11515,9 @@ class Transformation(msrest.serialization.Model): _attribute_map = { 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, + 'dataset': {'key': 'dataset', 'type': 'DatasetReference'}, + 'linked_service': {'key': 'linkedService', 'type': 'LinkedServiceReference'}, + 'flowlet': {'key': 'flowlet', 'type': 'DataFlowReference'}, } def __init__( @@ -11514,6 +11527,9 @@ def __init__( super(Transformation, self).__init__(**kwargs) self.name = kwargs['name'] self.description = kwargs.get('description', None) + self.dataset = kwargs.get('dataset', None) + self.linked_service = kwargs.get('linked_service', None) + self.flowlet = kwargs.get('flowlet', None) class DataFlowSink(Transformation): @@ -11529,6 +11545,8 @@ class DataFlowSink(Transformation): :type dataset: ~azure.mgmt.datafactory.models.DatasetReference :param linked_service: Linked service reference. :type linked_service: ~azure.mgmt.datafactory.models.LinkedServiceReference + :param flowlet: Flowlet Reference. + :type flowlet: ~azure.mgmt.datafactory.models.DataFlowReference :param schema_linked_service: Schema linked service reference. :type schema_linked_service: ~azure.mgmt.datafactory.models.LinkedServiceReference """ @@ -11542,6 +11560,7 @@ class DataFlowSink(Transformation): 'description': {'key': 'description', 'type': 'str'}, 'dataset': {'key': 'dataset', 'type': 'DatasetReference'}, 'linked_service': {'key': 'linkedService', 'type': 'LinkedServiceReference'}, + 'flowlet': {'key': 'flowlet', 'type': 'DataFlowReference'}, 'schema_linked_service': {'key': 'schemaLinkedService', 'type': 'LinkedServiceReference'}, } @@ -11550,8 +11569,6 @@ def __init__( **kwargs ): super(DataFlowSink, self).__init__(**kwargs) - self.dataset = kwargs.get('dataset', None) - self.linked_service = kwargs.get('linked_service', None) self.schema_linked_service = kwargs.get('schema_linked_service', None) @@ -11568,6 +11585,8 @@ class DataFlowSource(Transformation): :type dataset: ~azure.mgmt.datafactory.models.DatasetReference :param linked_service: Linked service reference. :type linked_service: ~azure.mgmt.datafactory.models.LinkedServiceReference + :param flowlet: Flowlet Reference. + :type flowlet: ~azure.mgmt.datafactory.models.DataFlowReference :param schema_linked_service: Schema linked service reference. :type schema_linked_service: ~azure.mgmt.datafactory.models.LinkedServiceReference """ @@ -11581,6 +11600,7 @@ class DataFlowSource(Transformation): 'description': {'key': 'description', 'type': 'str'}, 'dataset': {'key': 'dataset', 'type': 'DatasetReference'}, 'linked_service': {'key': 'linkedService', 'type': 'LinkedServiceReference'}, + 'flowlet': {'key': 'flowlet', 'type': 'DataFlowReference'}, 'schema_linked_service': {'key': 'schemaLinkedService', 'type': 'LinkedServiceReference'}, } @@ -11589,8 +11609,6 @@ def __init__( **kwargs ): super(DataFlowSource, self).__init__(**kwargs) - self.dataset = kwargs.get('dataset', None) - self.linked_service = kwargs.get('linked_service', None) self.schema_linked_service = kwargs.get('schema_linked_service', None) @@ -16009,6 +16027,70 @@ def __init__( self.condition = kwargs['condition'] +class Flowlet(DataFlow): + """Data flow flowlet. + + All required parameters must be populated in order to send to Azure. + + :param type: Required. Type of data flow.Constant filled by server. + :type type: str + :param description: The description of the data flow. + :type description: str + :param annotations: List of tags that can be used for describing the data flow. + :type annotations: list[any] + :param folder: The folder that this data flow is in. If not specified, Data flow will appear at + the root level. + :type folder: ~azure.mgmt.datafactory.models.DataFlowFolder + :param additional_properties: Unmatched properties from the message are deserialized to this + collection. + :type additional_properties: dict[str, any] + :param sources: List of sources in Flowlet. + :type sources: list[~azure.mgmt.datafactory.models.DataFlowSource] + :param sinks: List of sinks in Flowlet. + :type sinks: list[~azure.mgmt.datafactory.models.DataFlowSink] + :param transformations: List of transformations in Flowlet. + :type transformations: list[~azure.mgmt.datafactory.models.Transformation] + :param script: Flowlet script. + :type script: str + :param script_lines: Flowlet script lines. + :type script_lines: list[str] + :param additional_properties1: Any object. + :type additional_properties1: any + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'type': {'key': 'type', 'type': 'str'}, + 'description': {'key': 'description', 'type': 'str'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'folder': {'key': 'folder', 'type': 'DataFlowFolder'}, + 'additional_properties': {'key': '', 'type': '{object}'}, + 'sources': {'key': 'typeProperties.sources', 'type': '[DataFlowSource]'}, + 'sinks': {'key': 'typeProperties.sinks', 'type': '[DataFlowSink]'}, + 'transformations': {'key': 'typeProperties.transformations', 'type': '[Transformation]'}, + 'script': {'key': 'typeProperties.script', 'type': 'str'}, + 'script_lines': {'key': 'typeProperties.scriptLines', 'type': '[str]'}, + 'additional_properties1': {'key': 'typeProperties.additionalProperties', 'type': 'object'}, + } + + def __init__( + self, + **kwargs + ): + super(Flowlet, self).__init__(**kwargs) + self.type = 'Flowlet' # type: str + self.additional_properties = kwargs.get('additional_properties', None) + self.sources = kwargs.get('sources', None) + self.sinks = kwargs.get('sinks', None) + self.transformations = kwargs.get('transformations', None) + self.script = kwargs.get('script', None) + self.script_lines = kwargs.get('script_lines', None) + self.additional_properties1 = kwargs.get('additional_properties1', None) + + class ForEachActivity(ControlActivity): """This activity is used for iterating over a collection and execute given activities. @@ -22502,6 +22584,8 @@ class MappingDataFlow(DataFlow): :type transformations: list[~azure.mgmt.datafactory.models.Transformation] :param script: DataFlow script. :type script: str + :param script_lines: Data flow script lines. + :type script_lines: list[str] """ _validation = { @@ -22517,6 +22601,7 @@ class MappingDataFlow(DataFlow): 'sinks': {'key': 'typeProperties.sinks', 'type': '[DataFlowSink]'}, 'transformations': {'key': 'typeProperties.transformations', 'type': '[Transformation]'}, 'script': {'key': 'typeProperties.script', 'type': 'str'}, + 'script_lines': {'key': 'typeProperties.scriptLines', 'type': '[str]'}, } def __init__( @@ -22529,6 +22614,7 @@ def __init__( self.sinks = kwargs.get('sinks', None) self.transformations = kwargs.get('transformations', None) self.script = kwargs.get('script', None) + self.script_lines = kwargs.get('script_lines', None) class MariaDBLinkedService(LinkedService): @@ -27510,6 +27596,8 @@ class PowerQuerySink(DataFlowSink): :type dataset: ~azure.mgmt.datafactory.models.DatasetReference :param linked_service: Linked service reference. :type linked_service: ~azure.mgmt.datafactory.models.LinkedServiceReference + :param flowlet: Flowlet Reference. + :type flowlet: ~azure.mgmt.datafactory.models.DataFlowReference :param schema_linked_service: Schema linked service reference. :type schema_linked_service: ~azure.mgmt.datafactory.models.LinkedServiceReference :param script: sink script. @@ -27525,6 +27613,7 @@ class PowerQuerySink(DataFlowSink): 'description': {'key': 'description', 'type': 'str'}, 'dataset': {'key': 'dataset', 'type': 'DatasetReference'}, 'linked_service': {'key': 'linkedService', 'type': 'LinkedServiceReference'}, + 'flowlet': {'key': 'flowlet', 'type': 'DataFlowReference'}, 'schema_linked_service': {'key': 'schemaLinkedService', 'type': 'LinkedServiceReference'}, 'script': {'key': 'script', 'type': 'str'}, } @@ -27573,6 +27662,8 @@ class PowerQuerySource(DataFlowSource): :type dataset: ~azure.mgmt.datafactory.models.DatasetReference :param linked_service: Linked service reference. :type linked_service: ~azure.mgmt.datafactory.models.LinkedServiceReference + :param flowlet: Flowlet Reference. + :type flowlet: ~azure.mgmt.datafactory.models.DataFlowReference :param schema_linked_service: Schema linked service reference. :type schema_linked_service: ~azure.mgmt.datafactory.models.LinkedServiceReference :param script: source script. @@ -27588,6 +27679,7 @@ class PowerQuerySource(DataFlowSource): 'description': {'key': 'description', 'type': 'str'}, 'dataset': {'key': 'dataset', 'type': 'DatasetReference'}, 'linked_service': {'key': 'linkedService', 'type': 'LinkedServiceReference'}, + 'flowlet': {'key': 'flowlet', 'type': 'DataFlowReference'}, 'schema_linked_service': {'key': 'schemaLinkedService', 'type': 'LinkedServiceReference'}, 'script': {'key': 'script', 'type': 'str'}, } diff --git a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/_models_py3.py b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/_models_py3.py index 963145597ea9..81a39bf71410 100644 --- a/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/_models_py3.py +++ b/sdk/datafactory/azure-mgmt-datafactory/azure/mgmt/datafactory/models/_models_py3.py @@ -12696,7 +12696,7 @@ class DataFlow(msrest.serialization.Model): """Azure Data Factory nested object which contains a flow with data movements and transformations. You probably want to use the sub-classes and not this class directly. Known - sub-classes are: MappingDataFlow, WranglingDataFlow. + sub-classes are: Flowlet, MappingDataFlow, WranglingDataFlow. All required parameters must be populated in order to send to Azure. @@ -12723,7 +12723,7 @@ class DataFlow(msrest.serialization.Model): } _subtype_map = { - 'type': {'MappingDataFlow': 'MappingDataFlow', 'WranglingDataFlow': 'WranglingDataFlow'} + 'type': {'Flowlet': 'Flowlet', 'MappingDataFlow': 'MappingDataFlow', 'WranglingDataFlow': 'WranglingDataFlow'} } def __init__( @@ -12851,6 +12851,8 @@ class DataFlowDebugPackage(msrest.serialization.Model): :type session_id: str :param data_flow: Data flow instance. :type data_flow: ~azure.mgmt.datafactory.models.DataFlowDebugResource + :param data_flows: List of Data flows. + :type data_flows: list[~azure.mgmt.datafactory.models.DataFlowDebugResource] :param datasets: List of datasets. :type datasets: list[~azure.mgmt.datafactory.models.DatasetDebugResource] :param linked_services: List of linked services. @@ -12865,6 +12867,7 @@ class DataFlowDebugPackage(msrest.serialization.Model): 'additional_properties': {'key': '', 'type': '{object}'}, 'session_id': {'key': 'sessionId', 'type': 'str'}, 'data_flow': {'key': 'dataFlow', 'type': 'DataFlowDebugResource'}, + 'data_flows': {'key': 'dataFlows', 'type': '[DataFlowDebugResource]'}, 'datasets': {'key': 'datasets', 'type': '[DatasetDebugResource]'}, 'linked_services': {'key': 'linkedServices', 'type': '[LinkedServiceDebugResource]'}, 'staging': {'key': 'staging', 'type': 'DataFlowStagingInfo'}, @@ -12877,6 +12880,7 @@ def __init__( additional_properties: Optional[Dict[str, Any]] = None, session_id: Optional[str] = None, data_flow: Optional["DataFlowDebugResource"] = None, + data_flows: Optional[List["DataFlowDebugResource"]] = None, datasets: Optional[List["DatasetDebugResource"]] = None, linked_services: Optional[List["LinkedServiceDebugResource"]] = None, staging: Optional["DataFlowStagingInfo"] = None, @@ -12887,6 +12891,7 @@ def __init__( self.additional_properties = additional_properties self.session_id = session_id self.data_flow = data_flow + self.data_flows = data_flows self.datasets = datasets self.linked_services = linked_services self.staging = staging @@ -13196,6 +13201,12 @@ class Transformation(msrest.serialization.Model): :type name: str :param description: Transformation description. :type description: str + :param dataset: Dataset reference. + :type dataset: ~azure.mgmt.datafactory.models.DatasetReference + :param linked_service: Linked service reference. + :type linked_service: ~azure.mgmt.datafactory.models.LinkedServiceReference + :param flowlet: Flowlet Reference. + :type flowlet: ~azure.mgmt.datafactory.models.DataFlowReference """ _validation = { @@ -13205,6 +13216,9 @@ class Transformation(msrest.serialization.Model): _attribute_map = { 'name': {'key': 'name', 'type': 'str'}, 'description': {'key': 'description', 'type': 'str'}, + 'dataset': {'key': 'dataset', 'type': 'DatasetReference'}, + 'linked_service': {'key': 'linkedService', 'type': 'LinkedServiceReference'}, + 'flowlet': {'key': 'flowlet', 'type': 'DataFlowReference'}, } def __init__( @@ -13212,11 +13226,17 @@ def __init__( *, name: str, description: Optional[str] = None, + dataset: Optional["DatasetReference"] = None, + linked_service: Optional["LinkedServiceReference"] = None, + flowlet: Optional["DataFlowReference"] = None, **kwargs ): super(Transformation, self).__init__(**kwargs) self.name = name self.description = description + self.dataset = dataset + self.linked_service = linked_service + self.flowlet = flowlet class DataFlowSink(Transformation): @@ -13232,6 +13252,8 @@ class DataFlowSink(Transformation): :type dataset: ~azure.mgmt.datafactory.models.DatasetReference :param linked_service: Linked service reference. :type linked_service: ~azure.mgmt.datafactory.models.LinkedServiceReference + :param flowlet: Flowlet Reference. + :type flowlet: ~azure.mgmt.datafactory.models.DataFlowReference :param schema_linked_service: Schema linked service reference. :type schema_linked_service: ~azure.mgmt.datafactory.models.LinkedServiceReference """ @@ -13245,6 +13267,7 @@ class DataFlowSink(Transformation): 'description': {'key': 'description', 'type': 'str'}, 'dataset': {'key': 'dataset', 'type': 'DatasetReference'}, 'linked_service': {'key': 'linkedService', 'type': 'LinkedServiceReference'}, + 'flowlet': {'key': 'flowlet', 'type': 'DataFlowReference'}, 'schema_linked_service': {'key': 'schemaLinkedService', 'type': 'LinkedServiceReference'}, } @@ -13255,12 +13278,11 @@ def __init__( description: Optional[str] = None, dataset: Optional["DatasetReference"] = None, linked_service: Optional["LinkedServiceReference"] = None, + flowlet: Optional["DataFlowReference"] = None, schema_linked_service: Optional["LinkedServiceReference"] = None, **kwargs ): - super(DataFlowSink, self).__init__(name=name, description=description, **kwargs) - self.dataset = dataset - self.linked_service = linked_service + super(DataFlowSink, self).__init__(name=name, description=description, dataset=dataset, linked_service=linked_service, flowlet=flowlet, **kwargs) self.schema_linked_service = schema_linked_service @@ -13277,6 +13299,8 @@ class DataFlowSource(Transformation): :type dataset: ~azure.mgmt.datafactory.models.DatasetReference :param linked_service: Linked service reference. :type linked_service: ~azure.mgmt.datafactory.models.LinkedServiceReference + :param flowlet: Flowlet Reference. + :type flowlet: ~azure.mgmt.datafactory.models.DataFlowReference :param schema_linked_service: Schema linked service reference. :type schema_linked_service: ~azure.mgmt.datafactory.models.LinkedServiceReference """ @@ -13290,6 +13314,7 @@ class DataFlowSource(Transformation): 'description': {'key': 'description', 'type': 'str'}, 'dataset': {'key': 'dataset', 'type': 'DatasetReference'}, 'linked_service': {'key': 'linkedService', 'type': 'LinkedServiceReference'}, + 'flowlet': {'key': 'flowlet', 'type': 'DataFlowReference'}, 'schema_linked_service': {'key': 'schemaLinkedService', 'type': 'LinkedServiceReference'}, } @@ -13300,12 +13325,11 @@ def __init__( description: Optional[str] = None, dataset: Optional["DatasetReference"] = None, linked_service: Optional["LinkedServiceReference"] = None, + flowlet: Optional["DataFlowReference"] = None, schema_linked_service: Optional["LinkedServiceReference"] = None, **kwargs ): - super(DataFlowSource, self).__init__(name=name, description=description, **kwargs) - self.dataset = dataset - self.linked_service = linked_service + super(DataFlowSource, self).__init__(name=name, description=description, dataset=dataset, linked_service=linked_service, flowlet=flowlet, **kwargs) self.schema_linked_service = schema_linked_service @@ -18359,6 +18383,81 @@ def __init__( self.condition = condition +class Flowlet(DataFlow): + """Data flow flowlet. + + All required parameters must be populated in order to send to Azure. + + :param type: Required. Type of data flow.Constant filled by server. + :type type: str + :param description: The description of the data flow. + :type description: str + :param annotations: List of tags that can be used for describing the data flow. + :type annotations: list[any] + :param folder: The folder that this data flow is in. If not specified, Data flow will appear at + the root level. + :type folder: ~azure.mgmt.datafactory.models.DataFlowFolder + :param additional_properties: Unmatched properties from the message are deserialized to this + collection. + :type additional_properties: dict[str, any] + :param sources: List of sources in Flowlet. + :type sources: list[~azure.mgmt.datafactory.models.DataFlowSource] + :param sinks: List of sinks in Flowlet. + :type sinks: list[~azure.mgmt.datafactory.models.DataFlowSink] + :param transformations: List of transformations in Flowlet. + :type transformations: list[~azure.mgmt.datafactory.models.Transformation] + :param script: Flowlet script. + :type script: str + :param script_lines: Flowlet script lines. + :type script_lines: list[str] + :param additional_properties1: Any object. + :type additional_properties1: any + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'type': {'key': 'type', 'type': 'str'}, + 'description': {'key': 'description', 'type': 'str'}, + 'annotations': {'key': 'annotations', 'type': '[object]'}, + 'folder': {'key': 'folder', 'type': 'DataFlowFolder'}, + 'additional_properties': {'key': '', 'type': '{object}'}, + 'sources': {'key': 'typeProperties.sources', 'type': '[DataFlowSource]'}, + 'sinks': {'key': 'typeProperties.sinks', 'type': '[DataFlowSink]'}, + 'transformations': {'key': 'typeProperties.transformations', 'type': '[Transformation]'}, + 'script': {'key': 'typeProperties.script', 'type': 'str'}, + 'script_lines': {'key': 'typeProperties.scriptLines', 'type': '[str]'}, + 'additional_properties1': {'key': 'typeProperties.additionalProperties', 'type': 'object'}, + } + + def __init__( + self, + *, + description: Optional[str] = None, + annotations: Optional[List[Any]] = None, + folder: Optional["DataFlowFolder"] = None, + additional_properties: Optional[Dict[str, Any]] = None, + sources: Optional[List["DataFlowSource"]] = None, + sinks: Optional[List["DataFlowSink"]] = None, + transformations: Optional[List["Transformation"]] = None, + script: Optional[str] = None, + script_lines: Optional[List[str]] = None, + additional_properties1: Optional[Any] = None, + **kwargs + ): + super(Flowlet, self).__init__(description=description, annotations=annotations, folder=folder, **kwargs) + self.type = 'Flowlet' # type: str + self.additional_properties = additional_properties + self.sources = sources + self.sinks = sinks + self.transformations = transformations + self.script = script + self.script_lines = script_lines + self.additional_properties1 = additional_properties1 + + class ForEachActivity(ControlActivity): """This activity is used for iterating over a collection and execute given activities. @@ -25763,6 +25862,8 @@ class MappingDataFlow(DataFlow): :type transformations: list[~azure.mgmt.datafactory.models.Transformation] :param script: DataFlow script. :type script: str + :param script_lines: Data flow script lines. + :type script_lines: list[str] """ _validation = { @@ -25778,6 +25879,7 @@ class MappingDataFlow(DataFlow): 'sinks': {'key': 'typeProperties.sinks', 'type': '[DataFlowSink]'}, 'transformations': {'key': 'typeProperties.transformations', 'type': '[Transformation]'}, 'script': {'key': 'typeProperties.script', 'type': 'str'}, + 'script_lines': {'key': 'typeProperties.scriptLines', 'type': '[str]'}, } def __init__( @@ -25790,6 +25892,7 @@ def __init__( sinks: Optional[List["DataFlowSink"]] = None, transformations: Optional[List["Transformation"]] = None, script: Optional[str] = None, + script_lines: Optional[List[str]] = None, **kwargs ): super(MappingDataFlow, self).__init__(description=description, annotations=annotations, folder=folder, **kwargs) @@ -25798,6 +25901,7 @@ def __init__( self.sinks = sinks self.transformations = transformations self.script = script + self.script_lines = script_lines class MariaDBLinkedService(LinkedService): @@ -31535,6 +31639,8 @@ class PowerQuerySink(DataFlowSink): :type dataset: ~azure.mgmt.datafactory.models.DatasetReference :param linked_service: Linked service reference. :type linked_service: ~azure.mgmt.datafactory.models.LinkedServiceReference + :param flowlet: Flowlet Reference. + :type flowlet: ~azure.mgmt.datafactory.models.DataFlowReference :param schema_linked_service: Schema linked service reference. :type schema_linked_service: ~azure.mgmt.datafactory.models.LinkedServiceReference :param script: sink script. @@ -31550,6 +31656,7 @@ class PowerQuerySink(DataFlowSink): 'description': {'key': 'description', 'type': 'str'}, 'dataset': {'key': 'dataset', 'type': 'DatasetReference'}, 'linked_service': {'key': 'linkedService', 'type': 'LinkedServiceReference'}, + 'flowlet': {'key': 'flowlet', 'type': 'DataFlowReference'}, 'schema_linked_service': {'key': 'schemaLinkedService', 'type': 'LinkedServiceReference'}, 'script': {'key': 'script', 'type': 'str'}, } @@ -31561,11 +31668,12 @@ def __init__( description: Optional[str] = None, dataset: Optional["DatasetReference"] = None, linked_service: Optional["LinkedServiceReference"] = None, + flowlet: Optional["DataFlowReference"] = None, schema_linked_service: Optional["LinkedServiceReference"] = None, script: Optional[str] = None, **kwargs ): - super(PowerQuerySink, self).__init__(name=name, description=description, dataset=dataset, linked_service=linked_service, schema_linked_service=schema_linked_service, **kwargs) + super(PowerQuerySink, self).__init__(name=name, description=description, dataset=dataset, linked_service=linked_service, flowlet=flowlet, schema_linked_service=schema_linked_service, **kwargs) self.script = script @@ -31608,6 +31716,8 @@ class PowerQuerySource(DataFlowSource): :type dataset: ~azure.mgmt.datafactory.models.DatasetReference :param linked_service: Linked service reference. :type linked_service: ~azure.mgmt.datafactory.models.LinkedServiceReference + :param flowlet: Flowlet Reference. + :type flowlet: ~azure.mgmt.datafactory.models.DataFlowReference :param schema_linked_service: Schema linked service reference. :type schema_linked_service: ~azure.mgmt.datafactory.models.LinkedServiceReference :param script: source script. @@ -31623,6 +31733,7 @@ class PowerQuerySource(DataFlowSource): 'description': {'key': 'description', 'type': 'str'}, 'dataset': {'key': 'dataset', 'type': 'DatasetReference'}, 'linked_service': {'key': 'linkedService', 'type': 'LinkedServiceReference'}, + 'flowlet': {'key': 'flowlet', 'type': 'DataFlowReference'}, 'schema_linked_service': {'key': 'schemaLinkedService', 'type': 'LinkedServiceReference'}, 'script': {'key': 'script', 'type': 'str'}, } @@ -31634,11 +31745,12 @@ def __init__( description: Optional[str] = None, dataset: Optional["DatasetReference"] = None, linked_service: Optional["LinkedServiceReference"] = None, + flowlet: Optional["DataFlowReference"] = None, schema_linked_service: Optional["LinkedServiceReference"] = None, script: Optional[str] = None, **kwargs ): - super(PowerQuerySource, self).__init__(name=name, description=description, dataset=dataset, linked_service=linked_service, schema_linked_service=schema_linked_service, **kwargs) + super(PowerQuerySource, self).__init__(name=name, description=description, dataset=dataset, linked_service=linked_service, flowlet=flowlet, schema_linked_service=schema_linked_service, **kwargs) self.script = script