You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
/** Link connection status, please refer to this [articles](https://learn.microsoft.com/azure/synapse-analytics/synapse-link/sql-database-synapse-link#monitoring) for details. */
681
682
status?: string;
682
683
/** Link connection's corresponding continuous run id */
/** Link table status, please refer to this [articles](https://learn.microsoft.com/azure/synapse-analytics/synapse-link/sql-database-synapse-link#monitoring) for details. */
/** Required to specify MSI, if using system assigned managed identity as authentication method. Type: string (or Expression with resultType string). */
7093
7105
authentication?: any;
7106
+
/** The resource ID of the Synapse workspace. The format should be: /subscriptions/{subscriptionID}/resourceGroups/{resourceGroup}/providers/Microsoft.Synapse/workspaces/{workspaceName}. Type: string (or Expression with resultType string). */
/** The main file used for the job, which will override the 'file' of the spark job definition you provide. Type: string (or Expression with resultType string). */
9622
9646
file?: any;
9647
+
/** Scanning subfolders from the root folder of the main definition file, these files will be added as reference files. The folders named 'jars', 'pyFiles', 'files' or 'archives' will be scanned, and the folders name are case sensitive. Type: boolean (or Expression with resultType boolean). */
9648
+
scanFolder?: any;
9623
9649
/** The fully-qualified identifier or the main class that is in the main definition file, which will override the 'className' of the spark job definition you provide. Type: string (or Expression with resultType string). */
9624
9650
className?: any;
9625
-
/** Additional files used for reference in the main definition file, which will override the 'files' of the spark job definition you provide. */
9651
+
/** (Deprecated. Please use pythonCodeReference and filesV2) Additional files used for reference in the main definition file, which will override the 'files' of the spark job definition you provide. */
9626
9652
files?: any[];
9653
+
/** Additional python code files used for reference in the main definition file, which will override the 'pyFiles' of the spark job definition you provide. */
9654
+
pythonCodeReference?: any[];
9655
+
/** Additional files used for reference in the main definition file, which will override the 'jars' and 'files' of the spark job definition you provide. */
9656
+
filesV2?: any[];
9627
9657
/** The name of the big data pool which will be used to execute the spark batch job, which will override the 'targetBigDataPool' of the spark job definition you provide. */
/** Number of core and memory to be used for executors allocated in the specified Spark pool for the job, which will be used for overriding 'executorCores' and 'executorMemory' of the spark job definition you provide. Type: string (or Expression with resultType string). */
/** Number of core and memory to be used for driver allocated in the specified Spark pool for the job, which will be used for overriding 'driverCores' and 'driverMemory' of the spark job definition you provide. Type: string (or Expression with resultType string). */
9634
9664
driverSize?: any;
9635
-
/** Number of executors to launch for this job, which will override the 'numExecutors' of the spark job definition you provide. */
9636
-
numExecutors?: number;
9665
+
/** Number of executors to launch for this job, which will override the 'numExecutors' of the spark job definition you provide. Type: integer (or Expression with resultType integer). */
0 commit comments