diff --git a/eng/mgmt/mgmtmetadata/datafactory_resource-manager.txt b/eng/mgmt/mgmtmetadata/datafactory_resource-manager.txt index e45dc0b395e1..4c578db3441a 100644 --- a/eng/mgmt/mgmtmetadata/datafactory_resource-manager.txt +++ b/eng/mgmt/mgmtmetadata/datafactory_resource-manager.txt @@ -5,11 +5,11 @@ Generating CSharp code Executing AutoRest command cmd.exe /c autorest.cmd https://github.com/Azure/azure-rest-api-specs/blob/main/specification/datafactory/resource-manager/readme.md --csharp --version=v2 --reflect-api-versions --tag=package-2018-06 --csharp-sdks-folder=D:\Projects\azure-sdk-for-net\sdk Autorest CSharp Version: 2.3.82 -2022-09-30 02:08:47 UTC +2022-10-25 04:48:50 UTC Azure-rest-api-specs repository information GitHub fork: Azure Branch: main -Commit: 4994cbed850f3726721ec6fd3235a474e8d08fcc +Commit: fdd4e5c9b9225698c7f26c75c4b26be5c57e60f8 AutoRest information Requested version: v2 Bootstrapper version: autorest@3.6.1 diff --git a/sdk/datafactory/Microsoft.Azure.Management.DataFactory/src/CHANGELOG.md b/sdk/datafactory/Microsoft.Azure.Management.DataFactory/src/CHANGELOG.md index ff433fdb98ce..b3fd8dc11908 100644 --- a/sdk/datafactory/Microsoft.Azure.Management.DataFactory/src/CHANGELOG.md +++ b/sdk/datafactory/Microsoft.Azure.Management.DataFactory/src/CHANGELOG.md @@ -1,5 +1,12 @@ # Changelog for the Azure Data Factory V2 .NET SDK +## Version 8.0.0 +### Feature Additions +### Breaking Changes +- Added properties to spark job activity +- Added disablePublish property into FactoryRepoConfiguration +- Added scriptBlockExecutionTimeout property into Script Activity + ## Version 7.0.0 ### Feature Additions ### Breaking Changes diff --git a/sdk/datafactory/Microsoft.Azure.Management.DataFactory/src/Customizations/FactoryGitHubConfiguration.cs b/sdk/datafactory/Microsoft.Azure.Management.DataFactory/src/Customizations/FactoryGitHubConfiguration.cs new file mode 100644 index 000000000000..b3a6a5fdad95 --- /dev/null +++ b/sdk/datafactory/Microsoft.Azure.Management.DataFactory/src/Customizations/FactoryGitHubConfiguration.cs @@ -0,0 +1,28 @@ +namespace Microsoft.Azure.Management.DataFactory.Models +{ + public partial class FactoryGitHubConfiguration : FactoryRepoConfiguration + { + /// + /// Initializes a new instance of the FactoryGitHubConfiguration class. + /// + /// Account name. + /// Repository name. + /// Collaboration branch. + /// Root folder. + /// Last commit id. + /// ADF studio to favor automated publish. + /// GitHub Enterprise host name. For example: + /// `https://github.mydomain.com` + /// GitHub bring your own app client id. + /// GitHub bring your own app client secret + /// information. + public FactoryGitHubConfiguration(string accountName, string repositoryName, string collaborationBranch, string rootFolder, string lastCommitId, string hostName, string clientId = default(string), GitHubClientSecret clientSecret = default(GitHubClientSecret)) + : base(accountName, repositoryName, collaborationBranch, rootFolder, lastCommitId) + { + HostName = hostName; + ClientId = clientId; + ClientSecret = clientSecret; + CustomInit(); + } + } +} diff --git a/sdk/datafactory/Microsoft.Azure.Management.DataFactory/src/Customizations/FactoryVSTSConfiguration.cs b/sdk/datafactory/Microsoft.Azure.Management.DataFactory/src/Customizations/FactoryVSTSConfiguration.cs new file mode 100644 index 000000000000..3c7de53048ae --- /dev/null +++ b/sdk/datafactory/Microsoft.Azure.Management.DataFactory/src/Customizations/FactoryVSTSConfiguration.cs @@ -0,0 +1,23 @@ +namespace Microsoft.Azure.Management.DataFactory.Models +{ + public partial class FactoryVSTSConfiguration : FactoryRepoConfiguration + { + /// + /// Initializes a new instance of the FactoryVSTSConfiguration class. + /// + /// Account name. + /// Repository name. + /// Collaboration branch. + /// Root folder. + /// VSTS project name. + /// Last commit id. + /// VSTS tenant id. + public FactoryVSTSConfiguration(string accountName, string repositoryName, string collaborationBranch, string rootFolder, string projectName, string lastCommitId, string tenantId) + : base(accountName, repositoryName, collaborationBranch, rootFolder, lastCommitId) + { + ProjectName = projectName; + TenantId = tenantId; + CustomInit(); + } + } +} diff --git a/sdk/datafactory/Microsoft.Azure.Management.DataFactory/src/Customizations/ScriptActivity.cs b/sdk/datafactory/Microsoft.Azure.Management.DataFactory/src/Customizations/ScriptActivity.cs new file mode 100644 index 000000000000..7cbba1436378 --- /dev/null +++ b/sdk/datafactory/Microsoft.Azure.Management.DataFactory/src/Customizations/ScriptActivity.cs @@ -0,0 +1,28 @@ +using System.Collections.Generic; + +namespace Microsoft.Azure.Management.DataFactory.Models +{ + public partial class ScriptActivity : ExecutionActivity + { + /// + /// Initializes a new instance of the ScriptActivity class. + /// + /// Activity name. + /// Unmatched properties from the + /// message are deserialized this collection + /// Activity description. + /// Activity depends on condition. + /// Activity user properties. + /// Linked service reference. + /// Activity policy. + /// Array of script blocks. Type: array. + /// Log settings of script activity. + public ScriptActivity(string name, IDictionary additionalProperties, string description, IList dependsOn, IList userProperties, LinkedServiceReference linkedServiceName, ActivityPolicy policy, IList scripts, ScriptActivityTypePropertiesLogSettings logSettings = default(ScriptActivityTypePropertiesLogSettings)) + : base(name, additionalProperties, description, dependsOn, userProperties, linkedServiceName, policy) + { + Scripts = scripts; + LogSettings = logSettings; + CustomInit(); + } + } +} diff --git a/sdk/datafactory/Microsoft.Azure.Management.DataFactory/src/Customizations/SynapseSparkJobDefinitionActivity.cs b/sdk/datafactory/Microsoft.Azure.Management.DataFactory/src/Customizations/SynapseSparkJobDefinitionActivity.cs new file mode 100644 index 000000000000..31462f7b4d8c --- /dev/null +++ b/sdk/datafactory/Microsoft.Azure.Management.DataFactory/src/Customizations/SynapseSparkJobDefinitionActivity.cs @@ -0,0 +1,68 @@ +using System.Collections.Generic; + +namespace Microsoft.Azure.Management.DataFactory.Models +{ + public partial class SynapseSparkJobDefinitionActivity : ExecutionActivity + { + /// + /// Initializes a new instance of the SynapseSparkJobDefinitionActivity + /// class. + /// + /// Activity name. + /// Synapse spark job reference. + /// Unmatched properties from the + /// message are deserialized this collection + /// Activity description. + /// Activity depends on condition. + /// Activity user properties. + /// Linked service reference. + /// Activity policy. + /// User specified arguments to + /// SynapseSparkJobDefinitionActivity. + /// The main file used for the job, which will + /// override the 'file' of the spark job definition you provide. Type: + /// string (or Expression with resultType string). + /// The fully-qualified identifier or the main + /// class that is in the main definition file, which will override the + /// 'className' of the spark job definition you provide. Type: string + /// (or Expression with resultType string). + /// Additional files used for reference in the main + /// definition file, which will override the 'files' of the spark job + /// definition you provide. + /// The name of the big data pool which + /// will be used to execute the spark batch job, which will override + /// the 'targetBigDataPool' of the spark job definition you + /// provide. + /// Number of core and memory to be used for + /// executors allocated in the specified Spark pool for the job, which + /// will be used for overriding 'executorCores' and 'executorMemory' of + /// the spark job definition you provide. Type: string (or Expression + /// with resultType string). + /// Spark configuration properties, which will + /// override the 'conf' of the spark job definition you + /// provide. + /// Number of core and memory to be used for + /// driver allocated in the specified Spark pool for the job, which + /// will be used for overriding 'driverCores' and 'driverMemory' of the + /// spark job definition you provide. Type: string (or Expression with + /// resultType string). + /// Number of executors to launch for this + /// job, which will override the 'numExecutors' of the spark job + /// definition you provide. + public SynapseSparkJobDefinitionActivity(string name, SynapseSparkJobReference sparkJob, IDictionary additionalProperties, string description, IList dependsOn, IList userProperties, LinkedServiceReference linkedServiceName, ActivityPolicy policy, IList arguments, object file, object className, IList files, BigDataPoolParametrizationReference targetBigDataPool, object executorSize = default(object), object conf = default(object), object driverSize = default(object), int? numExecutors = default(int?)) + : base(name, additionalProperties, description, dependsOn, userProperties, linkedServiceName, policy) + { + SparkJob = sparkJob; + Arguments = arguments; + File = file; + ClassName = className; + Files = files; + TargetBigDataPool = targetBigDataPool; + ExecutorSize = executorSize; + Conf = conf; + DriverSize = driverSize; + NumExecutors = numExecutors; + CustomInit(); + } + } +} diff --git a/sdk/datafactory/Microsoft.Azure.Management.DataFactory/src/Generated/Models/AzureSynapseArtifactsLinkedService.cs b/sdk/datafactory/Microsoft.Azure.Management.DataFactory/src/Generated/Models/AzureSynapseArtifactsLinkedService.cs index e2cc0fd29de8..f35d26514789 100644 --- a/sdk/datafactory/Microsoft.Azure.Management.DataFactory/src/Generated/Models/AzureSynapseArtifactsLinkedService.cs +++ b/sdk/datafactory/Microsoft.Azure.Management.DataFactory/src/Generated/Models/AzureSynapseArtifactsLinkedService.cs @@ -51,11 +51,16 @@ public AzureSynapseArtifactsLinkedService() /// Required to specify MSI, if using /// system assigned managed identity as authentication method. Type: /// string (or Expression with resultType string). - public AzureSynapseArtifactsLinkedService(object endpoint, IDictionary additionalProperties = default(IDictionary), IntegrationRuntimeReference connectVia = default(IntegrationRuntimeReference), string description = default(string), IDictionary parameters = default(IDictionary), IList annotations = default(IList), object authentication = default(object)) + /// The resource ID of the Synapse + /// workspace. The format should be: + /// /subscriptions/{subscriptionID}/resourceGroups/{resourceGroup}/providers/Microsoft.Synapse/workspaces/{workspaceName}. + /// Type: string (or Expression with resultType string). + public AzureSynapseArtifactsLinkedService(object endpoint, IDictionary additionalProperties = default(IDictionary), IntegrationRuntimeReference connectVia = default(IntegrationRuntimeReference), string description = default(string), IDictionary parameters = default(IDictionary), IList annotations = default(IList), object authentication = default(object), object workspaceResourceId = default(object)) : base(additionalProperties, connectVia, description, parameters, annotations) { Endpoint = endpoint; Authentication = authentication; + WorkspaceResourceId = workspaceResourceId; CustomInit(); } @@ -81,6 +86,15 @@ public AzureSynapseArtifactsLinkedService() [JsonProperty(PropertyName = "typeProperties.authentication")] public object Authentication { get; set; } + /// + /// Gets or sets the resource ID of the Synapse workspace. The format + /// should be: + /// /subscriptions/{subscriptionID}/resourceGroups/{resourceGroup}/providers/Microsoft.Synapse/workspaces/{workspaceName}. + /// Type: string (or Expression with resultType string). + /// + [JsonProperty(PropertyName = "typeProperties.workspaceResourceId")] + public object WorkspaceResourceId { get; set; } + /// /// Validate the object. /// diff --git a/sdk/datafactory/Microsoft.Azure.Management.DataFactory/src/Generated/Models/FactoryGitHubConfiguration.cs b/sdk/datafactory/Microsoft.Azure.Management.DataFactory/src/Generated/Models/FactoryGitHubConfiguration.cs index 8722faac6464..9921f38eac7b 100644 --- a/sdk/datafactory/Microsoft.Azure.Management.DataFactory/src/Generated/Models/FactoryGitHubConfiguration.cs +++ b/sdk/datafactory/Microsoft.Azure.Management.DataFactory/src/Generated/Models/FactoryGitHubConfiguration.cs @@ -34,13 +34,15 @@ public FactoryGitHubConfiguration() /// Collaboration branch. /// Root folder. /// Last commit id. + /// Disable manual publish operation in + /// ADF studio to favor automated publish. /// GitHub Enterprise host name. For example: /// `https://github.mydomain.com` /// GitHub bring your own app client id. /// GitHub bring your own app client secret /// information. - public FactoryGitHubConfiguration(string accountName, string repositoryName, string collaborationBranch, string rootFolder, string lastCommitId = default(string), string hostName = default(string), string clientId = default(string), GitHubClientSecret clientSecret = default(GitHubClientSecret)) - : base(accountName, repositoryName, collaborationBranch, rootFolder, lastCommitId) + public FactoryGitHubConfiguration(string accountName, string repositoryName, string collaborationBranch, string rootFolder, string lastCommitId = default(string), bool? disablePublish = default(bool?), string hostName = default(string), string clientId = default(string), GitHubClientSecret clientSecret = default(GitHubClientSecret)) + : base(accountName, repositoryName, collaborationBranch, rootFolder, lastCommitId, disablePublish) { HostName = hostName; ClientId = clientId; diff --git a/sdk/datafactory/Microsoft.Azure.Management.DataFactory/src/Generated/Models/FactoryRepoConfiguration.cs b/sdk/datafactory/Microsoft.Azure.Management.DataFactory/src/Generated/Models/FactoryRepoConfiguration.cs index 539d782b3939..f0fe1ac4e56a 100644 --- a/sdk/datafactory/Microsoft.Azure.Management.DataFactory/src/Generated/Models/FactoryRepoConfiguration.cs +++ b/sdk/datafactory/Microsoft.Azure.Management.DataFactory/src/Generated/Models/FactoryRepoConfiguration.cs @@ -35,13 +35,16 @@ public FactoryRepoConfiguration() /// Collaboration branch. /// Root folder. /// Last commit id. - public FactoryRepoConfiguration(string accountName, string repositoryName, string collaborationBranch, string rootFolder, string lastCommitId = default(string)) + /// Disable manual publish operation in + /// ADF studio to favor automated publish. + public FactoryRepoConfiguration(string accountName, string repositoryName, string collaborationBranch, string rootFolder, string lastCommitId = default(string), bool? disablePublish = default(bool?)) { AccountName = accountName; RepositoryName = repositoryName; CollaborationBranch = collaborationBranch; RootFolder = rootFolder; LastCommitId = lastCommitId; + DisablePublish = disablePublish; CustomInit(); } @@ -80,6 +83,13 @@ public FactoryRepoConfiguration() [JsonProperty(PropertyName = "lastCommitId")] public string LastCommitId { get; set; } + /// + /// Gets or sets disable manual publish operation in ADF studio to + /// favor automated publish. + /// + [JsonProperty(PropertyName = "disablePublish")] + public bool? DisablePublish { get; set; } + /// /// Validate the object. /// diff --git a/sdk/datafactory/Microsoft.Azure.Management.DataFactory/src/Generated/Models/FactoryVSTSConfiguration.cs b/sdk/datafactory/Microsoft.Azure.Management.DataFactory/src/Generated/Models/FactoryVSTSConfiguration.cs index 0b5f5a1c67c8..62dd94b309db 100644 --- a/sdk/datafactory/Microsoft.Azure.Management.DataFactory/src/Generated/Models/FactoryVSTSConfiguration.cs +++ b/sdk/datafactory/Microsoft.Azure.Management.DataFactory/src/Generated/Models/FactoryVSTSConfiguration.cs @@ -36,9 +36,11 @@ public FactoryVSTSConfiguration() /// Root folder. /// VSTS project name. /// Last commit id. + /// Disable manual publish operation in + /// ADF studio to favor automated publish. /// VSTS tenant id. - public FactoryVSTSConfiguration(string accountName, string repositoryName, string collaborationBranch, string rootFolder, string projectName, string lastCommitId = default(string), string tenantId = default(string)) - : base(accountName, repositoryName, collaborationBranch, rootFolder, lastCommitId) + public FactoryVSTSConfiguration(string accountName, string repositoryName, string collaborationBranch, string rootFolder, string projectName, string lastCommitId = default(string), bool? disablePublish = default(bool?), string tenantId = default(string)) + : base(accountName, repositoryName, collaborationBranch, rootFolder, lastCommitId, disablePublish) { ProjectName = projectName; TenantId = tenantId; diff --git a/sdk/datafactory/Microsoft.Azure.Management.DataFactory/src/Generated/Models/ScriptActivity.cs b/sdk/datafactory/Microsoft.Azure.Management.DataFactory/src/Generated/Models/ScriptActivity.cs index 248ef5440521..7892f8512499 100644 --- a/sdk/datafactory/Microsoft.Azure.Management.DataFactory/src/Generated/Models/ScriptActivity.cs +++ b/sdk/datafactory/Microsoft.Azure.Management.DataFactory/src/Generated/Models/ScriptActivity.cs @@ -43,11 +43,16 @@ public ScriptActivity() /// Activity user properties. /// Linked service reference. /// Activity policy. + /// ScriptBlock execution + /// timeout. Type: string (or Expression with resultType string), + /// pattern: + /// ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). /// Array of script blocks. Type: array. /// Log settings of script activity. - public ScriptActivity(string name, IDictionary additionalProperties = default(IDictionary), string description = default(string), IList dependsOn = default(IList), IList userProperties = default(IList), LinkedServiceReference linkedServiceName = default(LinkedServiceReference), ActivityPolicy policy = default(ActivityPolicy), IList scripts = default(IList), ScriptActivityTypePropertiesLogSettings logSettings = default(ScriptActivityTypePropertiesLogSettings)) + public ScriptActivity(string name, IDictionary additionalProperties = default(IDictionary), string description = default(string), IList dependsOn = default(IList), IList userProperties = default(IList), LinkedServiceReference linkedServiceName = default(LinkedServiceReference), ActivityPolicy policy = default(ActivityPolicy), object scriptBlockExecutionTimeout = default(object), IList scripts = default(IList), ScriptActivityTypePropertiesLogSettings logSettings = default(ScriptActivityTypePropertiesLogSettings)) : base(name, additionalProperties, description, dependsOn, userProperties, linkedServiceName, policy) { + ScriptBlockExecutionTimeout = scriptBlockExecutionTimeout; Scripts = scripts; LogSettings = logSettings; CustomInit(); @@ -58,6 +63,14 @@ public ScriptActivity() /// partial void CustomInit(); + /// + /// Gets or sets scriptBlock execution timeout. Type: string (or + /// Expression with resultType string), pattern: + /// ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + /// + [JsonProperty(PropertyName = "typeProperties.scriptBlockExecutionTimeout")] + public object ScriptBlockExecutionTimeout { get; set; } + /// /// Gets or sets array of script blocks. Type: array. /// diff --git a/sdk/datafactory/Microsoft.Azure.Management.DataFactory/src/Generated/Models/SynapseSparkJobDefinitionActivity.cs b/sdk/datafactory/Microsoft.Azure.Management.DataFactory/src/Generated/Models/SynapseSparkJobDefinitionActivity.cs index c427ef65ab77..e39a5a1f3c50 100644 --- a/sdk/datafactory/Microsoft.Azure.Management.DataFactory/src/Generated/Models/SynapseSparkJobDefinitionActivity.cs +++ b/sdk/datafactory/Microsoft.Azure.Management.DataFactory/src/Generated/Models/SynapseSparkJobDefinitionActivity.cs @@ -56,9 +56,16 @@ public SynapseSparkJobDefinitionActivity() /// class that is in the main definition file, which will override the /// 'className' of the spark job definition you provide. Type: string /// (or Expression with resultType string). - /// Additional files used for reference in the main - /// definition file, which will override the 'files' of the spark job - /// definition you provide. + /// (Deprecated. Please use pythonCodeReference and + /// filesV2) Additional files used for reference in the main definition + /// file, which will override the 'files' of the spark job definition + /// you provide. + /// Additional python code files used + /// for reference in the main definition file, which will override the + /// 'pyFiles' of the spark job definition you provide. + /// Additional files used for reference in the + /// main definition file, which will override the 'jars' and 'files' of + /// the spark job definition you provide. /// The name of the big data pool which /// will be used to execute the spark batch job, which will override /// the 'targetBigDataPool' of the spark job definition you @@ -79,7 +86,7 @@ public SynapseSparkJobDefinitionActivity() /// Number of executors to launch for this /// job, which will override the 'numExecutors' of the spark job /// definition you provide. - public SynapseSparkJobDefinitionActivity(string name, SynapseSparkJobReference sparkJob, IDictionary additionalProperties = default(IDictionary), string description = default(string), IList dependsOn = default(IList), IList userProperties = default(IList), LinkedServiceReference linkedServiceName = default(LinkedServiceReference), ActivityPolicy policy = default(ActivityPolicy), IList arguments = default(IList), object file = default(object), object className = default(object), IList files = default(IList), BigDataPoolParametrizationReference targetBigDataPool = default(BigDataPoolParametrizationReference), object executorSize = default(object), object conf = default(object), object driverSize = default(object), int? numExecutors = default(int?)) + public SynapseSparkJobDefinitionActivity(string name, SynapseSparkJobReference sparkJob, IDictionary additionalProperties = default(IDictionary), string description = default(string), IList dependsOn = default(IList), IList userProperties = default(IList), LinkedServiceReference linkedServiceName = default(LinkedServiceReference), ActivityPolicy policy = default(ActivityPolicy), IList arguments = default(IList), object file = default(object), object className = default(object), IList files = default(IList), IList pythonCodeReference = default(IList), IList filesV2 = default(IList), BigDataPoolParametrizationReference targetBigDataPool = default(BigDataPoolParametrizationReference), object executorSize = default(object), object conf = default(object), object driverSize = default(object), int? numExecutors = default(int?)) : base(name, additionalProperties, description, dependsOn, userProperties, linkedServiceName, policy) { SparkJob = sparkJob; @@ -87,6 +94,8 @@ public SynapseSparkJobDefinitionActivity() File = file; ClassName = className; Files = files; + PythonCodeReference = pythonCodeReference; + FilesV2 = filesV2; TargetBigDataPool = targetBigDataPool; ExecutorSize = executorSize; Conf = conf; @@ -131,13 +140,30 @@ public SynapseSparkJobDefinitionActivity() public object ClassName { get; set; } /// - /// Gets or sets additional files used for reference in the main - /// definition file, which will override the 'files' of the spark job - /// definition you provide. + /// Gets or sets (Deprecated. Please use pythonCodeReference and + /// filesV2) Additional files used for reference in the main definition + /// file, which will override the 'files' of the spark job definition + /// you provide. /// [JsonProperty(PropertyName = "typeProperties.files")] public IList Files { get; set; } + /// + /// Gets or sets additional python code files used for reference in the + /// main definition file, which will override the 'pyFiles' of the + /// spark job definition you provide. + /// + [JsonProperty(PropertyName = "typeProperties.pythonCodeReference")] + public IList PythonCodeReference { get; set; } + + /// + /// Gets or sets additional files used for reference in the main + /// definition file, which will override the 'jars' and 'files' of the + /// spark job definition you provide. + /// + [JsonProperty(PropertyName = "typeProperties.filesV2")] + public IList FilesV2 { get; set; } + /// /// Gets or sets the name of the big data pool which will be used to /// execute the spark batch job, which will override the diff --git a/sdk/datafactory/Microsoft.Azure.Management.DataFactory/src/Generated/Models/SynapseSparkJobReference.cs b/sdk/datafactory/Microsoft.Azure.Management.DataFactory/src/Generated/Models/SynapseSparkJobReference.cs index 35872c2c622e..702ad4738d2e 100644 --- a/sdk/datafactory/Microsoft.Azure.Management.DataFactory/src/Generated/Models/SynapseSparkJobReference.cs +++ b/sdk/datafactory/Microsoft.Azure.Management.DataFactory/src/Generated/Models/SynapseSparkJobReference.cs @@ -30,8 +30,9 @@ public SynapseSparkJobReference() /// /// Initializes a new instance of the SynapseSparkJobReference class. /// - /// Reference spark job name. - public SynapseSparkJobReference(string referenceName) + /// Reference spark job name. Expression + /// with resultType string. + public SynapseSparkJobReference(object referenceName) { ReferenceName = referenceName; CustomInit(); @@ -50,10 +51,11 @@ static SynapseSparkJobReference() partial void CustomInit(); /// - /// Gets or sets reference spark job name. + /// Gets or sets reference spark job name. Expression with resultType + /// string. /// [JsonProperty(PropertyName = "referenceName")] - public string ReferenceName { get; set; } + public object ReferenceName { get; set; } /// /// Synapse spark job reference type. diff --git a/sdk/datafactory/Microsoft.Azure.Management.DataFactory/src/Microsoft.Azure.Management.DataFactory.csproj b/sdk/datafactory/Microsoft.Azure.Management.DataFactory/src/Microsoft.Azure.Management.DataFactory.csproj index 519c86b301be..ba843ed7aea0 100644 --- a/sdk/datafactory/Microsoft.Azure.Management.DataFactory/src/Microsoft.Azure.Management.DataFactory.csproj +++ b/sdk/datafactory/Microsoft.Azure.Management.DataFactory/src/Microsoft.Azure.Management.DataFactory.csproj @@ -5,12 +5,14 @@ Microsoft.Azure.Management.DataFactory Azure Data Factory V2 is the data integration platform that goes beyond Azure Data Factory V1's orchestration and batch-processing of time-series data, with a general purpose app model supporting modern data warehousing patterns and scenarios, lift-and-shift SSIS, and data-driven SaaS applications. Compose and manage reliable and secure data integration workflows at scale. Use native ADF data connectors and Integration Runtimes to move and transform cloud and on-premises data that can be unstructured, semi-structured, and structured with Hadoop, Azure Data Lake, Spark, SQL Server, Cosmos DB and many other data platforms. - 7.0.0 + 8.0.0 Microsoft.Azure.Management.DataFactory Microsoft Azure resource management;Data Factory;ADF; diff --git a/sdk/datafactory/Microsoft.Azure.Management.DataFactory/src/Properties/AssemblyInfo.cs b/sdk/datafactory/Microsoft.Azure.Management.DataFactory/src/Properties/AssemblyInfo.cs index b41427aab773..d49f92a18d78 100644 --- a/sdk/datafactory/Microsoft.Azure.Management.DataFactory/src/Properties/AssemblyInfo.cs +++ b/sdk/datafactory/Microsoft.Azure.Management.DataFactory/src/Properties/AssemblyInfo.cs @@ -6,8 +6,8 @@ [assembly: AssemblyTitle("Microsoft Azure Data Factory Management Library")] [assembly: AssemblyDescription("Provides management functionality for Microsoft Azure Data Factory Resources.")] -[assembly: AssemblyVersion("7.0.0.0")] -[assembly: AssemblyFileVersion("7.0.0.0")] +[assembly: AssemblyVersion("8.0.0.0")] +[assembly: AssemblyFileVersion("8.0.0.0")] [assembly: AssemblyConfiguration("")] [assembly: AssemblyCompany("Microsoft")] [assembly: AssemblyProduct("Microsoft Azure .NET SDK")] diff --git a/sdk/datafactory/Microsoft.Azure.Management.DataFactory/tests/JsonSamples/PipelineJsonSamples.cs b/sdk/datafactory/Microsoft.Azure.Management.DataFactory/tests/JsonSamples/PipelineJsonSamples.cs index d5b3448d9b78..49510a3a2b07 100644 --- a/sdk/datafactory/Microsoft.Azure.Management.DataFactory/tests/JsonSamples/PipelineJsonSamples.cs +++ b/sdk/datafactory/Microsoft.Azure.Management.DataFactory/tests/JsonSamples/PipelineJsonSamples.cs @@ -7604,6 +7604,7 @@ public class PipelineJsonSamples : JsonSampleCollection ""type"": ""LinkedServiceReference"" }, ""typeProperties"": { + ""scriptBlockExecutionTimeout"": ""12:00:00"", ""scripts"": [ { ""text"": ""@pipeline().parameters.query"",