diff --git a/sdk/datafactory/Azure.ResourceManager.DataFactory/api/Azure.ResourceManager.DataFactory.netstandard2.0.cs b/sdk/datafactory/Azure.ResourceManager.DataFactory/api/Azure.ResourceManager.DataFactory.netstandard2.0.cs index b57bb6c2dbe2..15e4fecb01b6 100644 --- a/sdk/datafactory/Azure.ResourceManager.DataFactory/api/Azure.ResourceManager.DataFactory.netstandard2.0.cs +++ b/sdk/datafactory/Azure.ResourceManager.DataFactory/api/Azure.ResourceManager.DataFactory.netstandard2.0.cs @@ -2584,9 +2584,11 @@ public partial class AzureTableStorageLinkedService : Azure.ResourceManager.Data public AzureTableStorageLinkedService() { } public Azure.Core.Expressions.DataFactory.DataFactoryKeyVaultSecret AccountKey { get { throw null; } set { } } public Azure.Core.Expressions.DataFactory.DataFactoryElement ConnectionString { get { throw null; } set { } } + public Azure.ResourceManager.DataFactory.Models.DataFactoryCredentialReference Credential { get { throw null; } set { } } public string EncryptedCredential { get { throw null; } set { } } public Azure.Core.Expressions.DataFactory.DataFactoryKeyVaultSecret SasToken { get { throw null; } set { } } public Azure.Core.Expressions.DataFactory.DataFactoryElement SasUri { get { throw null; } set { } } + public Azure.Core.Expressions.DataFactory.DataFactoryElement ServiceEndpoint { get { throw null; } set { } } Azure.ResourceManager.DataFactory.Models.AzureTableStorageLinkedService System.ClientModel.Primitives.IJsonModel.Create(ref System.Text.Json.Utf8JsonReader reader, System.ClientModel.Primitives.ModelReaderWriterOptions options) { throw null; } void System.ClientModel.Primitives.IJsonModel.Write(System.Text.Json.Utf8JsonWriter writer, System.ClientModel.Primitives.ModelReaderWriterOptions options) { } Azure.ResourceManager.DataFactory.Models.AzureTableStorageLinkedService System.ClientModel.Primitives.IPersistableModel.Create(System.BinaryData data, System.ClientModel.Primitives.ModelReaderWriterOptions options) { throw null; } @@ -2874,6 +2876,18 @@ public ConnectionStateProperties() { } string System.ClientModel.Primitives.IPersistableModel.GetFormatFromOptions(System.ClientModel.Primitives.ModelReaderWriterOptions options) { throw null; } System.BinaryData System.ClientModel.Primitives.IPersistableModel.Write(System.ClientModel.Primitives.ModelReaderWriterOptions options) { throw null; } } + public partial class ContinuationSettingsReference : System.ClientModel.Primitives.IJsonModel, System.ClientModel.Primitives.IPersistableModel + { + public ContinuationSettingsReference() { } + public Azure.Core.Expressions.DataFactory.DataFactoryElement ContinuationTtlInMinutes { get { throw null; } set { } } + public Azure.Core.Expressions.DataFactory.DataFactoryElement CustomizedCheckpointKey { get { throw null; } set { } } + public Azure.Core.Expressions.DataFactory.DataFactoryElement IdleCondition { get { throw null; } set { } } + Azure.ResourceManager.DataFactory.Models.ContinuationSettingsReference System.ClientModel.Primitives.IJsonModel.Create(ref System.Text.Json.Utf8JsonReader reader, System.ClientModel.Primitives.ModelReaderWriterOptions options) { throw null; } + void System.ClientModel.Primitives.IJsonModel.Write(System.Text.Json.Utf8JsonWriter writer, System.ClientModel.Primitives.ModelReaderWriterOptions options) { } + Azure.ResourceManager.DataFactory.Models.ContinuationSettingsReference System.ClientModel.Primitives.IPersistableModel.Create(System.BinaryData data, System.ClientModel.Primitives.ModelReaderWriterOptions options) { throw null; } + string System.ClientModel.Primitives.IPersistableModel.GetFormatFromOptions(System.ClientModel.Primitives.ModelReaderWriterOptions options) { throw null; } + System.BinaryData System.ClientModel.Primitives.IPersistableModel.Write(System.ClientModel.Primitives.ModelReaderWriterOptions options) { throw null; } + } public partial class ControlActivity : Azure.ResourceManager.DataFactory.Models.PipelineActivity, System.ClientModel.Primitives.IJsonModel, System.ClientModel.Primitives.IPersistableModel { public ControlActivity(string name) : base (default(string)) { } @@ -5145,6 +5159,7 @@ public partial class ExecuteDataFlowActivity : Azure.ResourceManager.DataFactory { public ExecuteDataFlowActivity(string name, Azure.ResourceManager.DataFactory.Models.DataFlowReference dataFlow) : base (default(string)) { } public Azure.ResourceManager.DataFactory.Models.ExecuteDataFlowActivityComputeType Compute { get { throw null; } set { } } + public Azure.ResourceManager.DataFactory.Models.ContinuationSettingsReference ContinuationSettings { get { throw null; } set { } } public Azure.Core.Expressions.DataFactory.DataFactoryElement ContinueOnError { get { throw null; } set { } } public Azure.ResourceManager.DataFactory.Models.DataFlowReference DataFlow { get { throw null; } set { } } public Azure.ResourceManager.DataFactory.Models.IntegrationRuntimeReference IntegrationRuntime { get { throw null; } set { } } @@ -5218,6 +5233,7 @@ public partial class ExecuteWranglingDataflowActivity : Azure.ResourceManager.Da { public ExecuteWranglingDataflowActivity(string name, Azure.ResourceManager.DataFactory.Models.DataFlowReference dataFlow) : base (default(string)) { } public Azure.ResourceManager.DataFactory.Models.ExecuteDataFlowActivityComputeType Compute { get { throw null; } set { } } + public Azure.ResourceManager.DataFactory.Models.ContinuationSettingsReference ContinuationSettings { get { throw null; } set { } } public Azure.Core.Expressions.DataFactory.DataFactoryElement ContinueOnError { get { throw null; } set { } } public Azure.ResourceManager.DataFactory.Models.DataFlowReference DataFlow { get { throw null; } set { } } public Azure.ResourceManager.DataFactory.Models.IntegrationRuntimeReference IntegrationRuntime { get { throw null; } set { } } @@ -10425,6 +10441,7 @@ public partial class SnowflakeExportCopyCommand : Azure.ResourceManager.DataFact public SnowflakeExportCopyCommand() { } public System.Collections.Generic.IDictionary AdditionalCopyOptions { get { throw null; } } public System.Collections.Generic.IDictionary AdditionalFormatOptions { get { throw null; } } + public Azure.Core.Expressions.DataFactory.DataFactoryElement StorageIntegration { get { throw null; } set { } } Azure.ResourceManager.DataFactory.Models.SnowflakeExportCopyCommand System.ClientModel.Primitives.IJsonModel.Create(ref System.Text.Json.Utf8JsonReader reader, System.ClientModel.Primitives.ModelReaderWriterOptions options) { throw null; } void System.ClientModel.Primitives.IJsonModel.Write(System.Text.Json.Utf8JsonWriter writer, System.ClientModel.Primitives.ModelReaderWriterOptions options) { } Azure.ResourceManager.DataFactory.Models.SnowflakeExportCopyCommand System.ClientModel.Primitives.IPersistableModel.Create(System.BinaryData data, System.ClientModel.Primitives.ModelReaderWriterOptions options) { throw null; } @@ -10436,6 +10453,7 @@ public partial class SnowflakeImportCopyCommand : Azure.ResourceManager.DataFact public SnowflakeImportCopyCommand() { } public System.Collections.Generic.IDictionary AdditionalCopyOptions { get { throw null; } } public System.Collections.Generic.IDictionary AdditionalFormatOptions { get { throw null; } } + public Azure.Core.Expressions.DataFactory.DataFactoryElement StorageIntegration { get { throw null; } set { } } Azure.ResourceManager.DataFactory.Models.SnowflakeImportCopyCommand System.ClientModel.Primitives.IJsonModel.Create(ref System.Text.Json.Utf8JsonReader reader, System.ClientModel.Primitives.ModelReaderWriterOptions options) { throw null; } void System.ClientModel.Primitives.IJsonModel.Write(System.Text.Json.Utf8JsonWriter writer, System.ClientModel.Primitives.ModelReaderWriterOptions options) { } Azure.ResourceManager.DataFactory.Models.SnowflakeImportCopyCommand System.ClientModel.Primitives.IPersistableModel.Create(System.BinaryData data, System.ClientModel.Primitives.ModelReaderWriterOptions options) { throw null; } @@ -10808,6 +10826,7 @@ public SqlPartitionSettings() { } private readonly int _dummyPrimitive; public SqlServerAuthenticationType(string value) { throw null; } public static Azure.ResourceManager.DataFactory.Models.SqlServerAuthenticationType Sql { get { throw null; } } + public static Azure.ResourceManager.DataFactory.Models.SqlServerAuthenticationType UserAssignedManagedIdentity { get { throw null; } } public static Azure.ResourceManager.DataFactory.Models.SqlServerAuthenticationType Windows { get { throw null; } } public bool Equals(Azure.ResourceManager.DataFactory.Models.SqlServerAuthenticationType other) { throw null; } [System.ComponentModel.EditorBrowsableAttribute(System.ComponentModel.EditorBrowsableState.Never)] @@ -10832,6 +10851,7 @@ public SqlServerLinkedService(Azure.Core.Expressions.DataFactory.DataFactoryElem public Azure.Core.Expressions.DataFactory.DataFactoryElement ConnectRetryCount { get { throw null; } set { } } public Azure.Core.Expressions.DataFactory.DataFactoryElement ConnectRetryInterval { get { throw null; } set { } } public Azure.Core.Expressions.DataFactory.DataFactoryElement ConnectTimeout { get { throw null; } set { } } + public Azure.ResourceManager.DataFactory.Models.DataFactoryCredentialReference Credential { get { throw null; } set { } } public Azure.Core.Expressions.DataFactory.DataFactoryElement Database { get { throw null; } set { } } public Azure.Core.Expressions.DataFactory.DataFactoryElement Encrypt { get { throw null; } set { } } public string EncryptedCredential { get { throw null; } set { } } diff --git a/sdk/datafactory/Azure.ResourceManager.DataFactory/src/Generated/Models/AzureTableStorageLinkedService.Serialization.cs b/sdk/datafactory/Azure.ResourceManager.DataFactory/src/Generated/Models/AzureTableStorageLinkedService.Serialization.cs index 94a5465b17bb..30f9af7ff46e 100644 --- a/sdk/datafactory/Azure.ResourceManager.DataFactory/src/Generated/Models/AzureTableStorageLinkedService.Serialization.cs +++ b/sdk/datafactory/Azure.ResourceManager.DataFactory/src/Generated/Models/AzureTableStorageLinkedService.Serialization.cs @@ -99,6 +99,16 @@ void IJsonModel.Write(Utf8JsonWriter writer, Mod writer.WritePropertyName("encryptedCredential"u8); writer.WriteStringValue(EncryptedCredential); } + if (Optional.IsDefined(ServiceEndpoint)) + { + writer.WritePropertyName("serviceEndpoint"u8); + JsonSerializer.Serialize(writer, ServiceEndpoint); + } + if (Optional.IsDefined(Credential)) + { + writer.WritePropertyName("credential"u8); + writer.WriteObjectValue(Credential, options); + } writer.WriteEndObject(); foreach (var item in AdditionalProperties) { @@ -145,6 +155,8 @@ internal static AzureTableStorageLinkedService DeserializeAzureTableStorageLinke DataFactoryElement sasUri = default; DataFactoryKeyVaultSecret sasToken = default; string encryptedCredential = default; + DataFactoryElement serviceEndpoint = default; + DataFactoryCredentialReference credential = default; IDictionary additionalProperties = default; Dictionary additionalPropertiesDictionary = new Dictionary(); foreach (var property in element.EnumerateObject()) @@ -253,6 +265,24 @@ internal static AzureTableStorageLinkedService DeserializeAzureTableStorageLinke encryptedCredential = property0.Value.GetString(); continue; } + if (property0.NameEquals("serviceEndpoint"u8)) + { + if (property0.Value.ValueKind == JsonValueKind.Null) + { + continue; + } + serviceEndpoint = JsonSerializer.Deserialize>(property0.Value.GetRawText()); + continue; + } + if (property0.NameEquals("credential"u8)) + { + if (property0.Value.ValueKind == JsonValueKind.Null) + { + continue; + } + credential = DataFactoryCredentialReference.DeserializeDataFactoryCredentialReference(property0.Value, options); + continue; + } } continue; } @@ -270,7 +300,9 @@ internal static AzureTableStorageLinkedService DeserializeAzureTableStorageLinke accountKey, sasUri, sasToken, - encryptedCredential); + encryptedCredential, + serviceEndpoint, + credential); } BinaryData IPersistableModel.Write(ModelReaderWriterOptions options) diff --git a/sdk/datafactory/Azure.ResourceManager.DataFactory/src/Generated/Models/AzureTableStorageLinkedService.cs b/sdk/datafactory/Azure.ResourceManager.DataFactory/src/Generated/Models/AzureTableStorageLinkedService.cs index 5a40300d9c75..47cdea858e6b 100644 --- a/sdk/datafactory/Azure.ResourceManager.DataFactory/src/Generated/Models/AzureTableStorageLinkedService.cs +++ b/sdk/datafactory/Azure.ResourceManager.DataFactory/src/Generated/Models/AzureTableStorageLinkedService.cs @@ -32,13 +32,17 @@ public AzureTableStorageLinkedService() /// SAS URI of the Azure Storage resource. It is mutually exclusive with connectionString property. Type: string, SecureString or AzureKeyVaultSecretReference. /// The Azure key vault secret reference of sasToken in sas uri. /// The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string. - internal AzureTableStorageLinkedService(string linkedServiceType, IntegrationRuntimeReference connectVia, string description, IDictionary parameters, IList annotations, IDictionary additionalProperties, DataFactoryElement connectionString, DataFactoryKeyVaultSecret accountKey, DataFactoryElement sasUri, DataFactoryKeyVaultSecret sasToken, string encryptedCredential) : base(linkedServiceType, connectVia, description, parameters, annotations, additionalProperties) + /// Table service endpoint of the Azure Table Storage resource. It is mutually exclusive with connectionString, sasUri property. + /// The credential reference containing authentication information. + internal AzureTableStorageLinkedService(string linkedServiceType, IntegrationRuntimeReference connectVia, string description, IDictionary parameters, IList annotations, IDictionary additionalProperties, DataFactoryElement connectionString, DataFactoryKeyVaultSecret accountKey, DataFactoryElement sasUri, DataFactoryKeyVaultSecret sasToken, string encryptedCredential, DataFactoryElement serviceEndpoint, DataFactoryCredentialReference credential) : base(linkedServiceType, connectVia, description, parameters, annotations, additionalProperties) { ConnectionString = connectionString; AccountKey = accountKey; SasUri = sasUri; SasToken = sasToken; EncryptedCredential = encryptedCredential; + ServiceEndpoint = serviceEndpoint; + Credential = credential; LinkedServiceType = linkedServiceType ?? "AzureTableStorage"; } @@ -52,5 +56,9 @@ internal AzureTableStorageLinkedService(string linkedServiceType, IntegrationRun public DataFactoryKeyVaultSecret SasToken { get; set; } /// The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string. public string EncryptedCredential { get; set; } + /// Table service endpoint of the Azure Table Storage resource. It is mutually exclusive with connectionString, sasUri property. + public DataFactoryElement ServiceEndpoint { get; set; } + /// The credential reference containing authentication information. + public DataFactoryCredentialReference Credential { get; set; } } } diff --git a/sdk/datafactory/Azure.ResourceManager.DataFactory/src/Generated/Models/ContinuationSettingsReference.Serialization.cs b/sdk/datafactory/Azure.ResourceManager.DataFactory/src/Generated/Models/ContinuationSettingsReference.Serialization.cs new file mode 100644 index 000000000000..8247cf12ab89 --- /dev/null +++ b/sdk/datafactory/Azure.ResourceManager.DataFactory/src/Generated/Models/ContinuationSettingsReference.Serialization.cs @@ -0,0 +1,157 @@ +// Copyright (c) Microsoft Corporation. All rights reserved. +// Licensed under the MIT License. + +// + +#nullable disable + +using System; +using System.ClientModel.Primitives; +using System.Collections.Generic; +using System.Text.Json; +using Azure.Core; +using Azure.Core.Expressions.DataFactory; + +namespace Azure.ResourceManager.DataFactory.Models +{ + public partial class ContinuationSettingsReference : IUtf8JsonSerializable, IJsonModel + { + void IUtf8JsonSerializable.Write(Utf8JsonWriter writer) => ((IJsonModel)this).Write(writer, ModelSerializationExtensions.WireOptions); + + void IJsonModel.Write(Utf8JsonWriter writer, ModelReaderWriterOptions options) + { + var format = options.Format == "W" ? ((IPersistableModel)this).GetFormatFromOptions(options) : options.Format; + if (format != "J") + { + throw new FormatException($"The model {nameof(ContinuationSettingsReference)} does not support writing '{format}' format."); + } + + writer.WriteStartObject(); + if (Optional.IsDefined(ContinuationTtlInMinutes)) + { + writer.WritePropertyName("continuationTtlInMinutes"u8); + JsonSerializer.Serialize(writer, ContinuationTtlInMinutes); + } + if (Optional.IsDefined(IdleCondition)) + { + writer.WritePropertyName("idleCondition"u8); + JsonSerializer.Serialize(writer, IdleCondition); + } + if (Optional.IsDefined(CustomizedCheckpointKey)) + { + writer.WritePropertyName("customizedCheckpointKey"u8); + JsonSerializer.Serialize(writer, CustomizedCheckpointKey); + } + if (options.Format != "W" && _serializedAdditionalRawData != null) + { + foreach (var item in _serializedAdditionalRawData) + { + writer.WritePropertyName(item.Key); +#if NET6_0_OR_GREATER + writer.WriteRawValue(item.Value); +#else + using (JsonDocument document = JsonDocument.Parse(item.Value)) + { + JsonSerializer.Serialize(writer, document.RootElement); + } +#endif + } + } + writer.WriteEndObject(); + } + + ContinuationSettingsReference IJsonModel.Create(ref Utf8JsonReader reader, ModelReaderWriterOptions options) + { + var format = options.Format == "W" ? ((IPersistableModel)this).GetFormatFromOptions(options) : options.Format; + if (format != "J") + { + throw new FormatException($"The model {nameof(ContinuationSettingsReference)} does not support reading '{format}' format."); + } + + using JsonDocument document = JsonDocument.ParseValue(ref reader); + return DeserializeContinuationSettingsReference(document.RootElement, options); + } + + internal static ContinuationSettingsReference DeserializeContinuationSettingsReference(JsonElement element, ModelReaderWriterOptions options = null) + { + options ??= ModelSerializationExtensions.WireOptions; + + if (element.ValueKind == JsonValueKind.Null) + { + return null; + } + DataFactoryElement continuationTtlInMinutes = default; + DataFactoryElement idleCondition = default; + DataFactoryElement customizedCheckpointKey = default; + IDictionary serializedAdditionalRawData = default; + Dictionary rawDataDictionary = new Dictionary(); + foreach (var property in element.EnumerateObject()) + { + if (property.NameEquals("continuationTtlInMinutes"u8)) + { + if (property.Value.ValueKind == JsonValueKind.Null) + { + continue; + } + continuationTtlInMinutes = JsonSerializer.Deserialize>(property.Value.GetRawText()); + continue; + } + if (property.NameEquals("idleCondition"u8)) + { + if (property.Value.ValueKind == JsonValueKind.Null) + { + continue; + } + idleCondition = JsonSerializer.Deserialize>(property.Value.GetRawText()); + continue; + } + if (property.NameEquals("customizedCheckpointKey"u8)) + { + if (property.Value.ValueKind == JsonValueKind.Null) + { + continue; + } + customizedCheckpointKey = JsonSerializer.Deserialize>(property.Value.GetRawText()); + continue; + } + if (options.Format != "W") + { + rawDataDictionary.Add(property.Name, BinaryData.FromString(property.Value.GetRawText())); + } + } + serializedAdditionalRawData = rawDataDictionary; + return new ContinuationSettingsReference(continuationTtlInMinutes, idleCondition, customizedCheckpointKey, serializedAdditionalRawData); + } + + BinaryData IPersistableModel.Write(ModelReaderWriterOptions options) + { + var format = options.Format == "W" ? ((IPersistableModel)this).GetFormatFromOptions(options) : options.Format; + + switch (format) + { + case "J": + return ModelReaderWriter.Write(this, options); + default: + throw new FormatException($"The model {nameof(ContinuationSettingsReference)} does not support writing '{options.Format}' format."); + } + } + + ContinuationSettingsReference IPersistableModel.Create(BinaryData data, ModelReaderWriterOptions options) + { + var format = options.Format == "W" ? ((IPersistableModel)this).GetFormatFromOptions(options) : options.Format; + + switch (format) + { + case "J": + { + using JsonDocument document = JsonDocument.Parse(data); + return DeserializeContinuationSettingsReference(document.RootElement, options); + } + default: + throw new FormatException($"The model {nameof(ContinuationSettingsReference)} does not support reading '{options.Format}' format."); + } + } + + string IPersistableModel.GetFormatFromOptions(ModelReaderWriterOptions options) => "J"; + } +} diff --git a/sdk/datafactory/Azure.ResourceManager.DataFactory/src/Generated/Models/ContinuationSettingsReference.cs b/sdk/datafactory/Azure.ResourceManager.DataFactory/src/Generated/Models/ContinuationSettingsReference.cs new file mode 100644 index 000000000000..e37280749aff --- /dev/null +++ b/sdk/datafactory/Azure.ResourceManager.DataFactory/src/Generated/Models/ContinuationSettingsReference.cs @@ -0,0 +1,74 @@ +// Copyright (c) Microsoft Corporation. All rights reserved. +// Licensed under the MIT License. + +// + +#nullable disable + +using System; +using System.Collections.Generic; +using Azure.Core.Expressions.DataFactory; + +namespace Azure.ResourceManager.DataFactory.Models +{ + /// Continuation settings for execute data flow activity. + public partial class ContinuationSettingsReference + { + /// + /// Keeps track of any properties unknown to the library. + /// + /// To assign an object to the value of this property use . + /// + /// + /// To assign an already formatted json string to this property use . + /// + /// + /// Examples: + /// + /// + /// BinaryData.FromObjectAsJson("foo") + /// Creates a payload of "foo". + /// + /// + /// BinaryData.FromString("\"foo\"") + /// Creates a payload of "foo". + /// + /// + /// BinaryData.FromObjectAsJson(new { key = "value" }) + /// Creates a payload of { "key": "value" }. + /// + /// + /// BinaryData.FromString("{\"key\": \"value\"}") + /// Creates a payload of { "key": "value" }. + /// + /// + /// + /// + private IDictionary _serializedAdditionalRawData; + + /// Initializes a new instance of . + public ContinuationSettingsReference() + { + } + + /// Initializes a new instance of . + /// Continuation TTL in minutes. + /// Idle condition. + /// Customized checkpoint key. + /// Keeps track of any properties unknown to the library. + internal ContinuationSettingsReference(DataFactoryElement continuationTtlInMinutes, DataFactoryElement idleCondition, DataFactoryElement customizedCheckpointKey, IDictionary serializedAdditionalRawData) + { + ContinuationTtlInMinutes = continuationTtlInMinutes; + IdleCondition = idleCondition; + CustomizedCheckpointKey = customizedCheckpointKey; + _serializedAdditionalRawData = serializedAdditionalRawData; + } + + /// Continuation TTL in minutes. + public DataFactoryElement ContinuationTtlInMinutes { get; set; } + /// Idle condition. + public DataFactoryElement IdleCondition { get; set; } + /// Customized checkpoint key. + public DataFactoryElement CustomizedCheckpointKey { get; set; } + } +} diff --git a/sdk/datafactory/Azure.ResourceManager.DataFactory/src/Generated/Models/ExecuteDataFlowActivity.Serialization.cs b/sdk/datafactory/Azure.ResourceManager.DataFactory/src/Generated/Models/ExecuteDataFlowActivity.Serialization.cs index 1cec95e43b7d..d04114c71ec7 100644 --- a/sdk/datafactory/Azure.ResourceManager.DataFactory/src/Generated/Models/ExecuteDataFlowActivity.Serialization.cs +++ b/sdk/datafactory/Azure.ResourceManager.DataFactory/src/Generated/Models/ExecuteDataFlowActivity.Serialization.cs @@ -90,6 +90,11 @@ void IJsonModel.Write(Utf8JsonWriter writer, ModelReade writer.WritePropertyName("integrationRuntime"u8); writer.WriteObjectValue(IntegrationRuntime, options); } + if (Optional.IsDefined(ContinuationSettings)) + { + writer.WritePropertyName("continuationSettings"u8); + writer.WriteObjectValue(ContinuationSettings, options); + } if (Optional.IsDefined(Compute)) { writer.WritePropertyName("compute"u8); @@ -163,6 +168,7 @@ internal static ExecuteDataFlowActivity DeserializeExecuteDataFlowActivity(JsonE DataFlowReference dataFlow = default; DataFlowStagingInfo staging = default; IntegrationRuntimeReference integrationRuntime = default; + ContinuationSettingsReference continuationSettings = default; ExecuteDataFlowActivityComputeType compute = default; DataFactoryElement traceLevel = default; DataFactoryElement continueOnError = default; @@ -283,6 +289,15 @@ internal static ExecuteDataFlowActivity DeserializeExecuteDataFlowActivity(JsonE integrationRuntime = IntegrationRuntimeReference.DeserializeIntegrationRuntimeReference(property0.Value, options); continue; } + if (property0.NameEquals("continuationSettings"u8)) + { + if (property0.Value.ValueKind == JsonValueKind.Null) + { + continue; + } + continuationSettings = ContinuationSettingsReference.DeserializeContinuationSettingsReference(property0.Value, options); + continue; + } if (property0.NameEquals("compute"u8)) { if (property0.Value.ValueKind == JsonValueKind.Null) @@ -348,6 +363,7 @@ internal static ExecuteDataFlowActivity DeserializeExecuteDataFlowActivity(JsonE dataFlow, staging, integrationRuntime, + continuationSettings, compute, traceLevel, continueOnError, diff --git a/sdk/datafactory/Azure.ResourceManager.DataFactory/src/Generated/Models/ExecuteDataFlowActivity.cs b/sdk/datafactory/Azure.ResourceManager.DataFactory/src/Generated/Models/ExecuteDataFlowActivity.cs index 1930f6714e7a..6921300e69ab 100644 --- a/sdk/datafactory/Azure.ResourceManager.DataFactory/src/Generated/Models/ExecuteDataFlowActivity.cs +++ b/sdk/datafactory/Azure.ResourceManager.DataFactory/src/Generated/Models/ExecuteDataFlowActivity.cs @@ -41,16 +41,18 @@ public ExecuteDataFlowActivity(string name, DataFlowReference dataFlow) : base(n /// Data flow reference. /// Staging info for execute data flow activity. /// The integration runtime reference. + /// Continuation settings for execute data flow activity. /// Compute properties for data flow activity. /// Trace level setting used for data flow monitoring output. Supported values are: 'coarse', 'fine', and 'none'. Type: string (or Expression with resultType string). /// Continue on error setting used for data flow execution. Enables processing to continue if a sink fails. Type: boolean (or Expression with resultType boolean). /// Concurrent run setting used for data flow execution. Allows sinks with the same save order to be processed concurrently. Type: boolean (or Expression with resultType boolean). /// Specify number of parallel staging for sources applicable to the sink. Type: integer (or Expression with resultType integer). - internal ExecuteDataFlowActivity(string name, string activityType, string description, PipelineActivityState? state, ActivityOnInactiveMarkAs? onInactiveMarkAs, IList dependsOn, IList userProperties, IDictionary additionalProperties, DataFactoryLinkedServiceReference linkedServiceName, PipelineActivityPolicy policy, DataFlowReference dataFlow, DataFlowStagingInfo staging, IntegrationRuntimeReference integrationRuntime, ExecuteDataFlowActivityComputeType compute, DataFactoryElement traceLevel, DataFactoryElement continueOnError, DataFactoryElement runConcurrently, DataFactoryElement sourceStagingConcurrency) : base(name, activityType, description, state, onInactiveMarkAs, dependsOn, userProperties, additionalProperties, linkedServiceName, policy) + internal ExecuteDataFlowActivity(string name, string activityType, string description, PipelineActivityState? state, ActivityOnInactiveMarkAs? onInactiveMarkAs, IList dependsOn, IList userProperties, IDictionary additionalProperties, DataFactoryLinkedServiceReference linkedServiceName, PipelineActivityPolicy policy, DataFlowReference dataFlow, DataFlowStagingInfo staging, IntegrationRuntimeReference integrationRuntime, ContinuationSettingsReference continuationSettings, ExecuteDataFlowActivityComputeType compute, DataFactoryElement traceLevel, DataFactoryElement continueOnError, DataFactoryElement runConcurrently, DataFactoryElement sourceStagingConcurrency) : base(name, activityType, description, state, onInactiveMarkAs, dependsOn, userProperties, additionalProperties, linkedServiceName, policy) { DataFlow = dataFlow; Staging = staging; IntegrationRuntime = integrationRuntime; + ContinuationSettings = continuationSettings; Compute = compute; TraceLevel = traceLevel; ContinueOnError = continueOnError; @@ -70,6 +72,8 @@ internal ExecuteDataFlowActivity() public DataFlowStagingInfo Staging { get; set; } /// The integration runtime reference. public IntegrationRuntimeReference IntegrationRuntime { get; set; } + /// Continuation settings for execute data flow activity. + public ContinuationSettingsReference ContinuationSettings { get; set; } /// Compute properties for data flow activity. public ExecuteDataFlowActivityComputeType Compute { get; set; } /// Trace level setting used for data flow monitoring output. Supported values are: 'coarse', 'fine', and 'none'. Type: string (or Expression with resultType string). diff --git a/sdk/datafactory/Azure.ResourceManager.DataFactory/src/Generated/Models/ExecuteWranglingDataflowActivity.Serialization.cs b/sdk/datafactory/Azure.ResourceManager.DataFactory/src/Generated/Models/ExecuteWranglingDataflowActivity.Serialization.cs index 4f61081fb142..307f3b8426d5 100644 --- a/sdk/datafactory/Azure.ResourceManager.DataFactory/src/Generated/Models/ExecuteWranglingDataflowActivity.Serialization.cs +++ b/sdk/datafactory/Azure.ResourceManager.DataFactory/src/Generated/Models/ExecuteWranglingDataflowActivity.Serialization.cs @@ -85,6 +85,11 @@ void IJsonModel.Write(Utf8JsonWriter writer, M writer.WritePropertyName("integrationRuntime"u8); writer.WriteObjectValue(IntegrationRuntime, options); } + if (Optional.IsDefined(ContinuationSettings)) + { + writer.WritePropertyName("continuationSettings"u8); + writer.WriteObjectValue(ContinuationSettings, options); + } if (Optional.IsDefined(Compute)) { writer.WritePropertyName("compute"u8); @@ -178,6 +183,7 @@ internal static ExecuteWranglingDataflowActivity DeserializeExecuteWranglingData DataFlowReference dataFlow = default; DataFlowStagingInfo staging = default; IntegrationRuntimeReference integrationRuntime = default; + ContinuationSettingsReference continuationSettings = default; ExecuteDataFlowActivityComputeType compute = default; DataFactoryElement traceLevel = default; DataFactoryElement continueOnError = default; @@ -291,6 +297,15 @@ internal static ExecuteWranglingDataflowActivity DeserializeExecuteWranglingData integrationRuntime = IntegrationRuntimeReference.DeserializeIntegrationRuntimeReference(property0.Value, options); continue; } + if (property0.NameEquals("continuationSettings"u8)) + { + if (property0.Value.ValueKind == JsonValueKind.Null) + { + continue; + } + continuationSettings = ContinuationSettingsReference.DeserializeContinuationSettingsReference(property0.Value, options); + continue; + } if (property0.NameEquals("compute"u8)) { if (property0.Value.ValueKind == JsonValueKind.Null) @@ -383,6 +398,7 @@ internal static ExecuteWranglingDataflowActivity DeserializeExecuteWranglingData dataFlow, staging, integrationRuntime, + continuationSettings, compute, traceLevel, continueOnError, diff --git a/sdk/datafactory/Azure.ResourceManager.DataFactory/src/Generated/Models/ExecuteWranglingDataflowActivity.cs b/sdk/datafactory/Azure.ResourceManager.DataFactory/src/Generated/Models/ExecuteWranglingDataflowActivity.cs index 14b68091969c..5c31f5da9589 100644 --- a/sdk/datafactory/Azure.ResourceManager.DataFactory/src/Generated/Models/ExecuteWranglingDataflowActivity.cs +++ b/sdk/datafactory/Azure.ResourceManager.DataFactory/src/Generated/Models/ExecuteWranglingDataflowActivity.cs @@ -42,6 +42,7 @@ public ExecuteWranglingDataflowActivity(string name, DataFlowReference dataFlow) /// Data flow reference. /// Staging info for execute data flow activity. /// The integration runtime reference. + /// Continuation settings for execute data flow activity. /// Compute properties for data flow activity. /// Trace level setting used for data flow monitoring output. Supported values are: 'coarse', 'fine', and 'none'. Type: string (or Expression with resultType string). /// Continue on error setting used for data flow execution. Enables processing to continue if a sink fails. Type: boolean (or Expression with resultType boolean). @@ -49,12 +50,13 @@ public ExecuteWranglingDataflowActivity(string name, DataFlowReference dataFlow) /// Specify number of parallel staging for sources applicable to the sink. Type: integer (or Expression with resultType integer). /// (Deprecated. Please use Queries). List of Power Query activity sinks mapped to a queryName. /// List of mapping for Power Query mashup query to sink dataset(s). - internal ExecuteWranglingDataflowActivity(string name, string activityType, string description, PipelineActivityState? state, ActivityOnInactiveMarkAs? onInactiveMarkAs, IList dependsOn, IList userProperties, IDictionary additionalProperties, PipelineActivityPolicy policy, DataFlowReference dataFlow, DataFlowStagingInfo staging, IntegrationRuntimeReference integrationRuntime, ExecuteDataFlowActivityComputeType compute, DataFactoryElement traceLevel, DataFactoryElement continueOnError, DataFactoryElement runConcurrently, DataFactoryElement sourceStagingConcurrency, IDictionary sinks, IList queries) : base(name, activityType, description, state, onInactiveMarkAs, dependsOn, userProperties, additionalProperties) + internal ExecuteWranglingDataflowActivity(string name, string activityType, string description, PipelineActivityState? state, ActivityOnInactiveMarkAs? onInactiveMarkAs, IList dependsOn, IList userProperties, IDictionary additionalProperties, PipelineActivityPolicy policy, DataFlowReference dataFlow, DataFlowStagingInfo staging, IntegrationRuntimeReference integrationRuntime, ContinuationSettingsReference continuationSettings, ExecuteDataFlowActivityComputeType compute, DataFactoryElement traceLevel, DataFactoryElement continueOnError, DataFactoryElement runConcurrently, DataFactoryElement sourceStagingConcurrency, IDictionary sinks, IList queries) : base(name, activityType, description, state, onInactiveMarkAs, dependsOn, userProperties, additionalProperties) { Policy = policy; DataFlow = dataFlow; Staging = staging; IntegrationRuntime = integrationRuntime; + ContinuationSettings = continuationSettings; Compute = compute; TraceLevel = traceLevel; ContinueOnError = continueOnError; @@ -78,6 +80,8 @@ internal ExecuteWranglingDataflowActivity() public DataFlowStagingInfo Staging { get; set; } /// The integration runtime reference. public IntegrationRuntimeReference IntegrationRuntime { get; set; } + /// Continuation settings for execute data flow activity. + public ContinuationSettingsReference ContinuationSettings { get; set; } /// Compute properties for data flow activity. public ExecuteDataFlowActivityComputeType Compute { get; set; } /// Trace level setting used for data flow monitoring output. Supported values are: 'coarse', 'fine', and 'none'. Type: string (or Expression with resultType string). diff --git a/sdk/datafactory/Azure.ResourceManager.DataFactory/src/Generated/Models/SnowflakeExportCopyCommand.Serialization.cs b/sdk/datafactory/Azure.ResourceManager.DataFactory/src/Generated/Models/SnowflakeExportCopyCommand.Serialization.cs index c0a7e7a965be..028d9a2dbcc0 100644 --- a/sdk/datafactory/Azure.ResourceManager.DataFactory/src/Generated/Models/SnowflakeExportCopyCommand.Serialization.cs +++ b/sdk/datafactory/Azure.ResourceManager.DataFactory/src/Generated/Models/SnowflakeExportCopyCommand.Serialization.cs @@ -10,6 +10,7 @@ using System.Collections.Generic; using System.Text.Json; using Azure.Core; +using Azure.Core.Expressions.DataFactory; namespace Azure.ResourceManager.DataFactory.Models { @@ -72,6 +73,11 @@ void IJsonModel.Write(Utf8JsonWriter writer, ModelRe } writer.WriteEndObject(); } + if (Optional.IsDefined(StorageIntegration)) + { + writer.WritePropertyName("storageIntegration"u8); + JsonSerializer.Serialize(writer, StorageIntegration); + } writer.WritePropertyName("type"u8); writer.WriteStringValue(ExportSettingsType); foreach (var item in AdditionalProperties) @@ -111,6 +117,7 @@ internal static SnowflakeExportCopyCommand DeserializeSnowflakeExportCopyCommand } IDictionary additionalCopyOptions = default; IDictionary additionalFormatOptions = default; + DataFactoryElement storageIntegration = default; string type = default; IDictionary additionalProperties = default; Dictionary additionalPropertiesDictionary = new Dictionary(); @@ -158,6 +165,15 @@ internal static SnowflakeExportCopyCommand DeserializeSnowflakeExportCopyCommand additionalFormatOptions = dictionary; continue; } + if (property.NameEquals("storageIntegration"u8)) + { + if (property.Value.ValueKind == JsonValueKind.Null) + { + continue; + } + storageIntegration = JsonSerializer.Deserialize>(property.Value.GetRawText()); + continue; + } if (property.NameEquals("type"u8)) { type = property.Value.GetString(); @@ -166,7 +182,7 @@ internal static SnowflakeExportCopyCommand DeserializeSnowflakeExportCopyCommand additionalPropertiesDictionary.Add(property.Name, BinaryData.FromString(property.Value.GetRawText())); } additionalProperties = additionalPropertiesDictionary; - return new SnowflakeExportCopyCommand(type, additionalProperties, additionalCopyOptions ?? new ChangeTrackingDictionary(), additionalFormatOptions ?? new ChangeTrackingDictionary()); + return new SnowflakeExportCopyCommand(type, additionalProperties, additionalCopyOptions ?? new ChangeTrackingDictionary(), additionalFormatOptions ?? new ChangeTrackingDictionary(), storageIntegration); } BinaryData IPersistableModel.Write(ModelReaderWriterOptions options) diff --git a/sdk/datafactory/Azure.ResourceManager.DataFactory/src/Generated/Models/SnowflakeExportCopyCommand.cs b/sdk/datafactory/Azure.ResourceManager.DataFactory/src/Generated/Models/SnowflakeExportCopyCommand.cs index 60f5f2777df7..ebe34d9b96cf 100644 --- a/sdk/datafactory/Azure.ResourceManager.DataFactory/src/Generated/Models/SnowflakeExportCopyCommand.cs +++ b/sdk/datafactory/Azure.ResourceManager.DataFactory/src/Generated/Models/SnowflakeExportCopyCommand.cs @@ -7,6 +7,7 @@ using System; using System.Collections.Generic; +using Azure.Core.Expressions.DataFactory; namespace Azure.ResourceManager.DataFactory.Models { @@ -26,10 +27,12 @@ public SnowflakeExportCopyCommand() /// Additional Properties. /// Additional copy options directly passed to snowflake Copy Command. Type: key value pairs (value should be string type) (or Expression with resultType object). Example: "additionalCopyOptions": { "DATE_FORMAT": "MM/DD/YYYY", "TIME_FORMAT": "'HH24:MI:SS.FF'" }. /// Additional format options directly passed to snowflake Copy Command. Type: key value pairs (value should be string type) (or Expression with resultType object). Example: "additionalFormatOptions": { "OVERWRITE": "TRUE", "MAX_FILE_SIZE": "'FALSE'" }. - internal SnowflakeExportCopyCommand(string exportSettingsType, IDictionary additionalProperties, IDictionary additionalCopyOptions, IDictionary additionalFormatOptions) : base(exportSettingsType, additionalProperties) + /// The name of the snowflake storage integration to use for the copy operation. Type: string (or Expression with resultType string). + internal SnowflakeExportCopyCommand(string exportSettingsType, IDictionary additionalProperties, IDictionary additionalCopyOptions, IDictionary additionalFormatOptions, DataFactoryElement storageIntegration) : base(exportSettingsType, additionalProperties) { AdditionalCopyOptions = additionalCopyOptions; AdditionalFormatOptions = additionalFormatOptions; + StorageIntegration = storageIntegration; ExportSettingsType = exportSettingsType ?? "SnowflakeExportCopyCommand"; } @@ -95,5 +98,7 @@ internal SnowflakeExportCopyCommand(string exportSettingsType, IDictionary /// public IDictionary AdditionalFormatOptions { get; } + /// The name of the snowflake storage integration to use for the copy operation. Type: string (or Expression with resultType string). + public DataFactoryElement StorageIntegration { get; set; } } } diff --git a/sdk/datafactory/Azure.ResourceManager.DataFactory/src/Generated/Models/SnowflakeImportCopyCommand.Serialization.cs b/sdk/datafactory/Azure.ResourceManager.DataFactory/src/Generated/Models/SnowflakeImportCopyCommand.Serialization.cs index 4b30ab7665e8..5e2428adc172 100644 --- a/sdk/datafactory/Azure.ResourceManager.DataFactory/src/Generated/Models/SnowflakeImportCopyCommand.Serialization.cs +++ b/sdk/datafactory/Azure.ResourceManager.DataFactory/src/Generated/Models/SnowflakeImportCopyCommand.Serialization.cs @@ -10,6 +10,7 @@ using System.Collections.Generic; using System.Text.Json; using Azure.Core; +using Azure.Core.Expressions.DataFactory; namespace Azure.ResourceManager.DataFactory.Models { @@ -72,6 +73,11 @@ void IJsonModel.Write(Utf8JsonWriter writer, ModelRe } writer.WriteEndObject(); } + if (Optional.IsDefined(StorageIntegration)) + { + writer.WritePropertyName("storageIntegration"u8); + JsonSerializer.Serialize(writer, StorageIntegration); + } writer.WritePropertyName("type"u8); writer.WriteStringValue(ImportSettingsType); foreach (var item in AdditionalProperties) @@ -111,6 +117,7 @@ internal static SnowflakeImportCopyCommand DeserializeSnowflakeImportCopyCommand } IDictionary additionalCopyOptions = default; IDictionary additionalFormatOptions = default; + DataFactoryElement storageIntegration = default; string type = default; IDictionary additionalProperties = default; Dictionary additionalPropertiesDictionary = new Dictionary(); @@ -158,6 +165,15 @@ internal static SnowflakeImportCopyCommand DeserializeSnowflakeImportCopyCommand additionalFormatOptions = dictionary; continue; } + if (property.NameEquals("storageIntegration"u8)) + { + if (property.Value.ValueKind == JsonValueKind.Null) + { + continue; + } + storageIntegration = JsonSerializer.Deserialize>(property.Value.GetRawText()); + continue; + } if (property.NameEquals("type"u8)) { type = property.Value.GetString(); @@ -166,7 +182,7 @@ internal static SnowflakeImportCopyCommand DeserializeSnowflakeImportCopyCommand additionalPropertiesDictionary.Add(property.Name, BinaryData.FromString(property.Value.GetRawText())); } additionalProperties = additionalPropertiesDictionary; - return new SnowflakeImportCopyCommand(type, additionalProperties, additionalCopyOptions ?? new ChangeTrackingDictionary(), additionalFormatOptions ?? new ChangeTrackingDictionary()); + return new SnowflakeImportCopyCommand(type, additionalProperties, additionalCopyOptions ?? new ChangeTrackingDictionary(), additionalFormatOptions ?? new ChangeTrackingDictionary(), storageIntegration); } BinaryData IPersistableModel.Write(ModelReaderWriterOptions options) diff --git a/sdk/datafactory/Azure.ResourceManager.DataFactory/src/Generated/Models/SnowflakeImportCopyCommand.cs b/sdk/datafactory/Azure.ResourceManager.DataFactory/src/Generated/Models/SnowflakeImportCopyCommand.cs index fd1231f7d147..7fcad3180abf 100644 --- a/sdk/datafactory/Azure.ResourceManager.DataFactory/src/Generated/Models/SnowflakeImportCopyCommand.cs +++ b/sdk/datafactory/Azure.ResourceManager.DataFactory/src/Generated/Models/SnowflakeImportCopyCommand.cs @@ -7,6 +7,7 @@ using System; using System.Collections.Generic; +using Azure.Core.Expressions.DataFactory; namespace Azure.ResourceManager.DataFactory.Models { @@ -26,10 +27,12 @@ public SnowflakeImportCopyCommand() /// Additional Properties. /// Additional copy options directly passed to snowflake Copy Command. Type: key value pairs (value should be string type) (or Expression with resultType object). Example: "additionalCopyOptions": { "DATE_FORMAT": "MM/DD/YYYY", "TIME_FORMAT": "'HH24:MI:SS.FF'" }. /// Additional format options directly passed to snowflake Copy Command. Type: key value pairs (value should be string type) (or Expression with resultType object). Example: "additionalFormatOptions": { "FORCE": "TRUE", "LOAD_UNCERTAIN_FILES": "'FALSE'" }. - internal SnowflakeImportCopyCommand(string importSettingsType, IDictionary additionalProperties, IDictionary additionalCopyOptions, IDictionary additionalFormatOptions) : base(importSettingsType, additionalProperties) + /// The name of the snowflake storage integration to use for the copy operation. Type: string (or Expression with resultType string). + internal SnowflakeImportCopyCommand(string importSettingsType, IDictionary additionalProperties, IDictionary additionalCopyOptions, IDictionary additionalFormatOptions, DataFactoryElement storageIntegration) : base(importSettingsType, additionalProperties) { AdditionalCopyOptions = additionalCopyOptions; AdditionalFormatOptions = additionalFormatOptions; + StorageIntegration = storageIntegration; ImportSettingsType = importSettingsType ?? "SnowflakeImportCopyCommand"; } @@ -95,5 +98,7 @@ internal SnowflakeImportCopyCommand(string importSettingsType, IDictionary /// public IDictionary AdditionalFormatOptions { get; } + /// The name of the snowflake storage integration to use for the copy operation. Type: string (or Expression with resultType string). + public DataFactoryElement StorageIntegration { get; set; } } } diff --git a/sdk/datafactory/Azure.ResourceManager.DataFactory/src/Generated/Models/SqlServerAuthenticationType.cs b/sdk/datafactory/Azure.ResourceManager.DataFactory/src/Generated/Models/SqlServerAuthenticationType.cs index a457d9f325cb..42b538727c88 100644 --- a/sdk/datafactory/Azure.ResourceManager.DataFactory/src/Generated/Models/SqlServerAuthenticationType.cs +++ b/sdk/datafactory/Azure.ResourceManager.DataFactory/src/Generated/Models/SqlServerAuthenticationType.cs @@ -24,11 +24,14 @@ public SqlServerAuthenticationType(string value) private const string SqlValue = "SQL"; private const string WindowsValue = "Windows"; + private const string UserAssignedManagedIdentityValue = "UserAssignedManagedIdentity"; /// SQL. public static SqlServerAuthenticationType Sql { get; } = new SqlServerAuthenticationType(SqlValue); /// Windows. public static SqlServerAuthenticationType Windows { get; } = new SqlServerAuthenticationType(WindowsValue); + /// UserAssignedManagedIdentity. + public static SqlServerAuthenticationType UserAssignedManagedIdentity { get; } = new SqlServerAuthenticationType(UserAssignedManagedIdentityValue); /// Determines if two values are the same. public static bool operator ==(SqlServerAuthenticationType left, SqlServerAuthenticationType right) => left.Equals(right); /// Determines if two values are not the same. diff --git a/sdk/datafactory/Azure.ResourceManager.DataFactory/src/Generated/Models/SqlServerLinkedService.Serialization.cs b/sdk/datafactory/Azure.ResourceManager.DataFactory/src/Generated/Models/SqlServerLinkedService.Serialization.cs index 1b7ba06b241a..ff59b68707c4 100644 --- a/sdk/datafactory/Azure.ResourceManager.DataFactory/src/Generated/Models/SqlServerLinkedService.Serialization.cs +++ b/sdk/datafactory/Azure.ResourceManager.DataFactory/src/Generated/Models/SqlServerLinkedService.Serialization.cs @@ -199,6 +199,11 @@ void IJsonModel.Write(Utf8JsonWriter writer, ModelReader writer.WritePropertyName("alwaysEncryptedSettings"u8); writer.WriteObjectValue(AlwaysEncryptedSettings, options); } + if (Optional.IsDefined(Credential)) + { + writer.WritePropertyName("credential"u8); + writer.WriteObjectValue(Credential, options); + } writer.WriteEndObject(); foreach (var item in AdditionalProperties) { @@ -265,6 +270,7 @@ internal static SqlServerLinkedService DeserializeSqlServerLinkedService(JsonEle DataFactorySecret password = default; string encryptedCredential = default; SqlAlwaysEncryptedProperties alwaysEncryptedSettings = default; + DataFactoryCredentialReference credential = default; IDictionary additionalProperties = default; Dictionary additionalPropertiesDictionary = new Dictionary(); foreach (var property in element.EnumerateObject()) @@ -553,6 +559,15 @@ internal static SqlServerLinkedService DeserializeSqlServerLinkedService(JsonEle alwaysEncryptedSettings = SqlAlwaysEncryptedProperties.DeserializeSqlAlwaysEncryptedProperties(property0.Value, options); continue; } + if (property0.NameEquals("credential"u8)) + { + if (property0.Value.ValueKind == JsonValueKind.Null) + { + continue; + } + credential = DataFactoryCredentialReference.DeserializeDataFactoryCredentialReference(property0.Value, options); + continue; + } } continue; } @@ -590,7 +605,8 @@ internal static SqlServerLinkedService DeserializeSqlServerLinkedService(JsonEle userName, password, encryptedCredential, - alwaysEncryptedSettings); + alwaysEncryptedSettings, + credential); } BinaryData IPersistableModel.Write(ModelReaderWriterOptions options) diff --git a/sdk/datafactory/Azure.ResourceManager.DataFactory/src/Generated/Models/SqlServerLinkedService.cs b/sdk/datafactory/Azure.ResourceManager.DataFactory/src/Generated/Models/SqlServerLinkedService.cs index 4e19c85206e7..0b52eb5adfb4 100644 --- a/sdk/datafactory/Azure.ResourceManager.DataFactory/src/Generated/Models/SqlServerLinkedService.cs +++ b/sdk/datafactory/Azure.ResourceManager.DataFactory/src/Generated/Models/SqlServerLinkedService.cs @@ -52,7 +52,8 @@ public SqlServerLinkedService() /// The on-premises Windows authentication password. /// The encrypted credential used for authentication. Credentials are encrypted using the integration runtime credential manager. Type: string. /// Sql always encrypted properties. - internal SqlServerLinkedService(string linkedServiceType, IntegrationRuntimeReference connectVia, string description, IDictionary parameters, IList annotations, IDictionary additionalProperties, DataFactoryElement server, DataFactoryElement database, DataFactoryElement encrypt, DataFactoryElement trustServerCertificate, DataFactoryElement hostNameInCertificate, DataFactoryElement applicationIntent, DataFactoryElement connectTimeout, DataFactoryElement connectRetryCount, DataFactoryElement connectRetryInterval, DataFactoryElement loadBalanceTimeout, DataFactoryElement commandTimeout, DataFactoryElement integratedSecurity, DataFactoryElement failoverPartner, DataFactoryElement maxPoolSize, DataFactoryElement minPoolSize, DataFactoryElement multipleActiveResultSets, DataFactoryElement multiSubnetFailover, DataFactoryElement packetSize, DataFactoryElement pooling, DataFactoryElement connectionString, SqlServerAuthenticationType? authenticationType, DataFactoryElement userName, DataFactorySecret password, string encryptedCredential, SqlAlwaysEncryptedProperties alwaysEncryptedSettings) : base(linkedServiceType, connectVia, description, parameters, annotations, additionalProperties) + /// The credential reference containing authentication information. + internal SqlServerLinkedService(string linkedServiceType, IntegrationRuntimeReference connectVia, string description, IDictionary parameters, IList annotations, IDictionary additionalProperties, DataFactoryElement server, DataFactoryElement database, DataFactoryElement encrypt, DataFactoryElement trustServerCertificate, DataFactoryElement hostNameInCertificate, DataFactoryElement applicationIntent, DataFactoryElement connectTimeout, DataFactoryElement connectRetryCount, DataFactoryElement connectRetryInterval, DataFactoryElement loadBalanceTimeout, DataFactoryElement commandTimeout, DataFactoryElement integratedSecurity, DataFactoryElement failoverPartner, DataFactoryElement maxPoolSize, DataFactoryElement minPoolSize, DataFactoryElement multipleActiveResultSets, DataFactoryElement multiSubnetFailover, DataFactoryElement packetSize, DataFactoryElement pooling, DataFactoryElement connectionString, SqlServerAuthenticationType? authenticationType, DataFactoryElement userName, DataFactorySecret password, string encryptedCredential, SqlAlwaysEncryptedProperties alwaysEncryptedSettings, DataFactoryCredentialReference credential) : base(linkedServiceType, connectVia, description, parameters, annotations, additionalProperties) { Server = server; Database = database; @@ -79,6 +80,7 @@ internal SqlServerLinkedService(string linkedServiceType, IntegrationRuntimeRefe Password = password; EncryptedCredential = encryptedCredential; AlwaysEncryptedSettings = alwaysEncryptedSettings; + Credential = credential; LinkedServiceType = linkedServiceType ?? "SqlServer"; } @@ -132,5 +134,7 @@ internal SqlServerLinkedService(string linkedServiceType, IntegrationRuntimeRefe public string EncryptedCredential { get; set; } /// Sql always encrypted properties. public SqlAlwaysEncryptedProperties AlwaysEncryptedSettings { get; set; } + /// The credential reference containing authentication information. + public DataFactoryCredentialReference Credential { get; set; } } } diff --git a/sdk/datafactory/Azure.ResourceManager.DataFactory/src/autorest.md b/sdk/datafactory/Azure.ResourceManager.DataFactory/src/autorest.md index 566adc8411ee..33651fa3b756 100644 --- a/sdk/datafactory/Azure.ResourceManager.DataFactory/src/autorest.md +++ b/sdk/datafactory/Azure.ResourceManager.DataFactory/src/autorest.md @@ -8,7 +8,7 @@ azure-arm: true csharp: true library-name: DataFactory namespace: Azure.ResourceManager.DataFactory -require: https://github.com/Azure/azure-rest-api-specs/blob/bc06df2282fe4d29f42dce56a930fac14ea0f6c4/specification/datafactory/resource-manager/readme.md +require: /mnt/vss/_work/1/s/azure-rest-api-specs/specification/datafactory/resource-manager/readme.md output-folder: $(this-folder)/Generated clear-output-folder: true sample-gen: