diff --git a/sdk/datafactory/arm-datafactory/src/dataFactoryManagementClient.ts b/sdk/datafactory/arm-datafactory/src/dataFactoryManagementClient.ts index 21184128a935..be96a827a17d 100644 --- a/sdk/datafactory/arm-datafactory/src/dataFactoryManagementClient.ts +++ b/sdk/datafactory/arm-datafactory/src/dataFactoryManagementClient.ts @@ -30,7 +30,6 @@ class DataFactoryManagementClient extends DataFactoryManagementClientContext { activityRuns: operations.ActivityRuns; triggers: operations.Triggers; triggerRuns: operations.TriggerRuns; - rerunTriggers: operations.RerunTriggers; dataFlows: operations.DataFlows; dataFlowDebugSession: operations.DataFlowDebugSession; @@ -55,7 +54,6 @@ class DataFactoryManagementClient extends DataFactoryManagementClientContext { this.activityRuns = new operations.ActivityRuns(this); this.triggers = new operations.Triggers(this); this.triggerRuns = new operations.TriggerRuns(this); - this.rerunTriggers = new operations.RerunTriggers(this); this.dataFlows = new operations.DataFlows(this); this.dataFlowDebugSession = new operations.DataFlowDebugSession(this); } diff --git a/sdk/datafactory/arm-datafactory/src/models/dataFlowsMappers.ts b/sdk/datafactory/arm-datafactory/src/models/dataFlowsMappers.ts index f302ede884dc..9d3716d531f6 100644 --- a/sdk/datafactory/arm-datafactory/src/models/dataFlowsMappers.ts +++ b/sdk/datafactory/arm-datafactory/src/models/dataFlowsMappers.ts @@ -11,6 +11,7 @@ export { Activity, ActivityDependency, ActivityPolicy, + AdditionalColumns, AmazonMWSLinkedService, AmazonMWSObjectDataset, AmazonMWSSource, @@ -374,7 +375,6 @@ export { RedshiftUnloadSettings, RelationalSource, RelationalTableDataset, - RerunTriggerResource, RerunTumblingWindowTrigger, Resource, ResponsysLinkedService, @@ -434,6 +434,7 @@ export { ShopifyLinkedService, ShopifyObjectDataset, ShopifySource, + SkipErrorFile, SparkLinkedService, SparkObjectDataset, SparkSource, diff --git a/sdk/datafactory/arm-datafactory/src/models/datasetsMappers.ts b/sdk/datafactory/arm-datafactory/src/models/datasetsMappers.ts index bb0bb8162f24..4fd2bf0b1967 100644 --- a/sdk/datafactory/arm-datafactory/src/models/datasetsMappers.ts +++ b/sdk/datafactory/arm-datafactory/src/models/datasetsMappers.ts @@ -11,6 +11,7 @@ export { Activity, ActivityDependency, ActivityPolicy, + AdditionalColumns, AmazonMWSLinkedService, AmazonMWSObjectDataset, AmazonMWSSource, @@ -374,7 +375,6 @@ export { RedshiftUnloadSettings, RelationalSource, RelationalTableDataset, - RerunTriggerResource, RerunTumblingWindowTrigger, Resource, ResponsysLinkedService, @@ -434,6 +434,7 @@ export { ShopifyLinkedService, ShopifyObjectDataset, ShopifySource, + SkipErrorFile, SparkLinkedService, SparkObjectDataset, SparkSource, diff --git a/sdk/datafactory/arm-datafactory/src/models/factoriesMappers.ts b/sdk/datafactory/arm-datafactory/src/models/factoriesMappers.ts index a29de38c9fba..1e2234e64d0a 100644 --- a/sdk/datafactory/arm-datafactory/src/models/factoriesMappers.ts +++ b/sdk/datafactory/arm-datafactory/src/models/factoriesMappers.ts @@ -12,6 +12,7 @@ export { Activity, ActivityDependency, ActivityPolicy, + AdditionalColumns, AmazonMWSLinkedService, AmazonMWSObjectDataset, AmazonMWSSource, @@ -379,7 +380,6 @@ export { RedshiftUnloadSettings, RelationalSource, RelationalTableDataset, - RerunTriggerResource, RerunTumblingWindowTrigger, Resource, ResponsysLinkedService, @@ -439,6 +439,7 @@ export { ShopifyLinkedService, ShopifyObjectDataset, ShopifySource, + SkipErrorFile, SparkLinkedService, SparkObjectDataset, SparkSource, diff --git a/sdk/datafactory/arm-datafactory/src/models/index.ts b/sdk/datafactory/arm-datafactory/src/models/index.ts index f9c994ec0132..3cb4579316c5 100644 --- a/sdk/datafactory/arm-datafactory/src/models/index.ts +++ b/sdk/datafactory/arm-datafactory/src/models/index.ts @@ -685,7 +685,7 @@ export interface PipelineResource extends SubResource { /** * Contains the possible cases for Trigger. */ -export type TriggerUnion = Trigger | RerunTumblingWindowTrigger | ChainingTrigger | TumblingWindowTrigger | MultiplePipelineTriggerUnion; +export type TriggerUnion = Trigger | ChainingTrigger | RerunTumblingWindowTrigger | TumblingWindowTrigger | MultiplePipelineTriggerUnion; /** * Azure data factory nested object which contains information about creating pipeline run @@ -725,6 +725,21 @@ export interface TriggerResource extends SubResource { properties: TriggerUnion; } +/** + * A query of triggers. + */ +export interface TriggerQueryResponse { + /** + * List of triggers. + */ + value: TriggerResource[]; + /** + * The continuation token for getting the next page of results, if any remaining results exist, + * null otherwise. + */ + continuationToken?: string; +} + /** * Response body with a run identifier. */ @@ -990,6 +1005,20 @@ export interface LinkedServiceDebugResource extends SubResourceDebugResource { properties: LinkedServiceUnion; } +/** + * Query parameters for triggers. + */ +export interface TriggerFilterParameters { + /** + * The continuation token for getting the next page of results. Null for first page. + */ + continuationToken?: string; + /** + * The name of the parent TumblingWindowTrigger to get the child rerun triggers + */ + parentTriggerName?: string; +} + /** * Query filter option for listing runs. */ @@ -1335,79 +1364,6 @@ export interface TriggerRunsQueryResponse { continuationToken?: string; } -/** - * Rerun tumbling window trigger Parameters. - */ -export interface RerunTumblingWindowTriggerActionParameters { - /** - * The start time for the time period for which restatement is initiated. Only UTC time is - * currently supported. - */ - startTime: Date; - /** - * The end time for the time period for which restatement is initiated. Only UTC time is - * currently supported. - */ - endTime: Date; - /** - * The max number of parallel time windows (ready for execution) for which a rerun is triggered. - */ - maxConcurrency: number; -} - -/** - * Trigger that schedules pipeline reruns for all fixed time interval windows from a requested - * start time to requested end time. - */ -export interface RerunTumblingWindowTrigger { - /** - * Polymorphic Discriminator - */ - type: "RerunTumblingWindowTrigger"; - /** - * Trigger description. - */ - description?: string; - /** - * Indicates if trigger is running or not. Updated when Start/Stop APIs are called on the - * Trigger. Possible values include: 'Started', 'Stopped', 'Disabled' - * **NOTE: This property will not be serialized. It can only be populated by the server.** - */ - readonly runtimeState?: TriggerRuntimeState; - /** - * List of tags that can be used for describing the trigger. - */ - annotations?: any[]; - /** - * The parent trigger reference. - */ - parentTrigger?: any; - /** - * The start time for the time period for which restatement is initiated. Only UTC time is - * currently supported. - */ - requestedStartTime: Date; - /** - * The end time for the time period for which restatement is initiated. Only UTC time is - * currently supported. - */ - requestedEndTime: Date; - /** - * The max number of parallel time windows (ready for execution) for which a rerun is triggered. - */ - maxConcurrency: number; -} - -/** - * RerunTrigger resource type. - */ -export interface RerunTriggerResource extends SubResource { - /** - * Properties of the rerun trigger. - */ - properties: RerunTumblingWindowTrigger; -} - /** * Metadata associated with the operation. */ @@ -2031,461 +1987,565 @@ export interface MappingDataFlow { } /** - * Trigger that allows the referenced pipeline to depend on other pipeline runs based on - * runDimension Name/Value pairs. Upstream pipelines should declare the same runDimension Name and - * their runs should have the values for those runDimensions. The referenced pipeline run would be - * triggered if the values for the runDimension match for all upstream pipeline runs. + * Azure Function linked service. */ -export interface ChainingTrigger { +export interface AzureFunctionLinkedService { /** * Polymorphic Discriminator */ - type: "ChainingTrigger"; + type: "AzureFunction"; /** - * Trigger description. + * The integration runtime reference. */ - description?: string; + connectVia?: IntegrationRuntimeReference; /** - * Indicates if trigger is running or not. Updated when Start/Stop APIs are called on the - * Trigger. Possible values include: 'Started', 'Stopped', 'Disabled' - * **NOTE: This property will not be serialized. It can only be populated by the server.** + * Linked service description. */ - readonly runtimeState?: TriggerRuntimeState; + description?: string; /** - * List of tags that can be used for describing the trigger. + * Parameters for linked service. */ - annotations?: any[]; + parameters?: { [propertyName: string]: ParameterSpecification }; /** - * Pipeline for which runs are created when all upstream pipelines complete successfully. + * List of tags that can be used for describing the linked service. */ - pipelineProperty: TriggerPipelineReference; + annotations?: any[]; /** - * Upstream Pipelines. + * The endpoint of the Azure Function App. URL will be in the format + * https://.azurewebsites.net. */ - dependsOn: PipelineReference[]; + functionAppUrl: any; /** - * Run Dimension property that needs to be emitted by upstream pipelines. + * Function or Host key for Azure Function App. */ - runDimension: string; -} - -/** - * Contains the possible cases for DependencyReference. - */ -export type DependencyReferenceUnion = DependencyReference | SelfDependencyTumblingWindowTriggerReference | TriggerDependencyReferenceUnion; - -/** - * Referenced dependency. - */ -export interface DependencyReference { + functionKey?: SecretBaseUnion; /** - * Polymorphic Discriminator + * The encrypted credential used for authentication. Credentials are encrypted using the + * integration runtime credential manager. Type: string (or Expression with resultType string). */ - type: "DependencyReference"; + encryptedCredential?: any; } /** - * Self referenced tumbling window trigger dependency. + * Azure Data Explorer (Kusto) linked service. */ -export interface SelfDependencyTumblingWindowTriggerReference { +export interface AzureDataExplorerLinkedService { /** * Polymorphic Discriminator */ - type: "SelfDependencyTumblingWindowTriggerReference"; + type: "AzureDataExplorer"; /** - * Timespan applied to the start time of a tumbling window when evaluating dependency. + * The integration runtime reference. */ - offset: string; + connectVia?: IntegrationRuntimeReference; /** - * The size of the window when evaluating the dependency. If undefined the frequency of the - * tumbling window will be used. + * Linked service description. */ - size?: string; -} - -/** - * Trigger reference type. - */ -export interface TriggerReference { + description?: string; /** - * Reference trigger name. + * Parameters for linked service. */ - referenceName: string; -} - -/** - * Contains the possible cases for TriggerDependencyReference. - */ -export type TriggerDependencyReferenceUnion = TriggerDependencyReference | TumblingWindowTriggerDependencyReference; - -/** - * Trigger referenced dependency. - */ -export interface TriggerDependencyReference { + parameters?: { [propertyName: string]: ParameterSpecification }; /** - * Polymorphic Discriminator + * List of tags that can be used for describing the linked service. */ - type: "TriggerDependencyReference"; + annotations?: any[]; /** - * Referenced trigger. + * The endpoint of Azure Data Explorer (the engine's endpoint). URL will be in the format + * https://..kusto.windows.net. Type: string (or Expression with + * resultType string) */ - referenceTrigger: TriggerReference; -} - -/** - * Referenced tumbling window trigger dependency. - */ -export interface TumblingWindowTriggerDependencyReference { + endpoint: any; /** - * Polymorphic Discriminator + * The ID of the service principal used to authenticate against Azure Data Explorer. Type: string + * (or Expression with resultType string). */ - type: "TumblingWindowTriggerDependencyReference"; + servicePrincipalId: any; /** - * Referenced trigger. + * The key of the service principal used to authenticate against Kusto. */ - referenceTrigger: TriggerReference; + servicePrincipalKey: SecretBaseUnion; /** - * Timespan applied to the start time of a tumbling window when evaluating dependency. + * Database name for connection. Type: string (or Expression with resultType string). */ - offset?: string; + database: any; /** - * The size of the window when evaluating the dependency. If undefined the frequency of the - * tumbling window will be used. + * The name or ID of the tenant to which the service principal belongs. Type: string (or + * Expression with resultType string). */ - size?: string; + tenant: any; } /** - * Execution policy for an activity. + * SAP Table Linked Service. */ -export interface RetryPolicy { - /** - * Maximum ordinary retry attempts. Default is 0. Type: integer (or Expression with resultType - * integer), minimum: 0. - */ - count?: any; +export interface SapTableLinkedService { /** - * Interval between retries in seconds. Default is 30. + * Polymorphic Discriminator */ - intervalInSeconds?: number; -} - -/** - * Trigger that schedules pipeline runs for all fixed time interval windows from a start time - * without gaps and also supports backfill scenarios (when start time is in the past). - */ -export interface TumblingWindowTrigger { + type: "SapTable"; /** - * Polymorphic Discriminator + * The integration runtime reference. */ - type: "TumblingWindowTrigger"; + connectVia?: IntegrationRuntimeReference; /** - * Trigger description. + * Linked service description. */ description?: string; /** - * Indicates if trigger is running or not. Updated when Start/Stop APIs are called on the - * Trigger. Possible values include: 'Started', 'Stopped', 'Disabled' - * **NOTE: This property will not be serialized. It can only be populated by the server.** + * Parameters for linked service. */ - readonly runtimeState?: TriggerRuntimeState; + parameters?: { [propertyName: string]: ParameterSpecification }; /** - * List of tags that can be used for describing the trigger. + * List of tags that can be used for describing the linked service. */ annotations?: any[]; /** - * Pipeline for which runs are created when an event is fired for trigger window that is ready. + * Host name of the SAP instance where the table is located. Type: string (or Expression with + * resultType string). */ - pipelineProperty: TriggerPipelineReference; + server?: any; /** - * The frequency of the time windows. Possible values include: 'Minute', 'Hour' + * System number of the SAP system where the table is located. (Usually a two-digit decimal + * number represented as a string.) Type: string (or Expression with resultType string). */ - frequency: TumblingWindowFrequency; + systemNumber?: any; /** - * The interval of the time windows. The minimum interval allowed is 15 Minutes. + * Client ID of the client on the SAP system where the table is located. (Usually a three-digit + * decimal number represented as a string) Type: string (or Expression with resultType string). */ - interval: number; + clientId?: any; /** - * The start time for the time period for the trigger during which events are fired for windows - * that are ready. Only UTC time is currently supported. + * Language of the SAP system where the table is located. The default value is EN. Type: string + * (or Expression with resultType string). */ - startTime: Date; + language?: any; /** - * The end time for the time period for the trigger during which events are fired for windows - * that are ready. Only UTC time is currently supported. + * SystemID of the SAP system where the table is located. Type: string (or Expression with + * resultType string). */ - endTime?: Date; + systemId?: any; /** - * Specifies how long the trigger waits past due time before triggering new run. It doesn't alter - * window start and end time. The default is 0. Type: string (or Expression with resultType - * string), pattern: ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + * Username to access the SAP server where the table is located. Type: string (or Expression with + * resultType string). */ - delay?: any; + userName?: any; /** - * The max number of parallel time windows (ready for execution) for which a new run is - * triggered. + * Password to access the SAP server where the table is located. */ - maxConcurrency: number; + password?: SecretBaseUnion; /** - * Retry policy that will be applied for failed pipeline runs. + * The hostname of the SAP Message Server. Type: string (or Expression with resultType string). */ - retryPolicy?: RetryPolicy; + messageServer?: any; /** - * Triggers that this trigger depends on. Only tumbling window triggers are supported. + * The service name or port number of the Message Server. Type: string (or Expression with + * resultType string). */ - dependsOn?: DependencyReferenceUnion[]; -} - -/** - * Contains the possible cases for MultiplePipelineTrigger. - */ -export type MultiplePipelineTriggerUnion = MultiplePipelineTrigger | BlobEventsTrigger | BlobTrigger | ScheduleTrigger; - -/** - * Base class for all triggers that support one to many model for trigger to pipeline. - */ -export interface MultiplePipelineTrigger { + messageServerService?: any; /** - * Polymorphic Discriminator + * SNC activation indicator to access the SAP server where the table is located. Must be either 0 + * (off) or 1 (on). Type: string (or Expression with resultType string). */ - type: "MultiplePipelineTrigger"; + sncMode?: any; /** - * Trigger description. + * Initiator's SNC name to access the SAP server where the table is located. Type: string (or + * Expression with resultType string). */ - description?: string; + sncMyName?: any; /** - * Indicates if trigger is running or not. Updated when Start/Stop APIs are called on the - * Trigger. Possible values include: 'Started', 'Stopped', 'Disabled' - * **NOTE: This property will not be serialized. It can only be populated by the server.** + * Communication partner's SNC name to access the SAP server where the table is located. Type: + * string (or Expression with resultType string). */ - readonly runtimeState?: TriggerRuntimeState; + sncPartnerName?: any; /** - * List of tags that can be used for describing the trigger. + * External security product's library to access the SAP server where the table is located. Type: + * string (or Expression with resultType string). */ - annotations?: any[]; + sncLibraryPath?: any; /** - * Pipelines that need to be started. + * SNC Quality of Protection. Allowed value include: 1, 2, 3, 8, 9. Type: string (or Expression + * with resultType string). */ - pipelines?: TriggerPipelineReference[]; + sncQop?: any; + /** + * The Logon Group for the SAP System. Type: string (or Expression with resultType string). + */ + logonGroup?: any; + /** + * The encrypted credential used for authentication. Credentials are encrypted using the + * integration runtime credential manager. Type: string (or Expression with resultType string). + */ + encryptedCredential?: any; } /** - * Trigger that runs every time a Blob event occurs. + * Google AdWords service linked service. */ -export interface BlobEventsTrigger { +export interface GoogleAdWordsLinkedService { /** * Polymorphic Discriminator */ - type: "BlobEventsTrigger"; + type: "GoogleAdWords"; /** - * Trigger description. + * The integration runtime reference. + */ + connectVia?: IntegrationRuntimeReference; + /** + * Linked service description. */ description?: string; /** - * Indicates if trigger is running or not. Updated when Start/Stop APIs are called on the - * Trigger. Possible values include: 'Started', 'Stopped', 'Disabled' - * **NOTE: This property will not be serialized. It can only be populated by the server.** + * Parameters for linked service. */ - readonly runtimeState?: TriggerRuntimeState; + parameters?: { [propertyName: string]: ParameterSpecification }; /** - * List of tags that can be used for describing the trigger. + * List of tags that can be used for describing the linked service. */ annotations?: any[]; /** - * Pipelines that need to be started. + * The Client customer ID of the AdWords account that you want to fetch report data for. */ - pipelines?: TriggerPipelineReference[]; + clientCustomerID: any; /** - * The blob path must begin with the pattern provided for trigger to fire. For example, - * '/records/blobs/december/' will only fire the trigger for blobs in the december folder under - * the records container. At least one of these must be provided: blobPathBeginsWith, - * blobPathEndsWith. + * The developer token associated with the manager account that you use to grant access to the + * AdWords API. */ - blobPathBeginsWith?: string; + developerToken: SecretBaseUnion; /** - * The blob path must end with the pattern provided for trigger to fire. For example, - * 'december/boxes.csv' will only fire the trigger for blobs named boxes in a december folder. At - * least one of these must be provided: blobPathBeginsWith, blobPathEndsWith. + * The OAuth 2.0 authentication mechanism used for authentication. ServiceAuthentication can only + * be used on self-hosted IR. Possible values include: 'ServiceAuthentication', + * 'UserAuthentication' */ - blobPathEndsWith?: string; + authenticationType: GoogleAdWordsAuthenticationType; /** - * If set to true, blobs with zero bytes will be ignored. + * The refresh token obtained from Google for authorizing access to AdWords for + * UserAuthentication. */ - ignoreEmptyBlobs?: boolean; + refreshToken?: SecretBaseUnion; /** - * The type of events that cause this trigger to fire. + * The client id of the google application used to acquire the refresh token. Type: string (or + * Expression with resultType string). */ - events: BlobEventTypes[]; + clientId?: any; /** - * The ARM resource ID of the Storage Account. + * The client secret of the google application used to acquire the refresh token. */ - scope: string; + clientSecret?: SecretBaseUnion; + /** + * The service account email ID that is used for ServiceAuthentication and can only be used on + * self-hosted IR. + */ + email?: any; + /** + * The full path to the .p12 key file that is used to authenticate the service account email + * address and can only be used on self-hosted IR. + */ + keyFilePath?: any; + /** + * The full path of the .pem file containing trusted CA certificates for verifying the server + * when connecting over SSL. This property can only be set when using SSL on self-hosted IR. The + * default value is the cacerts.pem file installed with the IR. + */ + trustedCertPath?: any; + /** + * Specifies whether to use a CA certificate from the system trust store or from a specified PEM + * file. The default value is false. + */ + useSystemTrustStore?: any; + /** + * The encrypted credential used for authentication. Credentials are encrypted using the + * integration runtime credential manager. Type: string (or Expression with resultType string). + */ + encryptedCredential?: any; } /** - * Trigger that runs every time the selected Blob container changes. + * Oracle Service Cloud linked service. */ -export interface BlobTrigger { +export interface OracleServiceCloudLinkedService { /** * Polymorphic Discriminator */ - type: "BlobTrigger"; + type: "OracleServiceCloud"; /** - * Trigger description. + * The integration runtime reference. + */ + connectVia?: IntegrationRuntimeReference; + /** + * Linked service description. */ description?: string; /** - * Indicates if trigger is running or not. Updated when Start/Stop APIs are called on the - * Trigger. Possible values include: 'Started', 'Stopped', 'Disabled' - * **NOTE: This property will not be serialized. It can only be populated by the server.** + * Parameters for linked service. */ - readonly runtimeState?: TriggerRuntimeState; + parameters?: { [propertyName: string]: ParameterSpecification }; /** - * List of tags that can be used for describing the trigger. + * List of tags that can be used for describing the linked service. */ annotations?: any[]; /** - * Pipelines that need to be started. + * The URL of the Oracle Service Cloud instance. */ - pipelines?: TriggerPipelineReference[]; + host: any; /** - * The path of the container/folder that will trigger the pipeline. + * The user name that you use to access Oracle Service Cloud server. */ - folderPath: string; + username: any; /** - * The max number of parallel files to handle when it is triggered. + * The password corresponding to the user name that you provided in the username key. */ - maxConcurrency: number; + password: SecretBaseUnion; /** - * The Azure Storage linked service reference. + * Specifies whether the data source endpoints are encrypted using HTTPS. The default value is + * true. Type: boolean (or Expression with resultType boolean). */ - linkedService: LinkedServiceReference; -} - -/** - * The recurrence schedule occurrence. - */ -export interface RecurrenceScheduleOccurrence { + useEncryptedEndpoints?: any; /** - * The day of the week. Possible values include: 'Sunday', 'Monday', 'Tuesday', 'Wednesday', - * 'Thursday', 'Friday', 'Saturday' + * Specifies whether to require the host name in the server's certificate to match the host name + * of the server when connecting over SSL. The default value is true. Type: boolean (or + * Expression with resultType boolean). */ - day?: DayOfWeek; + useHostVerification?: any; /** - * The occurrence. + * Specifies whether to verify the identity of the server when connecting over SSL. The default + * value is true. Type: boolean (or Expression with resultType boolean). */ - occurrence?: number; + usePeerVerification?: any; /** - * Describes unknown properties. The value of an unknown property can be of "any" type. + * The encrypted credential used for authentication. Credentials are encrypted using the + * integration runtime credential manager. Type: string (or Expression with resultType string). */ - [property: string]: any; + encryptedCredential?: any; } /** - * The recurrence schedule. + * Dynamics AX linked service. */ -export interface RecurrenceSchedule { +export interface DynamicsAXLinkedService { /** - * The minutes. + * Polymorphic Discriminator */ - minutes?: number[]; + type: "DynamicsAX"; /** - * The hours. + * The integration runtime reference. */ - hours?: number[]; + connectVia?: IntegrationRuntimeReference; /** - * The days of the week. + * Linked service description. */ - weekDays?: DaysOfWeek[]; + description?: string; /** - * The month days. + * Parameters for linked service. */ - monthDays?: number[]; + parameters?: { [propertyName: string]: ParameterSpecification }; /** - * The monthly occurrences. + * List of tags that can be used for describing the linked service. */ - monthlyOccurrences?: RecurrenceScheduleOccurrence[]; + annotations?: any[]; /** - * Describes unknown properties. The value of an unknown property can be of "any" type. + * The Dynamics AX (or Dynamics 365 Finance and Operations) instance OData endpoint. */ - [property: string]: any; + url: any; + /** + * Specify the application's client ID. Type: string (or Expression with resultType string). + */ + servicePrincipalId: any; + /** + * Specify the application's key. Mark this field as a SecureString to store it securely in Data + * Factory, or reference a secret stored in Azure Key Vault. Type: string (or Expression with + * resultType string). + */ + servicePrincipalKey: SecretBaseUnion; + /** + * Specify the tenant information (domain name or tenant ID) under which your application + * resides. Retrieve it by hovering the mouse in the top-right corner of the Azure portal. Type: + * string (or Expression with resultType string). + */ + tenant: any; + /** + * Specify the resource you are requesting authorization. Type: string (or Expression with + * resultType string). + */ + aadResourceId: any; + /** + * The encrypted credential used for authentication. Credentials are encrypted using the + * integration runtime credential manager. Type: string (or Expression with resultType string). + */ + encryptedCredential?: any; } /** - * The workflow trigger recurrence. + * Responsys linked service. */ -export interface ScheduleTriggerRecurrence { +export interface ResponsysLinkedService { /** - * The frequency. Possible values include: 'NotSpecified', 'Minute', 'Hour', 'Day', 'Week', - * 'Month', 'Year' + * Polymorphic Discriminator */ - frequency?: RecurrenceFrequency; + type: "Responsys"; /** - * The interval. + * The integration runtime reference. */ - interval?: number; + connectVia?: IntegrationRuntimeReference; /** - * The start time. + * Linked service description. */ - startTime?: Date; + description?: string; /** - * The end time. + * Parameters for linked service. */ - endTime?: Date; + parameters?: { [propertyName: string]: ParameterSpecification }; /** - * The time zone. + * List of tags that can be used for describing the linked service. */ - timeZone?: string; + annotations?: any[]; /** - * The recurrence schedule. + * The endpoint of the Responsys server. */ - schedule?: RecurrenceSchedule; + endpoint: any; /** - * Describes unknown properties. The value of an unknown property can be of "any" type. + * The client ID associated with the Responsys application. Type: string (or Expression with + * resultType string). */ - [property: string]: any; + clientId: any; + /** + * The client secret associated with the Responsys application. Type: string (or Expression with + * resultType string). + */ + clientSecret?: SecretBaseUnion; + /** + * Specifies whether the data source endpoints are encrypted using HTTPS. The default value is + * true. Type: boolean (or Expression with resultType boolean). + */ + useEncryptedEndpoints?: any; + /** + * Specifies whether to require the host name in the server's certificate to match the host name + * of the server when connecting over SSL. The default value is true. Type: boolean (or + * Expression with resultType boolean). + */ + useHostVerification?: any; + /** + * Specifies whether to verify the identity of the server when connecting over SSL. The default + * value is true. Type: boolean (or Expression with resultType boolean). + */ + usePeerVerification?: any; + /** + * The encrypted credential used for authentication. Credentials are encrypted using the + * integration runtime credential manager. Type: string (or Expression with resultType string). + */ + encryptedCredential?: any; } /** - * Trigger that creates pipeline runs periodically, on schedule. + * Azure Databricks linked service. */ -export interface ScheduleTrigger { +export interface AzureDatabricksLinkedService { /** * Polymorphic Discriminator */ - type: "ScheduleTrigger"; + type: "AzureDatabricks"; /** - * Trigger description. + * The integration runtime reference. + */ + connectVia?: IntegrationRuntimeReference; + /** + * Linked service description. */ description?: string; /** - * Indicates if trigger is running or not. Updated when Start/Stop APIs are called on the - * Trigger. Possible values include: 'Started', 'Stopped', 'Disabled' - * **NOTE: This property will not be serialized. It can only be populated by the server.** + * Parameters for linked service. */ - readonly runtimeState?: TriggerRuntimeState; + parameters?: { [propertyName: string]: ParameterSpecification }; /** - * List of tags that can be used for describing the trigger. + * List of tags that can be used for describing the linked service. */ annotations?: any[]; /** - * Pipelines that need to be started. + * .azuredatabricks.net, domain name of your Databricks deployment. Type: string (or + * Expression with resultType string). */ - pipelines?: TriggerPipelineReference[]; + domain: any; /** - * Recurrence schedule configuration. + * Access token for databricks REST API. Refer to + * https://docs.azuredatabricks.net/api/latest/authentication.html. Type: string (or Expression + * with resultType string). */ - recurrence: ScheduleTriggerRecurrence; + accessToken: SecretBaseUnion; + /** + * The id of an existing interactive cluster that will be used for all runs of this activity. + * Type: string (or Expression with resultType string). + */ + existingClusterId?: any; + /** + * The id of an existing instance pool that will be used for all runs of this activity. Type: + * string (or Expression with resultType string). + */ + instancePoolId?: any; + /** + * If not using an existing interactive cluster, this specifies the Spark version of a new job + * cluster or instance pool nodes created for each run of this activity. Required if + * instancePoolId is specified. Type: string (or Expression with resultType string). + */ + newClusterVersion?: any; + /** + * If not using an existing interactive cluster, this specifies the number of worker nodes to use + * for the new job cluster or instance pool. For new job clusters, this a string-formatted Int32, + * like '1' means numOfWorker is 1 or '1:10' means auto-scale from 1 (min) to 10 (max). For + * instance pools, this is a string-formatted Int32, and can only specify a fixed number of + * worker nodes, such as '2'. Required if newClusterVersion is specified. Type: string (or + * Expression with resultType string). + */ + newClusterNumOfWorker?: any; + /** + * The node type of the new job cluster. This property is required if newClusterVersion is + * specified and instancePoolId is not specified. If instancePoolId is specified, this property + * is ignored. Type: string (or Expression with resultType string). + */ + newClusterNodeType?: any; + /** + * A set of optional, user-specified Spark configuration key-value pairs. + */ + newClusterSparkConf?: { [propertyName: string]: any }; + /** + * A set of optional, user-specified Spark environment variables key-value pairs. + */ + newClusterSparkEnvVars?: { [propertyName: string]: any }; + /** + * Additional tags for cluster resources. This property is ignored in instance pool + * configurations. + */ + newClusterCustomTags?: { [propertyName: string]: any }; + /** + * The driver node type for the new job cluster. This property is ignored in instance pool + * configurations. Type: string (or Expression with resultType string). + */ + newClusterDriverNodeType?: any; + /** + * User-defined initialization scripts for the new cluster. Type: array of strings (or Expression + * with resultType array of strings). + */ + newClusterInitScripts?: any; + /** + * Enable the elastic disk on the new cluster. This property is now ignored, and takes the + * default elastic disk behavior in Databricks (elastic disks are always enabled). Type: boolean + * (or Expression with resultType boolean). + */ + newClusterEnableElasticDisk?: any; + /** + * The encrypted credential used for authentication. Credentials are encrypted using the + * integration runtime credential manager. Type: string (or Expression with resultType string). + */ + encryptedCredential?: any; } /** - * Azure Function linked service. + * Azure Data Lake Analytics linked service. */ -export interface AzureFunctionLinkedService { +export interface AzureDataLakeAnalyticsLinkedService { /** * Polymorphic Discriminator */ - type: "AzureFunction"; + type: "AzureDataLakeAnalytics"; /** * The integration runtime reference. */ @@ -2503,29 +2563,75 @@ export interface AzureFunctionLinkedService { */ annotations?: any[]; /** - * The endpoint of the Azure Function App. URL will be in the format - * https://.azurewebsites.net. + * The Azure Data Lake Analytics account name. Type: string (or Expression with resultType + * string). */ - functionAppUrl: any; + accountName: any; /** - * Function or Host key for Azure Function App. + * The ID of the application used to authenticate against the Azure Data Lake Analytics account. + * Type: string (or Expression with resultType string). + */ + servicePrincipalId?: any; + /** + * The Key of the application used to authenticate against the Azure Data Lake Analytics account. + */ + servicePrincipalKey?: SecretBaseUnion; + /** + * The name or ID of the tenant to which the service principal belongs. Type: string (or + * Expression with resultType string). + */ + tenant: any; + /** + * Data Lake Analytics account subscription ID (if different from Data Factory account). Type: + * string (or Expression with resultType string). + */ + subscriptionId?: any; + /** + * Data Lake Analytics account resource group name (if different from Data Factory account). + * Type: string (or Expression with resultType string). + */ + resourceGroupName?: any; + /** + * Azure Data Lake Analytics URI Type: string (or Expression with resultType string). + */ + dataLakeAnalyticsUri?: any; + /** + * The encrypted credential used for authentication. Credentials are encrypted using the + * integration runtime credential manager. Type: string (or Expression with resultType string). + */ + encryptedCredential?: any; +} + +/** + * Custom script action to run on HDI ondemand cluster once it's up. + */ +export interface ScriptAction { + /** + * The user provided name of the script action. + */ + name: string; + /** + * The URI for the script action. + */ + uri: string; + /** + * The node types on which the script action should be executed. */ - functionKey?: SecretBaseUnion; + roles: any; /** - * The encrypted credential used for authentication. Credentials are encrypted using the - * integration runtime credential manager. Type: string (or Expression with resultType string). + * The parameters for the script action. */ - encryptedCredential?: any; + parameters?: string; } /** - * Azure Data Explorer (Kusto) linked service. + * HDInsight ondemand linked service. */ -export interface AzureDataExplorerLinkedService { +export interface HDInsightOnDemandLinkedService { /** * Polymorphic Discriminator */ - type: "AzureDataExplorer"; + type: "HDInsightOnDemand"; /** * The integration runtime reference. */ @@ -2543,142 +2649,166 @@ export interface AzureDataExplorerLinkedService { */ annotations?: any[]; /** - * The endpoint of Azure Data Explorer (the engine's endpoint). URL will be in the format - * https://..kusto.windows.net. Type: string (or Expression with - * resultType string) + * Number of worker/data nodes in the cluster. Suggestion value: 4. Type: string (or Expression + * with resultType string). */ - endpoint: any; + clusterSize: any; /** - * The ID of the service principal used to authenticate against Azure Data Explorer. Type: string - * (or Expression with resultType string). + * The allowed idle time for the on-demand HDInsight cluster. Specifies how long the on-demand + * HDInsight cluster stays alive after completion of an activity run if there are no other active + * jobs in the cluster. The minimum value is 5 mins. Type: string (or Expression with resultType + * string). */ - servicePrincipalId: any; + timeToLive: any; /** - * The key of the service principal used to authenticate against Kusto. + * Version of the HDInsight cluster.  Type: string (or Expression with resultType string). */ - servicePrincipalKey: SecretBaseUnion; + version: any; /** - * Database name for connection. Type: string (or Expression with resultType string). + * Azure Storage linked service to be used by the on-demand cluster for storing and processing + * data. */ - database: any; + linkedServiceName: LinkedServiceReference; /** - * The name or ID of the tenant to which the service principal belongs. Type: string (or - * Expression with resultType string). + * The customer’s subscription to host the cluster. Type: string (or Expression with resultType + * string). */ - tenant: any; -} - -/** - * SAP Table Linked Service. - */ -export interface SapTableLinkedService { + hostSubscriptionId: any; /** - * Polymorphic Discriminator + * The service principal id for the hostSubscriptionId. Type: string (or Expression with + * resultType string). */ - type: "SapTable"; + servicePrincipalId?: any; /** - * The integration runtime reference. + * The key for the service principal id. */ - connectVia?: IntegrationRuntimeReference; + servicePrincipalKey?: SecretBaseUnion; /** - * Linked service description. + * The Tenant id/name to which the service principal belongs. Type: string (or Expression with + * resultType string). */ - description?: string; + tenant: any; /** - * Parameters for linked service. + * The resource group where the cluster belongs. Type: string (or Expression with resultType + * string). */ - parameters?: { [propertyName: string]: ParameterSpecification }; + clusterResourceGroup: any; /** - * List of tags that can be used for describing the linked service. + * The prefix of cluster name, postfix will be distinct with timestamp. Type: string (or + * Expression with resultType string). */ - annotations?: any[]; + clusterNamePrefix?: any; /** - * Host name of the SAP instance where the table is located. Type: string (or Expression with - * resultType string). + * The username to access the cluster. Type: string (or Expression with resultType string). */ - server?: any; + clusterUserName?: any; /** - * System number of the SAP system where the table is located. (Usually a two-digit decimal - * number represented as a string.) Type: string (or Expression with resultType string). + * The password to access the cluster. */ - systemNumber?: any; + clusterPassword?: SecretBaseUnion; /** - * Client ID of the client on the SAP system where the table is located. (Usually a three-digit - * decimal number represented as a string) Type: string (or Expression with resultType string). + * The username to SSH remotely connect to cluster’s node (for Linux). Type: string (or + * Expression with resultType string). */ - clientId?: any; + clusterSshUserName?: any; /** - * Language of the SAP system where the table is located. The default value is EN. Type: string - * (or Expression with resultType string). + * The password to SSH remotely connect cluster’s node (for Linux). */ - language?: any; + clusterSshPassword?: SecretBaseUnion; /** - * SystemID of the SAP system where the table is located. Type: string (or Expression with - * resultType string). + * Specifies additional storage accounts for the HDInsight linked service so that the Data + * Factory service can register them on your behalf. */ - systemId?: any; + additionalLinkedServiceNames?: LinkedServiceReference[]; /** - * Username to access the SAP server where the table is located. Type: string (or Expression with - * resultType string). + * The name of Azure SQL linked service that point to the HCatalog database. The on-demand + * HDInsight cluster is created by using the Azure SQL database as the metastore. */ - userName?: any; + hcatalogLinkedServiceName?: LinkedServiceReference; /** - * Password to access the SAP server where the table is located. + * The cluster type. Type: string (or Expression with resultType string). */ - password?: SecretBaseUnion; + clusterType?: any; /** - * The hostname of the SAP Message Server. Type: string (or Expression with resultType string). + * The version of spark if the cluster type is 'spark'. Type: string (or Expression with + * resultType string). */ - messageServer?: any; + sparkVersion?: any; /** - * The service name or port number of the Message Server. Type: string (or Expression with - * resultType string). + * Specifies the core configuration parameters (as in core-site.xml) for the HDInsight cluster to + * be created. */ - messageServerService?: any; + coreConfiguration?: any; /** - * SNC activation indicator to access the SAP server where the table is located. Must be either 0 - * (off) or 1 (on). Type: string (or Expression with resultType string). + * Specifies the HBase configuration parameters (hbase-site.xml) for the HDInsight cluster. */ - sncMode?: any; + hBaseConfiguration?: any; /** - * Initiator's SNC name to access the SAP server where the table is located. Type: string (or - * Expression with resultType string). + * Specifies the HDFS configuration parameters (hdfs-site.xml) for the HDInsight cluster. */ - sncMyName?: any; + hdfsConfiguration?: any; /** - * Communication partner's SNC name to access the SAP server where the table is located. Type: - * string (or Expression with resultType string). + * Specifies the hive configuration parameters (hive-site.xml) for the HDInsight cluster. */ - sncPartnerName?: any; + hiveConfiguration?: any; /** - * External security product's library to access the SAP server where the table is located. Type: - * string (or Expression with resultType string). + * Specifies the MapReduce configuration parameters (mapred-site.xml) for the HDInsight cluster. */ - sncLibraryPath?: any; + mapReduceConfiguration?: any; /** - * SNC Quality of Protection. Allowed value include: 1, 2, 3, 8, 9. Type: string (or Expression - * with resultType string). + * Specifies the Oozie configuration parameters (oozie-site.xml) for the HDInsight cluster. */ - sncQop?: any; + oozieConfiguration?: any; /** - * The Logon Group for the SAP System. Type: string (or Expression with resultType string). + * Specifies the Storm configuration parameters (storm-site.xml) for the HDInsight cluster. */ - logonGroup?: any; + stormConfiguration?: any; + /** + * Specifies the Yarn configuration parameters (yarn-site.xml) for the HDInsight cluster. + */ + yarnConfiguration?: any; /** * The encrypted credential used for authentication. Credentials are encrypted using the * integration runtime credential manager. Type: string (or Expression with resultType string). */ encryptedCredential?: any; + /** + * Specifies the size of the head node for the HDInsight cluster. + */ + headNodeSize?: any; + /** + * Specifies the size of the data node for the HDInsight cluster. + */ + dataNodeSize?: any; + /** + * Specifies the size of the Zoo Keeper node for the HDInsight cluster. + */ + zookeeperNodeSize?: any; + /** + * Custom script actions to run on HDI ondemand cluster once it's up. Please refer to + * https://docs.microsoft.com/en-us/azure/hdinsight/hdinsight-hadoop-customize-cluster-linux?toc=%2Fen-us%2Fazure%2Fhdinsight%2Fr-server%2FTOC.json&bc=%2Fen-us%2Fazure%2Fbread%2Ftoc.json#understanding-script-actions. + */ + scriptActions?: ScriptAction[]; + /** + * The ARM resource ID for the vNet to which the cluster should be joined after creation. Type: + * string (or Expression with resultType string). + */ + virtualNetworkId?: any; + /** + * The ARM resource ID for the subnet in the vNet. If virtualNetworkId was specified, then this + * property is required. Type: string (or Expression with resultType string). + */ + subnetName?: any; } /** - * Google AdWords service linked service. + * Salesforce Marketing Cloud linked service. */ -export interface GoogleAdWordsLinkedService { +export interface SalesforceMarketingCloudLinkedService { /** * Polymorphic Discriminator */ - type: "GoogleAdWords"; + type: "SalesforceMarketingCloud"; /** * The integration runtime reference. */ @@ -2696,55 +2826,70 @@ export interface GoogleAdWordsLinkedService { */ annotations?: any[]; /** - * The Client customer ID of the AdWords account that you want to fetch report data for. + * The client ID associated with the Salesforce Marketing Cloud application. Type: string (or + * Expression with resultType string). */ - clientCustomerID: any; + clientId: any; /** - * The developer token associated with the manager account that you use to grant access to the - * AdWords API. + * The client secret associated with the Salesforce Marketing Cloud application. Type: string (or + * Expression with resultType string). */ - developerToken: SecretBaseUnion; + clientSecret?: SecretBaseUnion; /** - * The OAuth 2.0 authentication mechanism used for authentication. ServiceAuthentication can only - * be used on self-hosted IR. Possible values include: 'ServiceAuthentication', - * 'UserAuthentication' + * Specifies whether the data source endpoints are encrypted using HTTPS. The default value is + * true. Type: boolean (or Expression with resultType boolean). */ - authenticationType: GoogleAdWordsAuthenticationType; + useEncryptedEndpoints?: any; /** - * The refresh token obtained from Google for authorizing access to AdWords for - * UserAuthentication. + * Specifies whether to require the host name in the server's certificate to match the host name + * of the server when connecting over SSL. The default value is true. Type: boolean (or + * Expression with resultType boolean). */ - refreshToken?: SecretBaseUnion; + useHostVerification?: any; /** - * The client id of the google application used to acquire the refresh token. Type: string (or - * Expression with resultType string). + * Specifies whether to verify the identity of the server when connecting over SSL. The default + * value is true. Type: boolean (or Expression with resultType boolean). */ - clientId?: any; + usePeerVerification?: any; /** - * The client secret of the google application used to acquire the refresh token. + * The encrypted credential used for authentication. Credentials are encrypted using the + * integration runtime credential manager. Type: string (or Expression with resultType string). */ - clientSecret?: SecretBaseUnion; + encryptedCredential?: any; +} + +/** + * Netezza linked service. + */ +export interface NetezzaLinkedService { /** - * The service account email ID that is used for ServiceAuthentication and can only be used on - * self-hosted IR. + * Polymorphic Discriminator */ - email?: any; + type: "Netezza"; /** - * The full path to the .p12 key file that is used to authenticate the service account email - * address and can only be used on self-hosted IR. + * The integration runtime reference. */ - keyFilePath?: any; + connectVia?: IntegrationRuntimeReference; /** - * The full path of the .pem file containing trusted CA certificates for verifying the server - * when connecting over SSL. This property can only be set when using SSL on self-hosted IR. The - * default value is the cacerts.pem file installed with the IR. + * Linked service description. */ - trustedCertPath?: any; + description?: string; + /** + * Parameters for linked service. + */ + parameters?: { [propertyName: string]: ParameterSpecification }; + /** + * List of tags that can be used for describing the linked service. + */ + annotations?: any[]; + /** + * An ODBC connection string. Type: string, SecureString or AzureKeyVaultSecretReference. + */ + connectionString?: any; /** - * Specifies whether to use a CA certificate from the system trust store or from a specified PEM - * file. The default value is false. + * The Azure key vault secret reference of password in connection string. */ - useSystemTrustStore?: any; + pwd?: AzureKeyVaultSecretReference; /** * The encrypted credential used for authentication. Credentials are encrypted using the * integration runtime credential manager. Type: string (or Expression with resultType string). @@ -2753,13 +2898,13 @@ export interface GoogleAdWordsLinkedService { } /** - * Oracle Service Cloud linked service. + * Vertica linked service. */ -export interface OracleServiceCloudLinkedService { +export interface VerticaLinkedService { /** * Polymorphic Discriminator */ - type: "OracleServiceCloud"; + type: "Vertica"; /** * The integration runtime reference. */ @@ -2777,33 +2922,13 @@ export interface OracleServiceCloudLinkedService { */ annotations?: any[]; /** - * The URL of the Oracle Service Cloud instance. - */ - host: any; - /** - * The user name that you use to access Oracle Service Cloud server. - */ - username: any; - /** - * The password corresponding to the user name that you provided in the username key. - */ - password: SecretBaseUnion; - /** - * Specifies whether the data source endpoints are encrypted using HTTPS. The default value is - * true. Type: boolean (or Expression with resultType boolean). - */ - useEncryptedEndpoints?: any; - /** - * Specifies whether to require the host name in the server's certificate to match the host name - * of the server when connecting over SSL. The default value is true. Type: boolean (or - * Expression with resultType boolean). + * An ODBC connection string. Type: string, SecureString or AzureKeyVaultSecretReference. */ - useHostVerification?: any; + connectionString?: any; /** - * Specifies whether to verify the identity of the server when connecting over SSL. The default - * value is true. Type: boolean (or Expression with resultType boolean). + * The Azure key vault secret reference of password in connection string. */ - usePeerVerification?: any; + pwd?: AzureKeyVaultSecretReference; /** * The encrypted credential used for authentication. Credentials are encrypted using the * integration runtime credential manager. Type: string (or Expression with resultType string). @@ -2812,13 +2937,13 @@ export interface OracleServiceCloudLinkedService { } /** - * Dynamics AX linked service. + * Zoho server linked service. */ -export interface DynamicsAXLinkedService { +export interface ZohoLinkedService { /** * Polymorphic Discriminator */ - type: "DynamicsAX"; + type: "Zoho"; /** * The integration runtime reference. */ @@ -2836,30 +2961,28 @@ export interface DynamicsAXLinkedService { */ annotations?: any[]; /** - * The Dynamics AX (or Dynamics 365 Finance and Operations) instance OData endpoint. + * The endpoint of the Zoho server. (i.e. crm.zoho.com/crm/private) */ - url: any; + endpoint: any; /** - * Specify the application's client ID. Type: string (or Expression with resultType string). + * The access token for Zoho authentication. */ - servicePrincipalId: any; + accessToken?: SecretBaseUnion; /** - * Specify the application's key. Mark this field as a SecureString to store it securely in Data - * Factory, or reference a secret stored in Azure Key Vault. Type: string (or Expression with - * resultType string). + * Specifies whether the data source endpoints are encrypted using HTTPS. The default value is + * true. */ - servicePrincipalKey: SecretBaseUnion; + useEncryptedEndpoints?: any; /** - * Specify the tenant information (domain name or tenant ID) under which your application - * resides. Retrieve it by hovering the mouse in the top-right corner of the Azure portal. Type: - * string (or Expression with resultType string). + * Specifies whether to require the host name in the server's certificate to match the host name + * of the server when connecting over SSL. The default value is true. */ - tenant: any; + useHostVerification?: any; /** - * Specify the resource you are requesting authorization. Type: string (or Expression with - * resultType string). + * Specifies whether to verify the identity of the server when connecting over SSL. The default + * value is true. */ - aadResourceId: any; + usePeerVerification?: any; /** * The encrypted credential used for authentication. Credentials are encrypted using the * integration runtime credential manager. Type: string (or Expression with resultType string). @@ -2868,13 +2991,13 @@ export interface DynamicsAXLinkedService { } /** - * Responsys linked service. + * Xero Service linked service. */ -export interface ResponsysLinkedService { +export interface XeroLinkedService { /** * Polymorphic Discriminator */ - type: "Responsys"; + type: "Xero"; /** * The integration runtime reference. */ @@ -2892,33 +3015,32 @@ export interface ResponsysLinkedService { */ annotations?: any[]; /** - * The endpoint of the Responsys server. + * The endpoint of the Xero server. (i.e. api.xero.com) */ - endpoint: any; + host: any; /** - * The client ID associated with the Responsys application. Type: string (or Expression with - * resultType string). + * The consumer key associated with the Xero application. */ - clientId: any; + consumerKey?: SecretBaseUnion; /** - * The client secret associated with the Responsys application. Type: string (or Expression with - * resultType string). + * The private key from the .pem file that was generated for your Xero private application. You + * must include all the text from the .pem file, including the Unix line endings( + * ). */ - clientSecret?: SecretBaseUnion; + privateKey?: SecretBaseUnion; /** * Specifies whether the data source endpoints are encrypted using HTTPS. The default value is - * true. Type: boolean (or Expression with resultType boolean). + * true. */ useEncryptedEndpoints?: any; /** * Specifies whether to require the host name in the server's certificate to match the host name - * of the server when connecting over SSL. The default value is true. Type: boolean (or - * Expression with resultType boolean). + * of the server when connecting over SSL. The default value is true. */ useHostVerification?: any; /** * Specifies whether to verify the identity of the server when connecting over SSL. The default - * value is true. Type: boolean (or Expression with resultType boolean). + * value is true. */ usePeerVerification?: any; /** @@ -2929,13 +3051,13 @@ export interface ResponsysLinkedService { } /** - * Azure Databricks linked service. + * Square Service linked service. */ -export interface AzureDatabricksLinkedService { +export interface SquareLinkedService { /** * Polymorphic Discriminator */ - type: "AzureDatabricks"; + type: "Square"; /** * The integration runtime reference. */ @@ -2953,76 +3075,36 @@ export interface AzureDatabricksLinkedService { */ annotations?: any[]; /** - * .azuredatabricks.net, domain name of your Databricks deployment. Type: string (or - * Expression with resultType string). - */ - domain: any; - /** - * Access token for databricks REST API. Refer to - * https://docs.azuredatabricks.net/api/latest/authentication.html. Type: string (or Expression - * with resultType string). - */ - accessToken: SecretBaseUnion; - /** - * The id of an existing interactive cluster that will be used for all runs of this activity. - * Type: string (or Expression with resultType string). - */ - existingClusterId?: any; - /** - * The id of an existing instance pool that will be used for all runs of this activity. Type: - * string (or Expression with resultType string). - */ - instancePoolId?: any; - /** - * If not using an existing interactive cluster, this specifies the Spark version of a new job - * cluster or instance pool nodes created for each run of this activity. Required if - * instancePoolId is specified. Type: string (or Expression with resultType string). - */ - newClusterVersion?: any; - /** - * If not using an existing interactive cluster, this specifies the number of worker nodes to use - * for the new job cluster or instance pool. For new job clusters, this a string-formatted Int32, - * like '1' means numOfWorker is 1 or '1:10' means auto-scale from 1 (min) to 10 (max). For - * instance pools, this is a string-formatted Int32, and can only specify a fixed number of - * worker nodes, such as '2'. Required if newClusterVersion is specified. Type: string (or - * Expression with resultType string). - */ - newClusterNumOfWorker?: any; - /** - * The node type of the new job cluster. This property is required if newClusterVersion is - * specified and instancePoolId is not specified. If instancePoolId is specified, this property - * is ignored. Type: string (or Expression with resultType string). + * The URL of the Square instance. (i.e. mystore.mysquare.com) */ - newClusterNodeType?: any; + host: any; /** - * A set of optional, user-specified Spark configuration key-value pairs. + * The client ID associated with your Square application. */ - newClusterSparkConf?: { [propertyName: string]: any }; + clientId: any; /** - * A set of optional, user-specified Spark environment variables key-value pairs. + * The client secret associated with your Square application. */ - newClusterSparkEnvVars?: { [propertyName: string]: any }; + clientSecret?: SecretBaseUnion; /** - * Additional tags for cluster resources. This property is ignored in instance pool - * configurations. + * The redirect URL assigned in the Square application dashboard. (i.e. http://localhost:2500) */ - newClusterCustomTags?: { [propertyName: string]: any }; + redirectUri: any; /** - * The driver node type for the new job cluster. This property is ignored in instance pool - * configurations. Type: string (or Expression with resultType string). + * Specifies whether the data source endpoints are encrypted using HTTPS. The default value is + * true. */ - newClusterDriverNodeType?: any; + useEncryptedEndpoints?: any; /** - * User-defined initialization scripts for the new cluster. Type: array of strings (or Expression - * with resultType array of strings). + * Specifies whether to require the host name in the server's certificate to match the host name + * of the server when connecting over SSL. The default value is true. */ - newClusterInitScripts?: any; + useHostVerification?: any; /** - * Enable the elastic disk on the new cluster. This property is now ignored, and takes the - * default elastic disk behavior in Databricks (elastic disks are always enabled). Type: boolean - * (or Expression with resultType boolean). + * Specifies whether to verify the identity of the server when connecting over SSL. The default + * value is true. */ - newClusterEnableElasticDisk?: any; + usePeerVerification?: any; /** * The encrypted credential used for authentication. Credentials are encrypted using the * integration runtime credential manager. Type: string (or Expression with resultType string). @@ -3031,13 +3113,13 @@ export interface AzureDatabricksLinkedService { } /** - * Azure Data Lake Analytics linked service. + * Spark Server linked service. */ -export interface AzureDataLakeAnalyticsLinkedService { +export interface SparkLinkedService { /** * Polymorphic Discriminator */ - type: "AzureDataLakeAnalytics"; + type: "Spark"; /** * The integration runtime reference. */ @@ -3055,75 +3137,81 @@ export interface AzureDataLakeAnalyticsLinkedService { */ annotations?: any[]; /** - * The Azure Data Lake Analytics account name. Type: string (or Expression with resultType - * string). + * IP address or host name of the Spark server */ - accountName: any; + host: any; /** - * The ID of the application used to authenticate against the Azure Data Lake Analytics account. - * Type: string (or Expression with resultType string). + * The TCP port that the Spark server uses to listen for client connections. */ - servicePrincipalId?: any; + port: any; /** - * The Key of the application used to authenticate against the Azure Data Lake Analytics account. + * The type of Spark server. Possible values include: 'SharkServer', 'SharkServer2', + * 'SparkThriftServer' */ - servicePrincipalKey?: SecretBaseUnion; + serverType?: SparkServerType; /** - * The name or ID of the tenant to which the service principal belongs. Type: string (or - * Expression with resultType string). + * The transport protocol to use in the Thrift layer. Possible values include: 'Binary', 'SASL', + * 'HTTP ' */ - tenant: any; + thriftTransportProtocol?: SparkThriftTransportProtocol; /** - * Data Lake Analytics account subscription ID (if different from Data Factory account). Type: - * string (or Expression with resultType string). + * The authentication method used to access the Spark server. Possible values include: + * 'Anonymous', 'Username', 'UsernameAndPassword', 'WindowsAzureHDInsightService' */ - subscriptionId?: any; + authenticationType: SparkAuthenticationType; /** - * Data Lake Analytics account resource group name (if different from Data Factory account). - * Type: string (or Expression with resultType string). + * The user name that you use to access Spark Server. */ - resourceGroupName?: any; + username?: any; /** - * Azure Data Lake Analytics URI Type: string (or Expression with resultType string). + * The password corresponding to the user name that you provided in the Username field */ - dataLakeAnalyticsUri?: any; + password?: SecretBaseUnion; /** - * The encrypted credential used for authentication. Credentials are encrypted using the - * integration runtime credential manager. Type: string (or Expression with resultType string). + * The partial URL corresponding to the Spark server. */ - encryptedCredential?: any; -} - -/** - * Custom script action to run on HDI ondemand cluster once it's up. - */ -export interface ScriptAction { + httpPath?: any; /** - * The user provided name of the script action. + * Specifies whether the connections to the server are encrypted using SSL. The default value is + * false. */ - name: string; + enableSsl?: any; /** - * The URI for the script action. + * The full path of the .pem file containing trusted CA certificates for verifying the server + * when connecting over SSL. This property can only be set when using SSL on self-hosted IR. The + * default value is the cacerts.pem file installed with the IR. */ - uri: string; + trustedCertPath?: any; + /** + * Specifies whether to use a CA certificate from the system trust store or from a specified PEM + * file. The default value is false. + */ + useSystemTrustStore?: any; + /** + * Specifies whether to require a CA-issued SSL certificate name to match the host name of the + * server when connecting over SSL. The default value is false. + */ + allowHostNameCNMismatch?: any; /** - * The node types on which the script action should be executed. + * Specifies whether to allow self-signed certificates from the server. The default value is + * false. */ - roles: any; + allowSelfSignedServerCert?: any; /** - * The parameters for the script action. + * The encrypted credential used for authentication. Credentials are encrypted using the + * integration runtime credential manager. Type: string (or Expression with resultType string). */ - parameters?: string; + encryptedCredential?: any; } /** - * HDInsight ondemand linked service. + * Shopify Service linked service. */ -export interface HDInsightOnDemandLinkedService { +export interface ShopifyLinkedService { /** * Polymorphic Discriminator */ - type: "HDInsightOnDemand"; + type: "Shopify"; /** * The integration runtime reference. */ @@ -3141,166 +3229,174 @@ export interface HDInsightOnDemandLinkedService { */ annotations?: any[]; /** - * Number of worker/data nodes in the cluster. Suggestion value: 4. Type: string (or Expression - * with resultType string). + * The endpoint of the Shopify server. (i.e. mystore.myshopify.com) */ - clusterSize: any; + host: any; /** - * The allowed idle time for the on-demand HDInsight cluster. Specifies how long the on-demand - * HDInsight cluster stays alive after completion of an activity run if there are no other active - * jobs in the cluster. The minimum value is 5 mins. Type: string (or Expression with resultType - * string). + * The API access token that can be used to access Shopify’s data. The token won't expire if it + * is offline mode. */ - timeToLive: any; + accessToken?: SecretBaseUnion; /** - * Version of the HDInsight cluster.  Type: string (or Expression with resultType string). + * Specifies whether the data source endpoints are encrypted using HTTPS. The default value is + * true. */ - version: any; + useEncryptedEndpoints?: any; /** - * Azure Storage linked service to be used by the on-demand cluster for storing and processing - * data. + * Specifies whether to require the host name in the server's certificate to match the host name + * of the server when connecting over SSL. The default value is true. */ - linkedServiceName: LinkedServiceReference; + useHostVerification?: any; /** - * The customer’s subscription to host the cluster. Type: string (or Expression with resultType - * string). + * Specifies whether to verify the identity of the server when connecting over SSL. The default + * value is true. */ - hostSubscriptionId: any; + usePeerVerification?: any; /** - * The service principal id for the hostSubscriptionId. Type: string (or Expression with - * resultType string). + * The encrypted credential used for authentication. Credentials are encrypted using the + * integration runtime credential manager. Type: string (or Expression with resultType string). */ - servicePrincipalId?: any; + encryptedCredential?: any; +} + +/** + * ServiceNow server linked service. + */ +export interface ServiceNowLinkedService { /** - * The key for the service principal id. + * Polymorphic Discriminator */ - servicePrincipalKey?: SecretBaseUnion; + type: "ServiceNow"; /** - * The Tenant id/name to which the service principal belongs. Type: string (or Expression with - * resultType string). + * The integration runtime reference. */ - tenant: any; + connectVia?: IntegrationRuntimeReference; /** - * The resource group where the cluster belongs. Type: string (or Expression with resultType - * string). + * Linked service description. */ - clusterResourceGroup: any; + description?: string; /** - * The prefix of cluster name, postfix will be distinct with timestamp. Type: string (or - * Expression with resultType string). + * Parameters for linked service. */ - clusterNamePrefix?: any; + parameters?: { [propertyName: string]: ParameterSpecification }; /** - * The username to access the cluster. Type: string (or Expression with resultType string). + * List of tags that can be used for describing the linked service. */ - clusterUserName?: any; + annotations?: any[]; /** - * The password to access the cluster. + * The endpoint of the ServiceNow server. (i.e. .service-now.com) */ - clusterPassword?: SecretBaseUnion; + endpoint: any; /** - * The username to SSH remotely connect to cluster’s node (for Linux). Type: string (or - * Expression with resultType string). + * The authentication type to use. Possible values include: 'Basic', 'OAuth2' */ - clusterSshUserName?: any; + authenticationType: ServiceNowAuthenticationType; /** - * The password to SSH remotely connect cluster’s node (for Linux). + * The user name used to connect to the ServiceNow server for Basic and OAuth2 authentication. */ - clusterSshPassword?: SecretBaseUnion; + username?: any; /** - * Specifies additional storage accounts for the HDInsight linked service so that the Data - * Factory service can register them on your behalf. + * The password corresponding to the user name for Basic and OAuth2 authentication. */ - additionalLinkedServiceNames?: LinkedServiceReference[]; + password?: SecretBaseUnion; /** - * The name of Azure SQL linked service that point to the HCatalog database. The on-demand - * HDInsight cluster is created by using the Azure SQL database as the metastore. + * The client id for OAuth2 authentication. */ - hcatalogLinkedServiceName?: LinkedServiceReference; + clientId?: any; /** - * The cluster type. Type: string (or Expression with resultType string). + * The client secret for OAuth2 authentication. */ - clusterType?: any; + clientSecret?: SecretBaseUnion; /** - * The version of spark if the cluster type is 'spark'. Type: string (or Expression with - * resultType string). + * Specifies whether the data source endpoints are encrypted using HTTPS. The default value is + * true. */ - sparkVersion?: any; + useEncryptedEndpoints?: any; /** - * Specifies the core configuration parameters (as in core-site.xml) for the HDInsight cluster to - * be created. + * Specifies whether to require the host name in the server's certificate to match the host name + * of the server when connecting over SSL. The default value is true. */ - coreConfiguration?: any; + useHostVerification?: any; /** - * Specifies the HBase configuration parameters (hbase-site.xml) for the HDInsight cluster. + * Specifies whether to verify the identity of the server when connecting over SSL. The default + * value is true. */ - hBaseConfiguration?: any; + usePeerVerification?: any; /** - * Specifies the HDFS configuration parameters (hdfs-site.xml) for the HDInsight cluster. + * The encrypted credential used for authentication. Credentials are encrypted using the + * integration runtime credential manager. Type: string (or Expression with resultType string). */ - hdfsConfiguration?: any; + encryptedCredential?: any; +} + +/** + * QuickBooks server linked service. + */ +export interface QuickBooksLinkedService { /** - * Specifies the hive configuration parameters (hive-site.xml) for the HDInsight cluster. + * Polymorphic Discriminator */ - hiveConfiguration?: any; + type: "QuickBooks"; /** - * Specifies the MapReduce configuration parameters (mapred-site.xml) for the HDInsight cluster. + * The integration runtime reference. */ - mapReduceConfiguration?: any; + connectVia?: IntegrationRuntimeReference; /** - * Specifies the Oozie configuration parameters (oozie-site.xml) for the HDInsight cluster. + * Linked service description. */ - oozieConfiguration?: any; + description?: string; /** - * Specifies the Storm configuration parameters (storm-site.xml) for the HDInsight cluster. + * Parameters for linked service. */ - stormConfiguration?: any; + parameters?: { [propertyName: string]: ParameterSpecification }; /** - * Specifies the Yarn configuration parameters (yarn-site.xml) for the HDInsight cluster. + * List of tags that can be used for describing the linked service. */ - yarnConfiguration?: any; + annotations?: any[]; /** - * The encrypted credential used for authentication. Credentials are encrypted using the - * integration runtime credential manager. Type: string (or Expression with resultType string). + * The endpoint of the QuickBooks server. (i.e. quickbooks.api.intuit.com) */ - encryptedCredential?: any; + endpoint: any; /** - * Specifies the size of the head node for the HDInsight cluster. + * The company ID of the QuickBooks company to authorize. */ - headNodeSize?: any; + companyId: any; /** - * Specifies the size of the data node for the HDInsight cluster. + * The consumer key for OAuth 1.0 authentication. */ - dataNodeSize?: any; + consumerKey: any; /** - * Specifies the size of the Zoo Keeper node for the HDInsight cluster. + * The consumer secret for OAuth 1.0 authentication. */ - zookeeperNodeSize?: any; + consumerSecret: SecretBaseUnion; /** - * Custom script actions to run on HDI ondemand cluster once it's up. Please refer to - * https://docs.microsoft.com/en-us/azure/hdinsight/hdinsight-hadoop-customize-cluster-linux?toc=%2Fen-us%2Fazure%2Fhdinsight%2Fr-server%2FTOC.json&bc=%2Fen-us%2Fazure%2Fbread%2Ftoc.json#understanding-script-actions. + * The access token for OAuth 1.0 authentication. */ - scriptActions?: ScriptAction[]; + accessToken: SecretBaseUnion; /** - * The ARM resource ID for the vNet to which the cluster should be joined after creation. Type: - * string (or Expression with resultType string). + * The access token secret for OAuth 1.0 authentication. */ - virtualNetworkId?: any; + accessTokenSecret: SecretBaseUnion; /** - * The ARM resource ID for the subnet in the vNet. If virtualNetworkId was specified, then this - * property is required. Type: string (or Expression with resultType string). + * Specifies whether the data source endpoints are encrypted using HTTPS. The default value is + * true. */ - subnetName?: any; + useEncryptedEndpoints?: any; + /** + * The encrypted credential used for authentication. Credentials are encrypted using the + * integration runtime credential manager. Type: string (or Expression with resultType string). + */ + encryptedCredential?: any; } /** - * Salesforce Marketing Cloud linked service. + * Presto server linked service. */ -export interface SalesforceMarketingCloudLinkedService { +export interface PrestoLinkedService { /** * Polymorphic Discriminator */ - type: "SalesforceMarketingCloud"; + type: "Presto"; /** * The integration runtime reference. */ @@ -3318,31 +3414,66 @@ export interface SalesforceMarketingCloudLinkedService { */ annotations?: any[]; /** - * The client ID associated with the Salesforce Marketing Cloud application. Type: string (or - * Expression with resultType string). + * The IP address or host name of the Presto server. (i.e. 192.168.222.160) */ - clientId: any; + host: any; /** - * The client secret associated with the Salesforce Marketing Cloud application. Type: string (or - * Expression with resultType string). + * The version of the Presto server. (i.e. 0.148-t) */ - clientSecret?: SecretBaseUnion; + serverVersion: any; /** - * Specifies whether the data source endpoints are encrypted using HTTPS. The default value is - * true. Type: boolean (or Expression with resultType boolean). + * The catalog context for all request against the server. */ - useEncryptedEndpoints?: any; + catalog: any; /** - * Specifies whether to require the host name in the server's certificate to match the host name - * of the server when connecting over SSL. The default value is true. Type: boolean (or - * Expression with resultType boolean). + * The TCP port that the Presto server uses to listen for client connections. The default value + * is 8080. + */ + port?: any; + /** + * The authentication mechanism used to connect to the Presto server. Possible values include: + * 'Anonymous', 'LDAP' + */ + authenticationType: PrestoAuthenticationType; + /** + * The user name used to connect to the Presto server. + */ + username?: any; + /** + * The password corresponding to the user name. + */ + password?: SecretBaseUnion; + /** + * Specifies whether the connections to the server are encrypted using SSL. The default value is + * false. + */ + enableSsl?: any; + /** + * The full path of the .pem file containing trusted CA certificates for verifying the server + * when connecting over SSL. This property can only be set when using SSL on self-hosted IR. The + * default value is the cacerts.pem file installed with the IR. + */ + trustedCertPath?: any; + /** + * Specifies whether to use a CA certificate from the system trust store or from a specified PEM + * file. The default value is false. + */ + useSystemTrustStore?: any; + /** + * Specifies whether to require a CA-issued SSL certificate name to match the host name of the + * server when connecting over SSL. The default value is false. + */ + allowHostNameCNMismatch?: any; + /** + * Specifies whether to allow self-signed certificates from the server. The default value is + * false. */ - useHostVerification?: any; + allowSelfSignedServerCert?: any; /** - * Specifies whether to verify the identity of the server when connecting over SSL. The default - * value is true. Type: boolean (or Expression with resultType boolean). + * The local time zone used by the connection. Valid values for this option are specified in the + * IANA Time Zone Database. The default value is the system time zone. */ - usePeerVerification?: any; + timeZoneID?: any; /** * The encrypted credential used for authentication. Credentials are encrypted using the * integration runtime credential manager. Type: string (or Expression with resultType string). @@ -3351,13 +3482,13 @@ export interface SalesforceMarketingCloudLinkedService { } /** - * Netezza linked service. + * Phoenix server linked service. */ -export interface NetezzaLinkedService { +export interface PhoenixLinkedService { /** * Polymorphic Discriminator */ - type: "Netezza"; + type: "Phoenix"; /** * The integration runtime reference. */ @@ -3375,52 +3506,58 @@ export interface NetezzaLinkedService { */ annotations?: any[]; /** - * An ODBC connection string. Type: string, SecureString or AzureKeyVaultSecretReference. + * The IP address or host name of the Phoenix server. (i.e. 192.168.222.160) */ - connectionString?: any; + host: any; /** - * The Azure key vault secret reference of password in connection string. + * The TCP port that the Phoenix server uses to listen for client connections. The default value + * is 8765. */ - pwd?: AzureKeyVaultSecretReference; + port?: any; /** - * The encrypted credential used for authentication. Credentials are encrypted using the - * integration runtime credential manager. Type: string (or Expression with resultType string). + * The partial URL corresponding to the Phoenix server. (i.e. /gateway/sandbox/phoenix/version). + * The default value is hbasephoenix if using WindowsAzureHDInsightService. */ - encryptedCredential?: any; -} - -/** - * Vertica linked service. - */ -export interface VerticaLinkedService { + httpPath?: any; /** - * Polymorphic Discriminator + * The authentication mechanism used to connect to the Phoenix server. Possible values include: + * 'Anonymous', 'UsernameAndPassword', 'WindowsAzureHDInsightService' */ - type: "Vertica"; + authenticationType: PhoenixAuthenticationType; /** - * The integration runtime reference. + * The user name used to connect to the Phoenix server. */ - connectVia?: IntegrationRuntimeReference; + username?: any; /** - * Linked service description. + * The password corresponding to the user name. */ - description?: string; + password?: SecretBaseUnion; /** - * Parameters for linked service. + * Specifies whether the connections to the server are encrypted using SSL. The default value is + * false. */ - parameters?: { [propertyName: string]: ParameterSpecification }; + enableSsl?: any; /** - * List of tags that can be used for describing the linked service. + * The full path of the .pem file containing trusted CA certificates for verifying the server + * when connecting over SSL. This property can only be set when using SSL on self-hosted IR. The + * default value is the cacerts.pem file installed with the IR. */ - annotations?: any[]; + trustedCertPath?: any; /** - * An ODBC connection string. Type: string, SecureString or AzureKeyVaultSecretReference. + * Specifies whether to use a CA certificate from the system trust store or from a specified PEM + * file. The default value is false. */ - connectionString?: any; + useSystemTrustStore?: any; /** - * The Azure key vault secret reference of password in connection string. + * Specifies whether to require a CA-issued SSL certificate name to match the host name of the + * server when connecting over SSL. The default value is false. */ - pwd?: AzureKeyVaultSecretReference; + allowHostNameCNMismatch?: any; + /** + * Specifies whether to allow self-signed certificates from the server. The default value is + * false. + */ + allowSelfSignedServerCert?: any; /** * The encrypted credential used for authentication. Credentials are encrypted using the * integration runtime credential manager. Type: string (or Expression with resultType string). @@ -3429,13 +3566,13 @@ export interface VerticaLinkedService { } /** - * Zoho server linked service. + * Paypal Service linked service. */ -export interface ZohoLinkedService { +export interface PaypalLinkedService { /** * Polymorphic Discriminator */ - type: "Zoho"; + type: "Paypal"; /** * The integration runtime reference. */ @@ -3453,13 +3590,17 @@ export interface ZohoLinkedService { */ annotations?: any[]; /** - * The endpoint of the Zoho server. (i.e. crm.zoho.com/crm/private) + * The URL of the PayPal instance. (i.e. api.sandbox.paypal.com) */ - endpoint: any; + host: any; /** - * The access token for Zoho authentication. + * The client ID associated with your PayPal application. */ - accessToken?: SecretBaseUnion; + clientId: any; + /** + * The client secret associated with your PayPal application. + */ + clientSecret?: SecretBaseUnion; /** * Specifies whether the data source endpoints are encrypted using HTTPS. The default value is * true. @@ -3483,13 +3624,13 @@ export interface ZohoLinkedService { } /** - * Xero Service linked service. + * Marketo server linked service. */ -export interface XeroLinkedService { +export interface MarketoLinkedService { /** * Polymorphic Discriminator */ - type: "Xero"; + type: "Marketo"; /** * The integration runtime reference. */ @@ -3507,19 +3648,17 @@ export interface XeroLinkedService { */ annotations?: any[]; /** - * The endpoint of the Xero server. (i.e. api.xero.com) + * The endpoint of the Marketo server. (i.e. 123-ABC-321.mktorest.com) */ - host: any; + endpoint: any; /** - * The consumer key associated with the Xero application. + * The client Id of your Marketo service. */ - consumerKey?: SecretBaseUnion; + clientId: any; /** - * The private key from the .pem file that was generated for your Xero private application. You - * must include all the text from the .pem file, including the Unix line endings( - * ). + * The client secret of your Marketo service. */ - privateKey?: SecretBaseUnion; + clientSecret?: SecretBaseUnion; /** * Specifies whether the data source endpoints are encrypted using HTTPS. The default value is * true. @@ -3543,13 +3682,13 @@ export interface XeroLinkedService { } /** - * Square Service linked service. + * Azure Database for MariaDB linked service. */ -export interface SquareLinkedService { +export interface AzureMariaDBLinkedService { /** * Polymorphic Discriminator */ - type: "Square"; + type: "AzureMariaDB"; /** * The integration runtime reference. */ @@ -3567,36 +3706,13 @@ export interface SquareLinkedService { */ annotations?: any[]; /** - * The URL of the Square instance. (i.e. mystore.mysquare.com) - */ - host: any; - /** - * The client ID associated with your Square application. - */ - clientId: any; - /** - * The client secret associated with your Square application. - */ - clientSecret?: SecretBaseUnion; - /** - * The redirect URL assigned in the Square application dashboard. (i.e. http://localhost:2500) - */ - redirectUri: any; - /** - * Specifies whether the data source endpoints are encrypted using HTTPS. The default value is - * true. - */ - useEncryptedEndpoints?: any; - /** - * Specifies whether to require the host name in the server's certificate to match the host name - * of the server when connecting over SSL. The default value is true. + * An ODBC connection string. Type: string, SecureString or AzureKeyVaultSecretReference. */ - useHostVerification?: any; + connectionString?: any; /** - * Specifies whether to verify the identity of the server when connecting over SSL. The default - * value is true. + * The Azure key vault secret reference of password in connection string. */ - usePeerVerification?: any; + pwd?: AzureKeyVaultSecretReference; /** * The encrypted credential used for authentication. Credentials are encrypted using the * integration runtime credential manager. Type: string (or Expression with resultType string). @@ -3605,13 +3721,13 @@ export interface SquareLinkedService { } /** - * Spark Server linked service. + * MariaDB server linked service. */ -export interface SparkLinkedService { +export interface MariaDBLinkedService { /** * Polymorphic Discriminator */ - type: "Spark"; + type: "MariaDB"; /** * The integration runtime reference. */ @@ -3629,66 +3745,67 @@ export interface SparkLinkedService { */ annotations?: any[]; /** - * IP address or host name of the Spark server + * An ODBC connection string. Type: string, SecureString or AzureKeyVaultSecretReference. */ - host: any; + connectionString?: any; /** - * The TCP port that the Spark server uses to listen for client connections. + * The Azure key vault secret reference of password in connection string. */ - port: any; + pwd?: AzureKeyVaultSecretReference; /** - * The type of Spark server. Possible values include: 'SharkServer', 'SharkServer2', - * 'SparkThriftServer' + * The encrypted credential used for authentication. Credentials are encrypted using the + * integration runtime credential manager. Type: string (or Expression with resultType string). */ - serverType?: SparkServerType; + encryptedCredential?: any; +} + +/** + * Magento server linked service. + */ +export interface MagentoLinkedService { /** - * The transport protocol to use in the Thrift layer. Possible values include: 'Binary', 'SASL', - * 'HTTP ' + * Polymorphic Discriminator */ - thriftTransportProtocol?: SparkThriftTransportProtocol; + type: "Magento"; /** - * The authentication method used to access the Spark server. Possible values include: - * 'Anonymous', 'Username', 'UsernameAndPassword', 'WindowsAzureHDInsightService' + * The integration runtime reference. */ - authenticationType: SparkAuthenticationType; + connectVia?: IntegrationRuntimeReference; /** - * The user name that you use to access Spark Server. + * Linked service description. */ - username?: any; + description?: string; /** - * The password corresponding to the user name that you provided in the Username field + * Parameters for linked service. */ - password?: SecretBaseUnion; + parameters?: { [propertyName: string]: ParameterSpecification }; /** - * The partial URL corresponding to the Spark server. + * List of tags that can be used for describing the linked service. */ - httpPath?: any; + annotations?: any[]; /** - * Specifies whether the connections to the server are encrypted using SSL. The default value is - * false. + * The URL of the Magento instance. (i.e. 192.168.222.110/magento3) */ - enableSsl?: any; + host: any; /** - * The full path of the .pem file containing trusted CA certificates for verifying the server - * when connecting over SSL. This property can only be set when using SSL on self-hosted IR. The - * default value is the cacerts.pem file installed with the IR. + * The access token from Magento. */ - trustedCertPath?: any; + accessToken?: SecretBaseUnion; /** - * Specifies whether to use a CA certificate from the system trust store or from a specified PEM - * file. The default value is false. + * Specifies whether the data source endpoints are encrypted using HTTPS. The default value is + * true. */ - useSystemTrustStore?: any; + useEncryptedEndpoints?: any; /** - * Specifies whether to require a CA-issued SSL certificate name to match the host name of the - * server when connecting over SSL. The default value is false. + * Specifies whether to require the host name in the server's certificate to match the host name + * of the server when connecting over SSL. The default value is true. */ - allowHostNameCNMismatch?: any; + useHostVerification?: any; /** - * Specifies whether to allow self-signed certificates from the server. The default value is - * false. + * Specifies whether to verify the identity of the server when connecting over SSL. The default + * value is true. */ - allowSelfSignedServerCert?: any; + usePeerVerification?: any; /** * The encrypted credential used for authentication. Credentials are encrypted using the * integration runtime credential manager. Type: string (or Expression with resultType string). @@ -3697,13 +3814,13 @@ export interface SparkLinkedService { } /** - * Shopify Service linked service. + * Jira Service linked service. */ -export interface ShopifyLinkedService { +export interface JiraLinkedService { /** * Polymorphic Discriminator */ - type: "Shopify"; + type: "Jira"; /** * The integration runtime reference. */ @@ -3721,14 +3838,22 @@ export interface ShopifyLinkedService { */ annotations?: any[]; /** - * The endpoint of the Shopify server. (i.e. mystore.myshopify.com) + * The IP address or host name of the Jira service. (e.g. jira.example.com) */ host: any; /** - * The API access token that can be used to access Shopify’s data. The token won't expire if it - * is offline mode. + * The TCP port that the Jira server uses to listen for client connections. The default value is + * 443 if connecting through HTTPS, or 8080 if connecting through HTTP. */ - accessToken?: SecretBaseUnion; + port?: any; + /** + * The user name that you use to access Jira Service. + */ + username: any; + /** + * The password corresponding to the user name that you provided in the username field. + */ + password?: SecretBaseUnion; /** * Specifies whether the data source endpoints are encrypted using HTTPS. The default value is * true. @@ -3752,13 +3877,13 @@ export interface ShopifyLinkedService { } /** - * ServiceNow server linked service. + * Impala server linked service. */ -export interface ServiceNowLinkedService { +export interface ImpalaLinkedService { /** * Polymorphic Discriminator */ - type: "ServiceNow"; + type: "Impala"; /** * The integration runtime reference. */ @@ -3776,44 +3901,54 @@ export interface ServiceNowLinkedService { */ annotations?: any[]; /** - * The endpoint of the ServiceNow server. (i.e. .service-now.com) + * The IP address or host name of the Impala server. (i.e. 192.168.222.160) */ - endpoint: any; + host: any; /** - * The authentication type to use. Possible values include: 'Basic', 'OAuth2' + * The TCP port that the Impala server uses to listen for client connections. The default value + * is 21050. */ - authenticationType: ServiceNowAuthenticationType; + port?: any; /** - * The user name used to connect to the ServiceNow server for Basic and OAuth2 authentication. + * The authentication type to use. Possible values include: 'Anonymous', 'SASLUsername', + * 'UsernameAndPassword' + */ + authenticationType: ImpalaAuthenticationType; + /** + * The user name used to access the Impala server. The default value is anonymous when using + * SASLUsername. */ username?: any; /** - * The password corresponding to the user name for Basic and OAuth2 authentication. + * The password corresponding to the user name when using UsernameAndPassword. */ password?: SecretBaseUnion; /** - * The client id for OAuth2 authentication. + * Specifies whether the connections to the server are encrypted using SSL. The default value is + * false. */ - clientId?: any; + enableSsl?: any; /** - * The client secret for OAuth2 authentication. + * The full path of the .pem file containing trusted CA certificates for verifying the server + * when connecting over SSL. This property can only be set when using SSL on self-hosted IR. The + * default value is the cacerts.pem file installed with the IR. */ - clientSecret?: SecretBaseUnion; + trustedCertPath?: any; /** - * Specifies whether the data source endpoints are encrypted using HTTPS. The default value is - * true. + * Specifies whether to use a CA certificate from the system trust store or from a specified PEM + * file. The default value is false. */ - useEncryptedEndpoints?: any; + useSystemTrustStore?: any; /** - * Specifies whether to require the host name in the server's certificate to match the host name - * of the server when connecting over SSL. The default value is true. + * Specifies whether to require a CA-issued SSL certificate name to match the host name of the + * server when connecting over SSL. The default value is false. */ - useHostVerification?: any; + allowHostNameCNMismatch?: any; /** - * Specifies whether to verify the identity of the server when connecting over SSL. The default - * value is true. + * Specifies whether to allow self-signed certificates from the server. The default value is + * false. */ - usePeerVerification?: any; + allowSelfSignedServerCert?: any; /** * The encrypted credential used for authentication. Credentials are encrypted using the * integration runtime credential manager. Type: string (or Expression with resultType string). @@ -3822,13 +3957,13 @@ export interface ServiceNowLinkedService { } /** - * QuickBooks server linked service. + * Hubspot Service linked service. */ -export interface QuickBooksLinkedService { +export interface HubspotLinkedService { /** * Polymorphic Discriminator */ - type: "QuickBooks"; + type: "Hubspot"; /** * The integration runtime reference. */ @@ -3846,34 +3981,36 @@ export interface QuickBooksLinkedService { */ annotations?: any[]; /** - * The endpoint of the QuickBooks server. (i.e. quickbooks.api.intuit.com) + * The client ID associated with your Hubspot application. */ - endpoint: any; + clientId: any; /** - * The company ID of the QuickBooks company to authorize. + * The client secret associated with your Hubspot application. */ - companyId: any; + clientSecret?: SecretBaseUnion; /** - * The consumer key for OAuth 1.0 authentication. + * The access token obtained when initially authenticating your OAuth integration. */ - consumerKey: any; + accessToken?: SecretBaseUnion; /** - * The consumer secret for OAuth 1.0 authentication. + * The refresh token obtained when initially authenticating your OAuth integration. */ - consumerSecret: SecretBaseUnion; + refreshToken?: SecretBaseUnion; /** - * The access token for OAuth 1.0 authentication. + * Specifies whether the data source endpoints are encrypted using HTTPS. The default value is + * true. */ - accessToken: SecretBaseUnion; + useEncryptedEndpoints?: any; /** - * The access token secret for OAuth 1.0 authentication. + * Specifies whether to require the host name in the server's certificate to match the host name + * of the server when connecting over SSL. The default value is true. */ - accessTokenSecret: SecretBaseUnion; + useHostVerification?: any; /** - * Specifies whether the data source endpoints are encrypted using HTTPS. The default value is - * true. + * Specifies whether to verify the identity of the server when connecting over SSL. The default + * value is true. */ - useEncryptedEndpoints?: any; + usePeerVerification?: any; /** * The encrypted credential used for authentication. Credentials are encrypted using the * integration runtime credential manager. Type: string (or Expression with resultType string). @@ -3882,13 +4019,13 @@ export interface QuickBooksLinkedService { } /** - * Presto server linked service. + * Hive Server linked service. */ -export interface PrestoLinkedService { +export interface HiveLinkedService { /** * Polymorphic Discriminator */ - type: "Presto"; + type: "Hive"; /** * The integration runtime reference. */ @@ -3906,35 +4043,54 @@ export interface PrestoLinkedService { */ annotations?: any[]; /** - * The IP address or host name of the Presto server. (i.e. 192.168.222.160) + * IP address or host name of the Hive server, separated by ';' for multiple hosts (only when + * serviceDiscoveryMode is enable). */ host: any; /** - * The version of the Presto server. (i.e. 0.148-t) + * The TCP port that the Hive server uses to listen for client connections. */ - serverVersion: any; + port?: any; /** - * The catalog context for all request against the server. + * The type of Hive server. Possible values include: 'HiveServer1', 'HiveServer2', + * 'HiveThriftServer' */ - catalog: any; + serverType?: HiveServerType; /** - * The TCP port that the Presto server uses to listen for client connections. The default value - * is 8080. + * The transport protocol to use in the Thrift layer. Possible values include: 'Binary', 'SASL', + * 'HTTP ' */ - port?: any; + thriftTransportProtocol?: HiveThriftTransportProtocol; /** - * The authentication mechanism used to connect to the Presto server. Possible values include: - * 'Anonymous', 'LDAP' + * The authentication method used to access the Hive server. Possible values include: + * 'Anonymous', 'Username', 'UsernameAndPassword', 'WindowsAzureHDInsightService' */ - authenticationType: PrestoAuthenticationType; + authenticationType: HiveAuthenticationType; /** - * The user name used to connect to the Presto server. + * true to indicate using the ZooKeeper service, false not. + */ + serviceDiscoveryMode?: any; + /** + * The namespace on ZooKeeper under which Hive Server 2 nodes are added. + */ + zooKeeperNameSpace?: any; + /** + * Specifies whether the driver uses native HiveQL queries,or converts them into an equivalent + * form in HiveQL. + */ + useNativeQuery?: any; + /** + * The user name that you use to access Hive Server. */ username?: any; /** - * The password corresponding to the user name. + * The password corresponding to the user name that you provided in the Username field */ password?: SecretBaseUnion; + /** + * The partial URL corresponding to the Hive server. + */ + httpPath?: any; /** * Specifies whether the connections to the server are encrypted using SSL. The default value is * false. @@ -3961,11 +4117,6 @@ export interface PrestoLinkedService { * false. */ allowSelfSignedServerCert?: any; - /** - * The local time zone used by the connection. Valid values for this option are specified in the - * IANA Time Zone Database. The default value is the system time zone. - */ - timeZoneID?: any; /** * The encrypted credential used for authentication. Credentials are encrypted using the * integration runtime credential manager. Type: string (or Expression with resultType string). @@ -3974,13 +4125,13 @@ export interface PrestoLinkedService { } /** - * Phoenix server linked service. + * HBase server linked service. */ -export interface PhoenixLinkedService { +export interface HBaseLinkedService { /** * Polymorphic Discriminator */ - type: "Phoenix"; + type: "HBase"; /** * The integration runtime reference. */ @@ -3998,26 +4149,25 @@ export interface PhoenixLinkedService { */ annotations?: any[]; /** - * The IP address or host name of the Phoenix server. (i.e. 192.168.222.160) + * The IP address or host name of the HBase server. (i.e. 192.168.222.160) */ host: any; /** - * The TCP port that the Phoenix server uses to listen for client connections. The default value - * is 8765. + * The TCP port that the HBase instance uses to listen for client connections. The default value + * is 9090. */ port?: any; /** - * The partial URL corresponding to the Phoenix server. (i.e. /gateway/sandbox/phoenix/version). - * The default value is hbasephoenix if using WindowsAzureHDInsightService. + * The partial URL corresponding to the HBase server. (i.e. /gateway/sandbox/hbase/version) */ httpPath?: any; /** - * The authentication mechanism used to connect to the Phoenix server. Possible values include: - * 'Anonymous', 'UsernameAndPassword', 'WindowsAzureHDInsightService' + * The authentication mechanism to use to connect to the HBase server. Possible values include: + * 'Anonymous', 'Basic' */ - authenticationType: PhoenixAuthenticationType; + authenticationType: HBaseAuthenticationType; /** - * The user name used to connect to the Phoenix server. + * The user name used to connect to the HBase instance. */ username?: any; /** @@ -4035,11 +4185,6 @@ export interface PhoenixLinkedService { * default value is the cacerts.pem file installed with the IR. */ trustedCertPath?: any; - /** - * Specifies whether to use a CA certificate from the system trust store or from a specified PEM - * file. The default value is false. - */ - useSystemTrustStore?: any; /** * Specifies whether to require a CA-issued SSL certificate name to match the host name of the * server when connecting over SSL. The default value is false. @@ -4058,13 +4203,13 @@ export interface PhoenixLinkedService { } /** - * Paypal Service linked service. + * Greenplum Database linked service. */ -export interface PaypalLinkedService { +export interface GreenplumLinkedService { /** * Polymorphic Discriminator */ - type: "Paypal"; + type: "Greenplum"; /** * The integration runtime reference. */ @@ -4082,32 +4227,99 @@ export interface PaypalLinkedService { */ annotations?: any[]; /** - * The URL of the PayPal instance. (i.e. api.sandbox.paypal.com) + * An ODBC connection string. Type: string, SecureString or AzureKeyVaultSecretReference. */ - host: any; + connectionString?: any; /** - * The client ID associated with your PayPal application. + * The Azure key vault secret reference of password in connection string. */ - clientId: any; + pwd?: AzureKeyVaultSecretReference; /** - * The client secret associated with your PayPal application. + * The encrypted credential used for authentication. Credentials are encrypted using the + * integration runtime credential manager. Type: string (or Expression with resultType string). + */ + encryptedCredential?: any; +} + +/** + * Google BigQuery service linked service. + */ +export interface GoogleBigQueryLinkedService { + /** + * Polymorphic Discriminator + */ + type: "GoogleBigQuery"; + /** + * The integration runtime reference. + */ + connectVia?: IntegrationRuntimeReference; + /** + * Linked service description. + */ + description?: string; + /** + * Parameters for linked service. + */ + parameters?: { [propertyName: string]: ParameterSpecification }; + /** + * List of tags that can be used for describing the linked service. + */ + annotations?: any[]; + /** + * The default BigQuery project to query against. + */ + project: any; + /** + * A comma-separated list of public BigQuery projects to access. + */ + additionalProjects?: any; + /** + * Whether to request access to Google Drive. Allowing Google Drive access enables support for + * federated tables that combine BigQuery data with data from Google Drive. The default value is + * false. + */ + requestGoogleDriveScope?: any; + /** + * The OAuth 2.0 authentication mechanism used for authentication. ServiceAuthentication can only + * be used on self-hosted IR. Possible values include: 'ServiceAuthentication', + * 'UserAuthentication' + */ + authenticationType: GoogleBigQueryAuthenticationType; + /** + * The refresh token obtained from Google for authorizing access to BigQuery for + * UserAuthentication. + */ + refreshToken?: SecretBaseUnion; + /** + * The client id of the google application used to acquire the refresh token. Type: string (or + * Expression with resultType string). + */ + clientId?: any; + /** + * The client secret of the google application used to acquire the refresh token. */ clientSecret?: SecretBaseUnion; /** - * Specifies whether the data source endpoints are encrypted using HTTPS. The default value is - * true. + * The service account email ID that is used for ServiceAuthentication and can only be used on + * self-hosted IR. + */ + email?: any; + /** + * The full path to the .p12 key file that is used to authenticate the service account email + * address and can only be used on self-hosted IR. */ - useEncryptedEndpoints?: any; + keyFilePath?: any; /** - * Specifies whether to require the host name in the server's certificate to match the host name - * of the server when connecting over SSL. The default value is true. + * The full path of the .pem file containing trusted CA certificates for verifying the server + * when connecting over SSL. This property can only be set when using SSL on self-hosted IR. The + * default value is the cacerts.pem file installed with the IR. */ - useHostVerification?: any; + trustedCertPath?: any; /** - * Specifies whether to verify the identity of the server when connecting over SSL. The default - * value is true. + * Specifies whether to use a CA certificate from the system trust store or from a specified PEM + * file. The default value is false. */ - usePeerVerification?: any; + useSystemTrustStore?: any; /** * The encrypted credential used for authentication. Credentials are encrypted using the * integration runtime credential manager. Type: string (or Expression with resultType string). @@ -4116,13 +4328,13 @@ export interface PaypalLinkedService { } /** - * Marketo server linked service. + * Eloqua server linked service. */ -export interface MarketoLinkedService { +export interface EloquaLinkedService { /** * Polymorphic Discriminator */ - type: "Marketo"; + type: "Eloqua"; /** * The integration runtime reference. */ @@ -4140,17 +4352,18 @@ export interface MarketoLinkedService { */ annotations?: any[]; /** - * The endpoint of the Marketo server. (i.e. 123-ABC-321.mktorest.com) + * The endpoint of the Eloqua server. (i.e. eloqua.example.com) */ endpoint: any; /** - * The client Id of your Marketo service. + * The site name and user name of your Eloqua account in the form: sitename/username. (i.e. + * Eloqua/Alice) */ - clientId: any; + username: any; /** - * The client secret of your Marketo service. + * The password corresponding to the user name. */ - clientSecret?: SecretBaseUnion; + password?: SecretBaseUnion; /** * Specifies whether the data source endpoints are encrypted using HTTPS. The default value is * true. @@ -4174,13 +4387,13 @@ export interface MarketoLinkedService { } /** - * Azure Database for MariaDB linked service. + * Drill server linked service. */ -export interface AzureMariaDBLinkedService { +export interface DrillLinkedService { /** * Polymorphic Discriminator */ - type: "AzureMariaDB"; + type: "Drill"; /** * The integration runtime reference. */ @@ -4213,13 +4426,13 @@ export interface AzureMariaDBLinkedService { } /** - * MariaDB server linked service. + * Couchbase server linked service. */ -export interface MariaDBLinkedService { +export interface CouchbaseLinkedService { /** * Polymorphic Discriminator */ - type: "MariaDB"; + type: "Couchbase"; /** * The integration runtime reference. */ @@ -4241,63 +4454,9 @@ export interface MariaDBLinkedService { */ connectionString?: any; /** - * The Azure key vault secret reference of password in connection string. - */ - pwd?: AzureKeyVaultSecretReference; - /** - * The encrypted credential used for authentication. Credentials are encrypted using the - * integration runtime credential manager. Type: string (or Expression with resultType string). - */ - encryptedCredential?: any; -} - -/** - * Magento server linked service. - */ -export interface MagentoLinkedService { - /** - * Polymorphic Discriminator - */ - type: "Magento"; - /** - * The integration runtime reference. - */ - connectVia?: IntegrationRuntimeReference; - /** - * Linked service description. - */ - description?: string; - /** - * Parameters for linked service. - */ - parameters?: { [propertyName: string]: ParameterSpecification }; - /** - * List of tags that can be used for describing the linked service. - */ - annotations?: any[]; - /** - * The URL of the Magento instance. (i.e. 192.168.222.110/magento3) - */ - host: any; - /** - * The access token from Magento. - */ - accessToken?: SecretBaseUnion; - /** - * Specifies whether the data source endpoints are encrypted using HTTPS. The default value is - * true. - */ - useEncryptedEndpoints?: any; - /** - * Specifies whether to require the host name in the server's certificate to match the host name - * of the server when connecting over SSL. The default value is true. - */ - useHostVerification?: any; - /** - * Specifies whether to verify the identity of the server when connecting over SSL. The default - * value is true. + * The Azure key vault secret reference of credString in connection string. */ - usePeerVerification?: any; + credString?: AzureKeyVaultSecretReference; /** * The encrypted credential used for authentication. Credentials are encrypted using the * integration runtime credential manager. Type: string (or Expression with resultType string). @@ -4306,13 +4465,13 @@ export interface MagentoLinkedService { } /** - * Jira Service linked service. + * Concur Service linked service. */ -export interface JiraLinkedService { +export interface ConcurLinkedService { /** * Polymorphic Discriminator */ - type: "Jira"; + type: "Concur"; /** * The integration runtime reference. */ @@ -4330,16 +4489,11 @@ export interface JiraLinkedService { */ annotations?: any[]; /** - * The IP address or host name of the Jira service. (e.g. jira.example.com) - */ - host: any; - /** - * The TCP port that the Jira server uses to listen for client connections. The default value is - * 443 if connecting through HTTPS, or 8080 if connecting through HTTP. + * Application client_id supplied by Concur App Management. */ - port?: any; + clientId: any; /** - * The user name that you use to access Jira Service. + * The user name that you use to access Concur Service. */ username: any; /** @@ -4369,13 +4523,13 @@ export interface JiraLinkedService { } /** - * Impala server linked service. + * Azure PostgreSQL linked service. */ -export interface ImpalaLinkedService { +export interface AzurePostgreSqlLinkedService { /** * Polymorphic Discriminator */ - type: "Impala"; + type: "AzurePostgreSql"; /** * The integration runtime reference. */ @@ -4393,54 +4547,13 @@ export interface ImpalaLinkedService { */ annotations?: any[]; /** - * The IP address or host name of the Impala server. (i.e. 192.168.222.160) - */ - host: any; - /** - * The TCP port that the Impala server uses to listen for client connections. The default value - * is 21050. - */ - port?: any; - /** - * The authentication type to use. Possible values include: 'Anonymous', 'SASLUsername', - * 'UsernameAndPassword' - */ - authenticationType: ImpalaAuthenticationType; - /** - * The user name used to access the Impala server. The default value is anonymous when using - * SASLUsername. - */ - username?: any; - /** - * The password corresponding to the user name when using UsernameAndPassword. - */ - password?: SecretBaseUnion; - /** - * Specifies whether the connections to the server are encrypted using SSL. The default value is - * false. - */ - enableSsl?: any; - /** - * The full path of the .pem file containing trusted CA certificates for verifying the server - * when connecting over SSL. This property can only be set when using SSL on self-hosted IR. The - * default value is the cacerts.pem file installed with the IR. - */ - trustedCertPath?: any; - /** - * Specifies whether to use a CA certificate from the system trust store or from a specified PEM - * file. The default value is false. - */ - useSystemTrustStore?: any; - /** - * Specifies whether to require a CA-issued SSL certificate name to match the host name of the - * server when connecting over SSL. The default value is false. + * An ODBC connection string. Type: string, SecureString or AzureKeyVaultSecretReference. */ - allowHostNameCNMismatch?: any; + connectionString?: any; /** - * Specifies whether to allow self-signed certificates from the server. The default value is - * false. + * The Azure key vault secret reference of password in connection string. */ - allowSelfSignedServerCert?: any; + password?: AzureKeyVaultSecretReference; /** * The encrypted credential used for authentication. Credentials are encrypted using the * integration runtime credential manager. Type: string (or Expression with resultType string). @@ -4449,13 +4562,13 @@ export interface ImpalaLinkedService { } /** - * Hubspot Service linked service. + * Amazon Marketplace Web Service linked service. */ -export interface HubspotLinkedService { +export interface AmazonMWSLinkedService { /** * Polymorphic Discriminator */ - type: "Hubspot"; + type: "AmazonMWS"; /** * The integration runtime reference. */ @@ -4473,21 +4586,30 @@ export interface HubspotLinkedService { */ annotations?: any[]; /** - * The client ID associated with your Hubspot application. + * The endpoint of the Amazon MWS server, (i.e. mws.amazonservices.com) */ - clientId: any; + endpoint: any; /** - * The client secret associated with your Hubspot application. + * The Amazon Marketplace ID you want to retrieve data from. To retrieve data from multiple + * Marketplace IDs, separate them with a comma (,). (i.e. A2EUQ1WTGCTBG2) */ - clientSecret?: SecretBaseUnion; + marketplaceID: any; /** - * The access token obtained when initially authenticating your OAuth integration. + * The Amazon seller ID. */ - accessToken?: SecretBaseUnion; + sellerID: any; /** - * The refresh token obtained when initially authenticating your OAuth integration. + * The Amazon MWS authentication token. */ - refreshToken?: SecretBaseUnion; + mwsAuthToken?: SecretBaseUnion; + /** + * The access key id used to access data. + */ + accessKeyId: any; + /** + * The secret key used to access data. + */ + secretKey?: SecretBaseUnion; /** * Specifies whether the data source endpoints are encrypted using HTTPS. The default value is * true. @@ -4511,13 +4633,13 @@ export interface HubspotLinkedService { } /** - * Hive Server linked service. + * SAP HANA Linked Service. */ -export interface HiveLinkedService { +export interface SapHanaLinkedService { /** * Polymorphic Discriminator */ - type: "Hive"; + type: "SapHana"; /** * The integration runtime reference. */ @@ -4535,80 +4657,79 @@ export interface HiveLinkedService { */ annotations?: any[]; /** - * IP address or host name of the Hive server, separated by ';' for multiple hosts (only when - * serviceDiscoveryMode is enable). + * SAP HANA ODBC connection string. Type: string, SecureString or AzureKeyVaultSecretReference. */ - host: any; + connectionString?: any; /** - * The TCP port that the Hive server uses to listen for client connections. + * Host name of the SAP HANA server. Type: string (or Expression with resultType string). */ - port?: any; + server: any; /** - * The type of Hive server. Possible values include: 'HiveServer1', 'HiveServer2', - * 'HiveThriftServer' + * The authentication type to be used to connect to the SAP HANA server. Possible values include: + * 'Basic', 'Windows' */ - serverType?: HiveServerType; + authenticationType?: SapHanaAuthenticationType; /** - * The transport protocol to use in the Thrift layer. Possible values include: 'Binary', 'SASL', - * 'HTTP ' + * Username to access the SAP HANA server. Type: string (or Expression with resultType string). */ - thriftTransportProtocol?: HiveThriftTransportProtocol; + userName?: any; /** - * The authentication method used to access the Hive server. Possible values include: - * 'Anonymous', 'Username', 'UsernameAndPassword', 'WindowsAzureHDInsightService' + * Password to access the SAP HANA server. */ - authenticationType: HiveAuthenticationType; + password?: SecretBaseUnion; /** - * true to indicate using the ZooKeeper service, false not. + * The encrypted credential used for authentication. Credentials are encrypted using the + * integration runtime credential manager. Type: string (or Expression with resultType string). */ - serviceDiscoveryMode?: any; + encryptedCredential?: any; +} + +/** + * SAP Business Warehouse Linked Service. + */ +export interface SapBWLinkedService { /** - * The namespace on ZooKeeper under which Hive Server 2 nodes are added. + * Polymorphic Discriminator */ - zooKeeperNameSpace?: any; + type: "SapBW"; /** - * Specifies whether the driver uses native HiveQL queries,or converts them into an equivalent - * form in HiveQL. + * The integration runtime reference. */ - useNativeQuery?: any; + connectVia?: IntegrationRuntimeReference; /** - * The user name that you use to access Hive Server. + * Linked service description. */ - username?: any; + description?: string; /** - * The password corresponding to the user name that you provided in the Username field + * Parameters for linked service. */ - password?: SecretBaseUnion; + parameters?: { [propertyName: string]: ParameterSpecification }; /** - * The partial URL corresponding to the Hive server. + * List of tags that can be used for describing the linked service. */ - httpPath?: any; + annotations?: any[]; /** - * Specifies whether the connections to the server are encrypted using SSL. The default value is - * false. + * Host name of the SAP BW instance. Type: string (or Expression with resultType string). */ - enableSsl?: any; + server: any; /** - * The full path of the .pem file containing trusted CA certificates for verifying the server - * when connecting over SSL. This property can only be set when using SSL on self-hosted IR. The - * default value is the cacerts.pem file installed with the IR. + * System number of the BW system. (Usually a two-digit decimal number represented as a string.) + * Type: string (or Expression with resultType string). */ - trustedCertPath?: any; + systemNumber: any; /** - * Specifies whether to use a CA certificate from the system trust store or from a specified PEM - * file. The default value is false. + * Client ID of the client on the BW system. (Usually a three-digit decimal number represented as + * a string) Type: string (or Expression with resultType string). */ - useSystemTrustStore?: any; + clientId: any; /** - * Specifies whether to require a CA-issued SSL certificate name to match the host name of the - * server when connecting over SSL. The default value is false. + * Username to access the SAP BW server. Type: string (or Expression with resultType string). */ - allowHostNameCNMismatch?: any; + userName?: any; /** - * Specifies whether to allow self-signed certificates from the server. The default value is - * false. + * Password to access the SAP BW server. */ - allowSelfSignedServerCert?: any; + password?: SecretBaseUnion; /** * The encrypted credential used for authentication. Credentials are encrypted using the * integration runtime credential manager. Type: string (or Expression with resultType string). @@ -4617,13 +4738,13 @@ export interface HiveLinkedService { } /** - * HBase server linked service. + * A linked service for an SSH File Transfer Protocol (SFTP) server. */ -export interface HBaseLinkedService { +export interface SftpServerLinkedService { /** * Polymorphic Discriminator */ - type: "HBase"; + type: "Sftp"; /** * The integration runtime reference. */ @@ -4641,67 +4762,70 @@ export interface HBaseLinkedService { */ annotations?: any[]; /** - * The IP address or host name of the HBase server. (i.e. 192.168.222.160) + * The SFTP server host name. Type: string (or Expression with resultType string). */ host: any; /** - * The TCP port that the HBase instance uses to listen for client connections. The default value - * is 9090. + * The TCP port number that the SFTP server uses to listen for client connections. Default value + * is 22. Type: integer (or Expression with resultType integer), minimum: 0. */ port?: any; /** - * The partial URL corresponding to the HBase server. (i.e. /gateway/sandbox/hbase/version) + * The authentication type to be used to connect to the FTP server. Possible values include: + * 'Basic', 'SshPublicKey' */ - httpPath?: any; + authenticationType?: SftpAuthenticationType; /** - * The authentication mechanism to use to connect to the HBase server. Possible values include: - * 'Anonymous', 'Basic' + * The username used to log on to the SFTP server. Type: string (or Expression with resultType + * string). */ - authenticationType: HBaseAuthenticationType; + userName?: any; /** - * The user name used to connect to the HBase instance. + * Password to logon the SFTP server for Basic authentication. */ - username?: any; + password?: SecretBaseUnion; /** - * The password corresponding to the user name. + * The encrypted credential used for authentication. Credentials are encrypted using the + * integration runtime credential manager. Type: string (or Expression with resultType string). */ - password?: SecretBaseUnion; + encryptedCredential?: any; /** - * Specifies whether the connections to the server are encrypted using SSL. The default value is - * false. + * The SSH private key file path for SshPublicKey authentication. Only valid for on-premises + * copy. For on-premises copy with SshPublicKey authentication, either PrivateKeyPath or + * PrivateKeyContent should be specified. SSH private key should be OpenSSH format. Type: string + * (or Expression with resultType string). */ - enableSsl?: any; + privateKeyPath?: any; /** - * The full path of the .pem file containing trusted CA certificates for verifying the server - * when connecting over SSL. This property can only be set when using SSL on self-hosted IR. The - * default value is the cacerts.pem file installed with the IR. + * Base64 encoded SSH private key content for SshPublicKey authentication. For on-premises copy + * with SshPublicKey authentication, either PrivateKeyPath or PrivateKeyContent should be + * specified. SSH private key should be OpenSSH format. */ - trustedCertPath?: any; + privateKeyContent?: SecretBaseUnion; /** - * Specifies whether to require a CA-issued SSL certificate name to match the host name of the - * server when connecting over SSL. The default value is false. + * The password to decrypt the SSH private key if the SSH private key is encrypted. */ - allowHostNameCNMismatch?: any; + passPhrase?: SecretBaseUnion; /** - * Specifies whether to allow self-signed certificates from the server. The default value is - * false. + * If true, skip the SSH host key validation. Default value is false. Type: boolean (or + * Expression with resultType boolean). */ - allowSelfSignedServerCert?: any; + skipHostKeyValidation?: any; /** - * The encrypted credential used for authentication. Credentials are encrypted using the - * integration runtime credential manager. Type: string (or Expression with resultType string). + * The host key finger-print of the SFTP server. When SkipHostKeyValidation is false, + * HostKeyFingerprint should be specified. Type: string (or Expression with resultType string). */ - encryptedCredential?: any; + hostKeyFingerprint?: any; } /** - * Greenplum Database linked service. + * A FTP server Linked Service. */ -export interface GreenplumLinkedService { +export interface FtpServerLinkedService { /** * Polymorphic Discriminator */ - type: "Greenplum"; + type: "FtpServer"; /** * The integration runtime reference. */ @@ -4719,28 +4843,52 @@ export interface GreenplumLinkedService { */ annotations?: any[]; /** - * An ODBC connection string. Type: string, SecureString or AzureKeyVaultSecretReference. + * Host name of the FTP server. Type: string (or Expression with resultType string). */ - connectionString?: any; + host: any; /** - * The Azure key vault secret reference of password in connection string. + * The TCP port number that the FTP server uses to listen for client connections. Default value + * is 21. Type: integer (or Expression with resultType integer), minimum: 0. */ - pwd?: AzureKeyVaultSecretReference; + port?: any; + /** + * The authentication type to be used to connect to the FTP server. Possible values include: + * 'Basic', 'Anonymous' + */ + authenticationType?: FtpAuthenticationType; + /** + * Username to logon the FTP server. Type: string (or Expression with resultType string). + */ + userName?: any; + /** + * Password to logon the FTP server. + */ + password?: SecretBaseUnion; /** * The encrypted credential used for authentication. Credentials are encrypted using the * integration runtime credential manager. Type: string (or Expression with resultType string). */ encryptedCredential?: any; + /** + * If true, connect to the FTP server over SSL/TLS channel. Default value is true. Type: boolean + * (or Expression with resultType boolean). + */ + enableSsl?: any; + /** + * If true, validate the FTP server SSL certificate when connect over SSL/TLS channel. Default + * value is true. Type: boolean (or Expression with resultType boolean). + */ + enableServerCertificateValidation?: any; } /** - * Google BigQuery service linked service. + * Linked service for an HTTP source. */ -export interface GoogleBigQueryLinkedService { +export interface HttpLinkedService { /** * Polymorphic Discriminator */ - type: "GoogleBigQuery"; + type: "HttpServer"; /** * The integration runtime reference. */ @@ -4758,75 +4906,58 @@ export interface GoogleBigQueryLinkedService { */ annotations?: any[]; /** - * The default BigQuery project to query against. - */ - project: any; - /** - * A comma-separated list of public BigQuery projects to access. - */ - additionalProjects?: any; - /** - * Whether to request access to Google Drive. Allowing Google Drive access enables support for - * federated tables that combine BigQuery data with data from Google Drive. The default value is - * false. - */ - requestGoogleDriveScope?: any; - /** - * The OAuth 2.0 authentication mechanism used for authentication. ServiceAuthentication can only - * be used on self-hosted IR. Possible values include: 'ServiceAuthentication', - * 'UserAuthentication' - */ - authenticationType: GoogleBigQueryAuthenticationType; - /** - * The refresh token obtained from Google for authorizing access to BigQuery for - * UserAuthentication. - */ - refreshToken?: SecretBaseUnion; - /** - * The client id of the google application used to acquire the refresh token. Type: string (or - * Expression with resultType string). + * The base URL of the HTTP endpoint, e.g. http://www.microsoft.com. Type: string (or Expression + * with resultType string). */ - clientId?: any; + url: any; /** - * The client secret of the google application used to acquire the refresh token. + * The authentication type to be used to connect to the HTTP server. Possible values include: + * 'Basic', 'Anonymous', 'Digest', 'Windows', 'ClientCertificate' */ - clientSecret?: SecretBaseUnion; + authenticationType?: HttpAuthenticationType; /** - * The service account email ID that is used for ServiceAuthentication and can only be used on - * self-hosted IR. + * User name for Basic, Digest, or Windows authentication. Type: string (or Expression with + * resultType string). */ - email?: any; + userName?: any; /** - * The full path to the .p12 key file that is used to authenticate the service account email - * address and can only be used on self-hosted IR. + * Password for Basic, Digest, Windows, or ClientCertificate with EmbeddedCertData + * authentication. */ - keyFilePath?: any; + password?: SecretBaseUnion; /** - * The full path of the .pem file containing trusted CA certificates for verifying the server - * when connecting over SSL. This property can only be set when using SSL on self-hosted IR. The - * default value is the cacerts.pem file installed with the IR. + * Base64 encoded certificate data for ClientCertificate authentication. For on-premises copy + * with ClientCertificate authentication, either CertThumbprint or EmbeddedCertData/Password + * should be specified. Type: string (or Expression with resultType string). */ - trustedCertPath?: any; + embeddedCertData?: any; /** - * Specifies whether to use a CA certificate from the system trust store or from a specified PEM - * file. The default value is false. + * Thumbprint of certificate for ClientCertificate authentication. Only valid for on-premises + * copy. For on-premises copy with ClientCertificate authentication, either CertThumbprint or + * EmbeddedCertData/Password should be specified. Type: string (or Expression with resultType + * string). */ - useSystemTrustStore?: any; + certThumbprint?: any; /** * The encrypted credential used for authentication. Credentials are encrypted using the * integration runtime credential manager. Type: string (or Expression with resultType string). */ encryptedCredential?: any; + /** + * If true, validate the HTTPS server SSL certificate. Default value is true. Type: boolean (or + * Expression with resultType boolean). + */ + enableServerCertificateValidation?: any; } /** - * Eloqua server linked service. + * Linked service for Windows Azure Search Service. */ -export interface EloquaLinkedService { +export interface AzureSearchLinkedService { /** * Polymorphic Discriminator */ - type: "Eloqua"; + type: "AzureSearch"; /** * The integration runtime reference. */ @@ -4844,48 +4975,58 @@ export interface EloquaLinkedService { */ annotations?: any[]; /** - * The endpoint of the Eloqua server. (i.e. eloqua.example.com) + * URL for Azure Search service. Type: string (or Expression with resultType string). */ - endpoint: any; + url: any; /** - * The site name and user name of your Eloqua account in the form: sitename/username. (i.e. - * Eloqua/Alice) + * Admin Key for Azure Search service */ - username: any; + key?: SecretBaseUnion; /** - * The password corresponding to the user name. + * The encrypted credential used for authentication. Credentials are encrypted using the + * integration runtime credential manager. Type: string (or Expression with resultType string). */ - password?: SecretBaseUnion; + encryptedCredential?: any; +} + +/** + * Custom linked service. + */ +export interface CustomDataSourceLinkedService { /** - * Specifies whether the data source endpoints are encrypted using HTTPS. The default value is - * true. + * Polymorphic Discriminator + */ + type: "CustomDataSource"; + /** + * The integration runtime reference. + */ + connectVia?: IntegrationRuntimeReference; + /** + * Linked service description. */ - useEncryptedEndpoints?: any; + description?: string; /** - * Specifies whether to require the host name in the server's certificate to match the host name - * of the server when connecting over SSL. The default value is true. + * Parameters for linked service. */ - useHostVerification?: any; + parameters?: { [propertyName: string]: ParameterSpecification }; /** - * Specifies whether to verify the identity of the server when connecting over SSL. The default - * value is true. + * List of tags that can be used for describing the linked service. */ - usePeerVerification?: any; + annotations?: any[]; /** - * The encrypted credential used for authentication. Credentials are encrypted using the - * integration runtime credential manager. Type: string (or Expression with resultType string). + * Custom linked service properties. */ - encryptedCredential?: any; + typeProperties: any; } /** - * Drill server linked service. + * Linked service for Amazon Redshift. */ -export interface DrillLinkedService { +export interface AmazonRedshiftLinkedService { /** * Polymorphic Discriminator */ - type: "Drill"; + type: "AmazonRedshift"; /** * The integration runtime reference. */ @@ -4903,13 +5044,28 @@ export interface DrillLinkedService { */ annotations?: any[]; /** - * An ODBC connection string. Type: string, SecureString or AzureKeyVaultSecretReference. + * The name of the Amazon Redshift server. Type: string (or Expression with resultType string). */ - connectionString?: any; + server: any; /** - * The Azure key vault secret reference of password in connection string. + * The username of the Amazon Redshift source. Type: string (or Expression with resultType + * string). */ - pwd?: AzureKeyVaultSecretReference; + username?: any; + /** + * The password of the Amazon Redshift source. + */ + password?: SecretBaseUnion; + /** + * The database name of the Amazon Redshift source. Type: string (or Expression with resultType + * string). + */ + database: any; + /** + * The TCP port number that the Amazon Redshift server uses to listen for client connections. The + * default value is 5439. Type: integer (or Expression with resultType integer). + */ + port?: any; /** * The encrypted credential used for authentication. Credentials are encrypted using the * integration runtime credential manager. Type: string (or Expression with resultType string). @@ -4918,13 +5074,13 @@ export interface DrillLinkedService { } /** - * Couchbase server linked service. + * Linked service for Amazon S3. */ -export interface CouchbaseLinkedService { +export interface AmazonS3LinkedService { /** * Polymorphic Discriminator */ - type: "Couchbase"; + type: "AmazonS3"; /** * The integration runtime reference. */ @@ -4942,13 +5098,20 @@ export interface CouchbaseLinkedService { */ annotations?: any[]; /** - * An ODBC connection string. Type: string, SecureString or AzureKeyVaultSecretReference. + * The access key identifier of the Amazon S3 Identity and Access Management (IAM) user. Type: + * string (or Expression with resultType string). */ - connectionString?: any; + accessKeyId?: any; /** - * The Azure key vault secret reference of credString in connection string. + * The secret access key of the Amazon S3 Identity and Access Management (IAM) user. */ - credString?: AzureKeyVaultSecretReference; + secretAccessKey?: SecretBaseUnion; + /** + * This value specifies the endpoint to access with the S3 Connector. This is an optional + * property; change it only if you want to try a different service endpoint or want to switch + * between https and http. Type: string (or Expression with resultType string). + */ + serviceUrl?: any; /** * The encrypted credential used for authentication. Credentials are encrypted using the * integration runtime credential manager. Type: string (or Expression with resultType string). @@ -4957,13 +5120,13 @@ export interface CouchbaseLinkedService { } /** - * Concur Service linked service. + * Rest Service linked service. */ -export interface ConcurLinkedService { +export interface RestServiceLinkedService { /** * Polymorphic Discriminator */ - type: "Concur"; + type: "RestService"; /** * The integration runtime reference. */ @@ -4981,32 +5144,44 @@ export interface ConcurLinkedService { */ annotations?: any[]; /** - * Application client_id supplied by Concur App Management. + * The base URL of the REST service. */ - clientId: any; + url: any; /** - * The user name that you use to access Concur Service. + * Whether to validate server side SSL certificate when connecting to the endpoint.The default + * value is true. Type: boolean (or Expression with resultType boolean). */ - username: any; + enableServerCertificateValidation?: any; /** - * The password corresponding to the user name that you provided in the username field. + * Type of authentication used to connect to the REST service. Possible values include: + * 'Anonymous', 'Basic', 'AadServicePrincipal', 'ManagedServiceIdentity' + */ + authenticationType: RestServiceAuthenticationType; + /** + * The user name used in Basic authentication type. + */ + userName?: any; + /** + * The password used in Basic authentication type. */ password?: SecretBaseUnion; /** - * Specifies whether the data source endpoints are encrypted using HTTPS. The default value is - * true. + * The application's client ID used in AadServicePrincipal authentication type. */ - useEncryptedEndpoints?: any; + servicePrincipalId?: any; /** - * Specifies whether to require the host name in the server's certificate to match the host name - * of the server when connecting over SSL. The default value is true. + * The application's key used in AadServicePrincipal authentication type. */ - useHostVerification?: any; + servicePrincipalKey?: SecretBaseUnion; /** - * Specifies whether to verify the identity of the server when connecting over SSL. The default - * value is true. + * The tenant information (domain name or tenant ID) used in AadServicePrincipal authentication + * type under which your application resides. */ - usePeerVerification?: any; + tenant?: any; + /** + * The resource you are requesting authorization to use. + */ + aadResourceId?: any; /** * The encrypted credential used for authentication. Credentials are encrypted using the * integration runtime credential manager. Type: string (or Expression with resultType string). @@ -5015,13 +5190,13 @@ export interface ConcurLinkedService { } /** - * Azure PostgreSQL linked service. + * SAP Business Warehouse Open Hub Destination Linked Service. */ -export interface AzurePostgreSqlLinkedService { +export interface SapOpenHubLinkedService { /** * Polymorphic Discriminator */ - type: "AzurePostgreSql"; + type: "SapOpenHub"; /** * The integration runtime reference. */ @@ -5039,13 +5214,35 @@ export interface AzurePostgreSqlLinkedService { */ annotations?: any[]; /** - * An ODBC connection string. Type: string, SecureString or AzureKeyVaultSecretReference. + * Host name of the SAP BW instance where the open hub destination is located. Type: string (or + * Expression with resultType string). */ - connectionString?: any; + server: any; /** - * The Azure key vault secret reference of password in connection string. + * System number of the BW system where the open hub destination is located. (Usually a two-digit + * decimal number represented as a string.) Type: string (or Expression with resultType string). */ - password?: AzureKeyVaultSecretReference; + systemNumber: any; + /** + * Client ID of the client on the BW system where the open hub destination is located. (Usually a + * three-digit decimal number represented as a string) Type: string (or Expression with + * resultType string). + */ + clientId: any; + /** + * Language of the BW system where the open hub destination is located. The default value is EN. + * Type: string (or Expression with resultType string). + */ + language?: any; + /** + * Username to access the SAP BW server where the open hub destination is located. Type: string + * (or Expression with resultType string). + */ + userName?: any; + /** + * Password to access the SAP BW server where the open hub destination is located. + */ + password?: SecretBaseUnion; /** * The encrypted credential used for authentication. Credentials are encrypted using the * integration runtime credential manager. Type: string (or Expression with resultType string). @@ -5054,13 +5251,13 @@ export interface AzurePostgreSqlLinkedService { } /** - * Amazon Marketplace Web Service linked service. + * Linked service for SAP ERP Central Component(SAP ECC). */ -export interface AmazonMWSLinkedService { +export interface SapEccLinkedService { /** * Polymorphic Discriminator */ - type: "AmazonMWS"; + type: "SapEcc"; /** * The integration runtime reference. */ @@ -5078,60 +5275,35 @@ export interface AmazonMWSLinkedService { */ annotations?: any[]; /** - * The endpoint of the Amazon MWS server, (i.e. mws.amazonservices.com) - */ - endpoint: any; - /** - * The Amazon Marketplace ID you want to retrieve data from. To retrieve data from multiple - * Marketplace IDs, separate them with a comma (,). (i.e. A2EUQ1WTGCTBG2) - */ - marketplaceID: any; - /** - * The Amazon seller ID. - */ - sellerID: any; - /** - * The Amazon MWS authentication token. - */ - mwsAuthToken?: SecretBaseUnion; - /** - * The access key id used to access data. - */ - accessKeyId: any; - /** - * The secret key used to access data. - */ - secretKey?: SecretBaseUnion; - /** - * Specifies whether the data source endpoints are encrypted using HTTPS. The default value is - * true. + * The URL of SAP ECC OData API. For example, + * '[https://hostname:port/sap/opu/odata/sap/servicename/]'. Type: string (or Expression with + * resultType string). */ - useEncryptedEndpoints?: any; + url: string; /** - * Specifies whether to require the host name in the server's certificate to match the host name - * of the server when connecting over SSL. The default value is true. + * The username for Basic authentication. Type: string (or Expression with resultType string). */ - useHostVerification?: any; + username?: string; /** - * Specifies whether to verify the identity of the server when connecting over SSL. The default - * value is true. + * The password for Basic authentication. */ - usePeerVerification?: any; + password?: SecretBaseUnion; /** * The encrypted credential used for authentication. Credentials are encrypted using the - * integration runtime credential manager. Type: string (or Expression with resultType string). + * integration runtime credential manager. Either encryptedCredential or username/password must + * be provided. Type: string (or Expression with resultType string). */ - encryptedCredential?: any; + encryptedCredential?: string; } /** - * SAP HANA Linked Service. + * Linked service for SAP Cloud for Customer. */ -export interface SapHanaLinkedService { +export interface SapCloudForCustomerLinkedService { /** * Polymorphic Discriminator */ - type: "SapHana"; + type: "SapCloudForCustomer"; /** * The integration runtime reference. */ @@ -5149,41 +5321,35 @@ export interface SapHanaLinkedService { */ annotations?: any[]; /** - * SAP HANA ODBC connection string. Type: string, SecureString or AzureKeyVaultSecretReference. - */ - connectionString?: any; - /** - * Host name of the SAP HANA server. Type: string (or Expression with resultType string). - */ - server: any; - /** - * The authentication type to be used to connect to the SAP HANA server. Possible values include: - * 'Basic', 'Windows' + * The URL of SAP Cloud for Customer OData API. For example, + * '[https://[tenantname].crm.ondemand.com/sap/c4c/odata/v1]'. Type: string (or Expression with + * resultType string). */ - authenticationType?: SapHanaAuthenticationType; + url: any; /** - * Username to access the SAP HANA server. Type: string (or Expression with resultType string). + * The username for Basic authentication. Type: string (or Expression with resultType string). */ - userName?: any; + username?: any; /** - * Password to access the SAP HANA server. + * The password for Basic authentication. */ password?: SecretBaseUnion; /** * The encrypted credential used for authentication. Credentials are encrypted using the - * integration runtime credential manager. Type: string (or Expression with resultType string). + * integration runtime credential manager. Either encryptedCredential or username/password must + * be provided. Type: string (or Expression with resultType string). */ encryptedCredential?: any; } /** - * SAP Business Warehouse Linked Service. + * Linked service for Salesforce Service Cloud. */ -export interface SapBWLinkedService { +export interface SalesforceServiceCloudLinkedService { /** * Polymorphic Discriminator */ - type: "SapBW"; + type: "SalesforceServiceCloud"; /** * The integration runtime reference. */ @@ -5201,27 +5367,34 @@ export interface SapBWLinkedService { */ annotations?: any[]; /** - * Host name of the SAP BW instance. Type: string (or Expression with resultType string). + * The URL of Salesforce Service Cloud instance. Default is 'https://login.salesforce.com'. To + * copy data from sandbox, specify 'https://test.salesforce.com'. To copy data from custom + * domain, specify, for example, 'https://[domain].my.salesforce.com'. Type: string (or + * Expression with resultType string). */ - server: any; + environmentUrl?: any; /** - * System number of the BW system. (Usually a two-digit decimal number represented as a string.) - * Type: string (or Expression with resultType string). + * The username for Basic authentication of the Salesforce instance. Type: string (or Expression + * with resultType string). */ - systemNumber: any; + username?: any; + /** + * The password for Basic authentication of the Salesforce instance. + */ + password?: SecretBaseUnion; /** - * Client ID of the client on the BW system. (Usually a three-digit decimal number represented as - * a string) Type: string (or Expression with resultType string). + * The security token is optional to remotely access Salesforce instance. */ - clientId: any; + securityToken?: SecretBaseUnion; /** - * Username to access the SAP BW server. Type: string (or Expression with resultType string). + * The Salesforce API version used in ADF. Type: string (or Expression with resultType string). */ - userName?: any; + apiVersion?: any; /** - * Password to access the SAP BW server. + * Extended properties appended to the connection string. Type: string (or Expression with + * resultType string). */ - password?: SecretBaseUnion; + extendedProperties?: any; /** * The encrypted credential used for authentication. Credentials are encrypted using the * integration runtime credential manager. Type: string (or Expression with resultType string). @@ -5230,13 +5403,13 @@ export interface SapBWLinkedService { } /** - * A linked service for an SSH File Transfer Protocol (SFTP) server. + * Linked service for Salesforce. */ -export interface SftpServerLinkedService { +export interface SalesforceLinkedService { /** * Polymorphic Discriminator */ - type: "Sftp"; + type: "Salesforce"; /** * The integration runtime reference. */ @@ -5254,70 +5427,44 @@ export interface SftpServerLinkedService { */ annotations?: any[]; /** - * The SFTP server host name. Type: string (or Expression with resultType string). + * The URL of Salesforce instance. Default is 'https://login.salesforce.com'. To copy data from + * sandbox, specify 'https://test.salesforce.com'. To copy data from custom domain, specify, for + * example, 'https://[domain].my.salesforce.com'. Type: string (or Expression with resultType + * string). */ - host: any; + environmentUrl?: any; /** - * The TCP port number that the SFTP server uses to listen for client connections. Default value - * is 22. Type: integer (or Expression with resultType integer), minimum: 0. + * The username for Basic authentication of the Salesforce instance. Type: string (or Expression + * with resultType string). */ - port?: any; + username?: any; /** - * The authentication type to be used to connect to the FTP server. Possible values include: - * 'Basic', 'SshPublicKey' + * The password for Basic authentication of the Salesforce instance. */ - authenticationType?: SftpAuthenticationType; + password?: SecretBaseUnion; /** - * The username used to log on to the SFTP server. Type: string (or Expression with resultType - * string). + * The security token is optional to remotely access Salesforce instance. */ - userName?: any; + securityToken?: SecretBaseUnion; /** - * Password to logon the SFTP server for Basic authentication. + * The Salesforce API version used in ADF. Type: string (or Expression with resultType string). */ - password?: SecretBaseUnion; + apiVersion?: any; /** * The encrypted credential used for authentication. Credentials are encrypted using the * integration runtime credential manager. Type: string (or Expression with resultType string). */ encryptedCredential?: any; - /** - * The SSH private key file path for SshPublicKey authentication. Only valid for on-premises - * copy. For on-premises copy with SshPublicKey authentication, either PrivateKeyPath or - * PrivateKeyContent should be specified. SSH private key should be OpenSSH format. Type: string - * (or Expression with resultType string). - */ - privateKeyPath?: any; - /** - * Base64 encoded SSH private key content for SshPublicKey authentication. For on-premises copy - * with SshPublicKey authentication, either PrivateKeyPath or PrivateKeyContent should be - * specified. SSH private key should be OpenSSH format. - */ - privateKeyContent?: SecretBaseUnion; - /** - * The password to decrypt the SSH private key if the SSH private key is encrypted. - */ - passPhrase?: SecretBaseUnion; - /** - * If true, skip the SSH host key validation. Default value is false. Type: boolean (or - * Expression with resultType boolean). - */ - skipHostKeyValidation?: any; - /** - * The host key finger-print of the SFTP server. When SkipHostKeyValidation is false, - * HostKeyFingerprint should be specified. Type: string (or Expression with resultType string). - */ - hostKeyFingerprint?: any; } /** - * A FTP server Linked Service. + * Office365 linked service. */ -export interface FtpServerLinkedService { +export interface Office365LinkedService { /** * Polymorphic Discriminator */ - type: "FtpServer"; + type: "Office365"; /** * The integration runtime reference. */ @@ -5335,52 +5482,38 @@ export interface FtpServerLinkedService { */ annotations?: any[]; /** - * Host name of the FTP server. Type: string (or Expression with resultType string). - */ - host: any; - /** - * The TCP port number that the FTP server uses to listen for client connections. Default value - * is 21. Type: integer (or Expression with resultType integer), minimum: 0. + * Azure tenant ID to which the Office 365 account belongs. Type: string (or Expression with + * resultType string). */ - port?: any; + office365TenantId: any; /** - * The authentication type to be used to connect to the FTP server. Possible values include: - * 'Basic', 'Anonymous' + * Specify the tenant information under which your Azure AD web application resides. Type: string + * (or Expression with resultType string). */ - authenticationType?: FtpAuthenticationType; + servicePrincipalTenantId: any; /** - * Username to logon the FTP server. Type: string (or Expression with resultType string). + * Specify the application's client ID. Type: string (or Expression with resultType string). */ - userName?: any; + servicePrincipalId: any; /** - * Password to logon the FTP server. + * Specify the application's key. */ - password?: SecretBaseUnion; + servicePrincipalKey: SecretBaseUnion; /** * The encrypted credential used for authentication. Credentials are encrypted using the * integration runtime credential manager. Type: string (or Expression with resultType string). */ encryptedCredential?: any; - /** - * If true, connect to the FTP server over SSL/TLS channel. Default value is true. Type: boolean - * (or Expression with resultType boolean). - */ - enableSsl?: any; - /** - * If true, validate the FTP server SSL certificate when connect over SSL/TLS channel. Default - * value is true. Type: boolean (or Expression with resultType boolean). - */ - enableServerCertificateValidation?: any; } /** - * Linked service for an HTTP source. + * Azure Data Lake Storage Gen2 linked service. */ -export interface HttpLinkedService { +export interface AzureBlobFSLinkedService { /** * Polymorphic Discriminator */ - type: "HttpServer"; + type: "AzureBlobFS"; /** * The integration runtime reference. */ @@ -5398,58 +5531,45 @@ export interface HttpLinkedService { */ annotations?: any[]; /** - * The base URL of the HTTP endpoint, e.g. http://www.microsoft.com. Type: string (or Expression - * with resultType string). + * Endpoint for the Azure Data Lake Storage Gen2 service. Type: string (or Expression with + * resultType string). */ url: any; /** - * The authentication type to be used to connect to the HTTP server. Possible values include: - * 'Basic', 'Anonymous', 'Digest', 'Windows', 'ClientCertificate' - */ - authenticationType?: HttpAuthenticationType; - /** - * User name for Basic, Digest, or Windows authentication. Type: string (or Expression with + * Account key for the Azure Data Lake Storage Gen2 service. Type: string (or Expression with * resultType string). */ - userName?: any; + accountKey?: any; /** - * Password for Basic, Digest, Windows, or ClientCertificate with EmbeddedCertData - * authentication. + * The ID of the application used to authenticate against the Azure Data Lake Storage Gen2 + * account. Type: string (or Expression with resultType string). */ - password?: SecretBaseUnion; + servicePrincipalId?: any; /** - * Base64 encoded certificate data for ClientCertificate authentication. For on-premises copy - * with ClientCertificate authentication, either CertThumbprint or EmbeddedCertData/Password - * should be specified. Type: string (or Expression with resultType string). + * The Key of the application used to authenticate against the Azure Data Lake Storage Gen2 + * account. */ - embeddedCertData?: any; + servicePrincipalKey?: SecretBaseUnion; /** - * Thumbprint of certificate for ClientCertificate authentication. Only valid for on-premises - * copy. For on-premises copy with ClientCertificate authentication, either CertThumbprint or - * EmbeddedCertData/Password should be specified. Type: string (or Expression with resultType - * string). + * The name or ID of the tenant to which the service principal belongs. Type: string (or + * Expression with resultType string). */ - certThumbprint?: any; + tenant?: any; /** * The encrypted credential used for authentication. Credentials are encrypted using the * integration runtime credential manager. Type: string (or Expression with resultType string). */ encryptedCredential?: any; - /** - * If true, validate the HTTPS server SSL certificate. Default value is true. Type: boolean (or - * Expression with resultType boolean). - */ - enableServerCertificateValidation?: any; } /** - * Linked service for Windows Azure Search Service. + * Azure Data Lake Store linked service. */ -export interface AzureSearchLinkedService { +export interface AzureDataLakeStoreLinkedService { /** * Polymorphic Discriminator */ - type: "AzureSearch"; + type: "AzureDataLakeStore"; /** * The integration runtime reference. */ @@ -5467,58 +5587,52 @@ export interface AzureSearchLinkedService { */ annotations?: any[]; /** - * URL for Azure Search service. Type: string (or Expression with resultType string). - */ - url: any; - /** - * Admin Key for Azure Search service + * Data Lake Store service URI. Type: string (or Expression with resultType string). */ - key?: SecretBaseUnion; + dataLakeStoreUri: any; /** - * The encrypted credential used for authentication. Credentials are encrypted using the - * integration runtime credential manager. Type: string (or Expression with resultType string). + * The ID of the application used to authenticate against the Azure Data Lake Store account. + * Type: string (or Expression with resultType string). */ - encryptedCredential?: any; -} - -/** - * Custom linked service. - */ -export interface CustomDataSourceLinkedService { + servicePrincipalId?: any; /** - * Polymorphic Discriminator + * The Key of the application used to authenticate against the Azure Data Lake Store account. */ - type: "CustomDataSource"; + servicePrincipalKey?: SecretBaseUnion; /** - * The integration runtime reference. + * The name or ID of the tenant to which the service principal belongs. Type: string (or + * Expression with resultType string). */ - connectVia?: IntegrationRuntimeReference; + tenant?: any; /** - * Linked service description. + * Data Lake Store account name. Type: string (or Expression with resultType string). */ - description?: string; + accountName?: any; /** - * Parameters for linked service. + * Data Lake Store account subscription ID (if different from Data Factory account). Type: string + * (or Expression with resultType string). */ - parameters?: { [propertyName: string]: ParameterSpecification }; + subscriptionId?: any; /** - * List of tags that can be used for describing the linked service. + * Data Lake Store account resource group name (if different from Data Factory account). Type: + * string (or Expression with resultType string). */ - annotations?: any[]; + resourceGroupName?: any; /** - * Custom linked service properties. + * The encrypted credential used for authentication. Credentials are encrypted using the + * integration runtime credential manager. Type: string (or Expression with resultType string). */ - typeProperties: any; + encryptedCredential?: any; } /** - * Linked service for Amazon Redshift. + * Linked service for CosmosDB (MongoDB API) data source. */ -export interface AmazonRedshiftLinkedService { +export interface CosmosDbMongoDbApiLinkedService { /** * Polymorphic Discriminator */ - type: "AmazonRedshift"; + type: "CosmosDbMongoDbApi"; /** * The integration runtime reference. */ @@ -5536,43 +5650,25 @@ export interface AmazonRedshiftLinkedService { */ annotations?: any[]; /** - * The name of the Amazon Redshift server. Type: string (or Expression with resultType string). - */ - server: any; - /** - * The username of the Amazon Redshift source. Type: string (or Expression with resultType - * string). - */ - username?: any; - /** - * The password of the Amazon Redshift source. + * The CosmosDB (MongoDB API) connection string. Type: string, SecureString or + * AzureKeyVaultSecretReference. Type: string, SecureString or AzureKeyVaultSecretReference. */ - password?: SecretBaseUnion; + connectionString: any; /** - * The database name of the Amazon Redshift source. Type: string (or Expression with resultType - * string). + * The name of the CosmosDB (MongoDB API) database that you want to access. Type: string (or + * Expression with resultType string). */ database: any; - /** - * The TCP port number that the Amazon Redshift server uses to listen for client connections. The - * default value is 5439. Type: integer (or Expression with resultType integer). - */ - port?: any; - /** - * The encrypted credential used for authentication. Credentials are encrypted using the - * integration runtime credential manager. Type: string (or Expression with resultType string). - */ - encryptedCredential?: any; } /** - * Linked service for Amazon S3. + * Linked service for MongoDB data source. */ -export interface AmazonS3LinkedService { +export interface MongoDbV2LinkedService { /** * Polymorphic Discriminator */ - type: "AmazonS3"; + type: "MongoDbV2"; /** * The integration runtime reference. */ @@ -5588,37 +5684,27 @@ export interface AmazonS3LinkedService { /** * List of tags that can be used for describing the linked service. */ - annotations?: any[]; - /** - * The access key identifier of the Amazon S3 Identity and Access Management (IAM) user. Type: - * string (or Expression with resultType string). - */ - accessKeyId?: any; - /** - * The secret access key of the Amazon S3 Identity and Access Management (IAM) user. - */ - secretAccessKey?: SecretBaseUnion; + annotations?: any[]; /** - * This value specifies the endpoint to access with the S3 Connector. This is an optional - * property; change it only if you want to try a different service endpoint or want to switch - * between https and http. Type: string (or Expression with resultType string). + * The MongoDB connection string. Type: string, SecureString or AzureKeyVaultSecretReference. + * Type: string, SecureString or AzureKeyVaultSecretReference. */ - serviceUrl?: any; + connectionString: any; /** - * The encrypted credential used for authentication. Credentials are encrypted using the - * integration runtime credential manager. Type: string (or Expression with resultType string). + * The name of the MongoDB database that you want to access. Type: string (or Expression with + * resultType string). */ - encryptedCredential?: any; + database: any; } /** - * Rest Service linked service. + * Linked service for MongoDb data source. */ -export interface RestServiceLinkedService { +export interface MongoDbLinkedService { /** * Polymorphic Discriminator */ - type: "RestService"; + type: "MongoDb"; /** * The integration runtime reference. */ @@ -5636,44 +5722,48 @@ export interface RestServiceLinkedService { */ annotations?: any[]; /** - * The base URL of the REST service. + * The IP address or server name of the MongoDB server. Type: string (or Expression with + * resultType string). */ - url: any; + server: any; /** - * Whether to validate server side SSL certificate when connecting to the endpoint.The default - * value is true. Type: boolean (or Expression with resultType boolean). + * The authentication type to be used to connect to the MongoDB database. Possible values + * include: 'Basic', 'Anonymous' */ - enableServerCertificateValidation?: any; + authenticationType?: MongoDbAuthenticationType; /** - * Type of authentication used to connect to the REST service. Possible values include: - * 'Anonymous', 'Basic', 'AadServicePrincipal', 'ManagedServiceIdentity' + * The name of the MongoDB database that you want to access. Type: string (or Expression with + * resultType string). */ - authenticationType: RestServiceAuthenticationType; + databaseName: any; /** - * The user name used in Basic authentication type. + * Username for authentication. Type: string (or Expression with resultType string). */ - userName?: any; + username?: any; /** - * The password used in Basic authentication type. + * Password for authentication. */ password?: SecretBaseUnion; /** - * The application's client ID used in AadServicePrincipal authentication type. + * Database to verify the username and password. Type: string (or Expression with resultType + * string). */ - servicePrincipalId?: any; + authSource?: any; /** - * The application's key used in AadServicePrincipal authentication type. + * The TCP port number that the MongoDB server uses to listen for client connections. The default + * value is 27017. Type: integer (or Expression with resultType integer), minimum: 0. */ - servicePrincipalKey?: SecretBaseUnion; + port?: any; /** - * The tenant information (domain name or tenant ID) used in AadServicePrincipal authentication - * type under which your application resides. + * Specifies whether the connections to the server are encrypted using SSL. The default value is + * false. Type: boolean (or Expression with resultType boolean). */ - tenant?: any; + enableSsl?: any; /** - * The resource you are requesting authorization to use. + * Specifies whether to allow self-signed certificates from the server. The default value is + * false. Type: boolean (or Expression with resultType boolean). */ - aadResourceId?: any; + allowSelfSignedServerCert?: any; /** * The encrypted credential used for authentication. Credentials are encrypted using the * integration runtime credential manager. Type: string (or Expression with resultType string). @@ -5682,13 +5772,13 @@ export interface RestServiceLinkedService { } /** - * SAP Business Warehouse Open Hub Destination Linked Service. + * Linked service for Cassandra data source. */ -export interface SapOpenHubLinkedService { +export interface CassandraLinkedService { /** * Polymorphic Discriminator */ - type: "SapOpenHub"; + type: "Cassandra"; /** * The integration runtime reference. */ @@ -5706,33 +5796,24 @@ export interface SapOpenHubLinkedService { */ annotations?: any[]; /** - * Host name of the SAP BW instance where the open hub destination is located. Type: string (or - * Expression with resultType string). - */ - server: any; - /** - * System number of the BW system where the open hub destination is located. (Usually a two-digit - * decimal number represented as a string.) Type: string (or Expression with resultType string). + * Host name for connection. Type: string (or Expression with resultType string). */ - systemNumber: any; + host: any; /** - * Client ID of the client on the BW system where the open hub destination is located. (Usually a - * three-digit decimal number represented as a string) Type: string (or Expression with - * resultType string). + * AuthenticationType to be used for connection. Type: string (or Expression with resultType + * string). */ - clientId: any; + authenticationType?: any; /** - * Language of the BW system where the open hub destination is located. The default value is EN. - * Type: string (or Expression with resultType string). + * The port for the connection. Type: integer (or Expression with resultType integer). */ - language?: any; + port?: any; /** - * Username to access the SAP BW server where the open hub destination is located. Type: string - * (or Expression with resultType string). + * Username for authentication. Type: string (or Expression with resultType string). */ - userName?: any; + username?: any; /** - * Password to access the SAP BW server where the open hub destination is located. + * Password for authentication. */ password?: SecretBaseUnion; /** @@ -5743,59 +5824,97 @@ export interface SapOpenHubLinkedService { } /** - * Linked service for SAP ERP Central Component(SAP ECC). + * Contains the possible cases for WebLinkedServiceTypeProperties. */ -export interface SapEccLinkedService { +export type WebLinkedServiceTypePropertiesUnion = WebLinkedServiceTypeProperties | WebClientCertificateAuthentication | WebBasicAuthentication | WebAnonymousAuthentication; + +/** + * Base definition of WebLinkedServiceTypeProperties, this typeProperties is polymorphic based on + * authenticationType, so not flattened in SDK models. + */ +export interface WebLinkedServiceTypeProperties { /** * Polymorphic Discriminator */ - type: "SapEcc"; + authenticationType: "WebLinkedServiceTypeProperties"; /** - * The integration runtime reference. + * The URL of the web service endpoint, e.g. http://www.microsoft.com . Type: string (or + * Expression with resultType string). */ - connectVia?: IntegrationRuntimeReference; + url: any; +} + +/** + * A WebLinkedService that uses client certificate based authentication to communicate with an HTTP + * endpoint. This scheme follows mutual authentication; the server must also provide valid + * credentials to the client. + */ +export interface WebClientCertificateAuthentication { /** - * Linked service description. + * Polymorphic Discriminator */ - description?: string; + authenticationType: "ClientCertificate"; /** - * Parameters for linked service. + * The URL of the web service endpoint, e.g. http://www.microsoft.com . Type: string (or + * Expression with resultType string). */ - parameters?: { [propertyName: string]: ParameterSpecification }; + url: any; /** - * List of tags that can be used for describing the linked service. + * Base64-encoded contents of a PFX file. */ - annotations?: any[]; + pfx: SecretBaseUnion; /** - * The URL of SAP ECC OData API. For example, - * '[https://hostname:port/sap/opu/odata/sap/servicename/]'. Type: string (or Expression with - * resultType string). + * Password for the PFX file. */ - url: string; + password: SecretBaseUnion; +} + +/** + * A WebLinkedService that uses basic authentication to communicate with an HTTP endpoint. + */ +export interface WebBasicAuthentication { /** - * The username for Basic authentication. Type: string (or Expression with resultType string). + * Polymorphic Discriminator */ - username?: string; + authenticationType: "Basic"; + /** + * The URL of the web service endpoint, e.g. http://www.microsoft.com . Type: string (or + * Expression with resultType string). + */ + url: any; + /** + * User name for Basic authentication. Type: string (or Expression with resultType string). + */ + username: any; /** * The password for Basic authentication. */ - password?: SecretBaseUnion; + password: SecretBaseUnion; +} + +/** + * A WebLinkedService that uses anonymous authentication to communicate with an HTTP endpoint. + */ +export interface WebAnonymousAuthentication { /** - * The encrypted credential used for authentication. Credentials are encrypted using the - * integration runtime credential manager. Either encryptedCredential or username/password must - * be provided. Type: string (or Expression with resultType string). + * Polymorphic Discriminator */ - encryptedCredential?: string; + authenticationType: "Anonymous"; + /** + * The URL of the web service endpoint, e.g. http://www.microsoft.com . Type: string (or + * Expression with resultType string). + */ + url: any; } /** - * Linked service for SAP Cloud for Customer. + * Web linked service. */ -export interface SapCloudForCustomerLinkedService { +export interface WebLinkedService { /** * Polymorphic Discriminator */ - type: "SapCloudForCustomer"; + type: "Web"; /** * The integration runtime reference. */ @@ -5813,35 +5932,19 @@ export interface SapCloudForCustomerLinkedService { */ annotations?: any[]; /** - * The URL of SAP Cloud for Customer OData API. For example, - * '[https://[tenantname].crm.ondemand.com/sap/c4c/odata/v1]'. Type: string (or Expression with - * resultType string). - */ - url: any; - /** - * The username for Basic authentication. Type: string (or Expression with resultType string). - */ - username?: any; - /** - * The password for Basic authentication. - */ - password?: SecretBaseUnion; - /** - * The encrypted credential used for authentication. Credentials are encrypted using the - * integration runtime credential manager. Either encryptedCredential or username/password must - * be provided. Type: string (or Expression with resultType string). + * Web linked service properties. */ - encryptedCredential?: any; + typeProperties: WebLinkedServiceTypePropertiesUnion; } /** - * Linked service for Salesforce Service Cloud. + * Open Data Protocol (OData) linked service. */ -export interface SalesforceServiceCloudLinkedService { +export interface ODataLinkedService { /** * Polymorphic Discriminator */ - type: "SalesforceServiceCloud"; + type: "OData"; /** * The integration runtime reference. */ @@ -5859,30 +5962,57 @@ export interface SalesforceServiceCloudLinkedService { */ annotations?: any[]; /** - * The URL of Salesforce Service Cloud instance. Default is 'https://login.salesforce.com'. To - * copy data from sandbox, specify 'https://test.salesforce.com'. To copy data from custom - * domain, specify, for example, 'https://[domain].my.salesforce.com'. Type: string (or + * The URL of the OData service endpoint. Type: string (or Expression with resultType string). + */ + url: any; + /** + * Type of authentication used to connect to the OData service. Possible values include: 'Basic', + * 'Anonymous', 'Windows', 'AadServicePrincipal', 'ManagedServiceIdentity' + */ + authenticationType?: ODataAuthenticationType; + /** + * User name of the OData service. Type: string (or Expression with resultType string). + */ + userName?: any; + /** + * Password of the OData service. + */ + password?: SecretBaseUnion; + /** + * Specify the tenant information (domain name or tenant ID) under which your application + * resides. Type: string (or Expression with resultType string). + */ + tenant?: any; + /** + * Specify the application id of your application registered in Azure Active Directory. Type: + * string (or Expression with resultType string). + */ + servicePrincipalId?: any; + /** + * Specify the resource you are requesting authorization to use Directory. Type: string (or * Expression with resultType string). */ - environmentUrl?: any; + aadResourceId?: any; /** - * The username for Basic authentication of the Salesforce instance. Type: string (or Expression - * with resultType string). + * Specify the credential type (key or cert) is used for service principal. Possible values + * include: 'ServicePrincipalKey', 'ServicePrincipalCert' */ - username?: any; + aadServicePrincipalCredentialType?: ODataAadServicePrincipalCredentialType; /** - * The password for Basic authentication of the Salesforce instance. + * Specify the secret of your application registered in Azure Active Directory. Type: string (or + * Expression with resultType string). */ - password?: SecretBaseUnion; + servicePrincipalKey?: SecretBaseUnion; /** - * The security token is required to remotely access Salesforce instance. + * Specify the base64 encoded certificate of your application registered in Azure Active + * Directory. Type: string (or Expression with resultType string). */ - securityToken?: SecretBaseUnion; + servicePrincipalEmbeddedCert?: SecretBaseUnion; /** - * Extended properties appended to the connection string. Type: string (or Expression with - * resultType string). + * Specify the password of your certificate if your certificate has a password and you are using + * AadServicePrincipal authentication. Type: string (or Expression with resultType string). */ - extendedProperties?: any; + servicePrincipalEmbeddedCertPassword?: SecretBaseUnion; /** * The encrypted credential used for authentication. Credentials are encrypted using the * integration runtime credential manager. Type: string (or Expression with resultType string). @@ -5891,13 +6021,13 @@ export interface SalesforceServiceCloudLinkedService { } /** - * Linked service for Salesforce. + * Hadoop Distributed File System (HDFS) linked service. */ -export interface SalesforceLinkedService { +export interface HdfsLinkedService { /** * Polymorphic Discriminator */ - type: "Salesforce"; + type: "Hdfs"; /** * The integration runtime reference. */ @@ -5915,40 +6045,38 @@ export interface SalesforceLinkedService { */ annotations?: any[]; /** - * The URL of Salesforce instance. Default is 'https://login.salesforce.com'. To copy data from - * sandbox, specify 'https://test.salesforce.com'. To copy data from custom domain, specify, for - * example, 'https://[domain].my.salesforce.com'. Type: string (or Expression with resultType - * string). + * The URL of the HDFS service endpoint, e.g. http://myhostname:50070/webhdfs/v1 . Type: string + * (or Expression with resultType string). */ - environmentUrl?: any; + url: any; /** - * The username for Basic authentication of the Salesforce instance. Type: string (or Expression - * with resultType string). + * Type of authentication used to connect to the HDFS. Possible values are: Anonymous and + * Windows. Type: string (or Expression with resultType string). */ - username?: any; + authenticationType?: any; /** - * The password for Basic authentication of the Salesforce instance. + * The encrypted credential used for authentication. Credentials are encrypted using the + * integration runtime credential manager. Type: string (or Expression with resultType string). */ - password?: SecretBaseUnion; + encryptedCredential?: any; /** - * The security token is required to remotely access Salesforce instance. + * User name for Windows authentication. Type: string (or Expression with resultType string). */ - securityToken?: SecretBaseUnion; + userName?: any; /** - * The encrypted credential used for authentication. Credentials are encrypted using the - * integration runtime credential manager. Type: string (or Expression with resultType string). + * Password for Windows authentication. */ - encryptedCredential?: any; + password?: SecretBaseUnion; } /** - * Office365 linked service. + * Microsoft Access linked service. */ -export interface Office365LinkedService { +export interface MicrosoftAccessLinkedService { /** * Polymorphic Discriminator */ - type: "Office365"; + type: "MicrosoftAccess"; /** * The integration runtime reference. */ @@ -5966,23 +6094,28 @@ export interface Office365LinkedService { */ annotations?: any[]; /** - * Azure tenant ID to which the Office 365 account belongs. Type: string (or Expression with - * resultType string). + * The non-access credential portion of the connection string as well as an optional encrypted + * credential. Type: string, SecureString or AzureKeyVaultSecretReference. */ - office365TenantId: any; + connectionString: any; /** - * Specify the tenant information under which your Azure AD web application resides. Type: string - * (or Expression with resultType string). + * Type of authentication used to connect to the Microsoft Access as ODBC data store. Possible + * values are: Anonymous and Basic. Type: string (or Expression with resultType string). */ - servicePrincipalTenantId: any; + authenticationType?: any; /** - * Specify the application's client ID. Type: string (or Expression with resultType string). + * The access credential portion of the connection string specified in driver-specific + * property-value format. */ - servicePrincipalId: any; + credential?: SecretBaseUnion; /** - * Specify the application's key. + * User name for Basic authentication. Type: string (or Expression with resultType string). */ - servicePrincipalKey: SecretBaseUnion; + userName?: any; + /** + * Password for Basic authentication. + */ + password?: SecretBaseUnion; /** * The encrypted credential used for authentication. Credentials are encrypted using the * integration runtime credential manager. Type: string (or Expression with resultType string). @@ -5991,13 +6124,13 @@ export interface Office365LinkedService { } /** - * Azure Data Lake Storage Gen2 linked service. + * Informix linked service. */ -export interface AzureBlobFSLinkedService { +export interface InformixLinkedService { /** * Polymorphic Discriminator */ - type: "AzureBlobFS"; + type: "Informix"; /** * The integration runtime reference. */ @@ -6015,30 +6148,28 @@ export interface AzureBlobFSLinkedService { */ annotations?: any[]; /** - * Endpoint for the Azure Data Lake Storage Gen2 service. Type: string (or Expression with - * resultType string). + * The non-access credential portion of the connection string as well as an optional encrypted + * credential. Type: string, SecureString or AzureKeyVaultSecretReference. */ - url: any; + connectionString: any; /** - * Account key for the Azure Data Lake Storage Gen2 service. Type: string (or Expression with - * resultType string). + * Type of authentication used to connect to the Informix as ODBC data store. Possible values + * are: Anonymous and Basic. Type: string (or Expression with resultType string). */ - accountKey?: any; + authenticationType?: any; /** - * The ID of the application used to authenticate against the Azure Data Lake Storage Gen2 - * account. Type: string (or Expression with resultType string). + * The access credential portion of the connection string specified in driver-specific + * property-value format. */ - servicePrincipalId?: any; + credential?: SecretBaseUnion; /** - * The Key of the application used to authenticate against the Azure Data Lake Storage Gen2 - * account. + * User name for Basic authentication. Type: string (or Expression with resultType string). */ - servicePrincipalKey?: SecretBaseUnion; + userName?: any; /** - * The name or ID of the tenant to which the service principal belongs. Type: string (or - * Expression with resultType string). + * Password for Basic authentication. */ - tenant?: any; + password?: SecretBaseUnion; /** * The encrypted credential used for authentication. Credentials are encrypted using the * integration runtime credential manager. Type: string (or Expression with resultType string). @@ -6047,13 +6178,13 @@ export interface AzureBlobFSLinkedService { } /** - * Azure Data Lake Store linked service. + * Open Database Connectivity (ODBC) linked service. */ -export interface AzureDataLakeStoreLinkedService { +export interface OdbcLinkedService { /** * Polymorphic Discriminator */ - type: "AzureDataLakeStore"; + type: "Odbc"; /** * The integration runtime reference. */ @@ -6071,37 +6202,28 @@ export interface AzureDataLakeStoreLinkedService { */ annotations?: any[]; /** - * Data Lake Store service URI. Type: string (or Expression with resultType string). - */ - dataLakeStoreUri: any; - /** - * The ID of the application used to authenticate against the Azure Data Lake Store account. - * Type: string (or Expression with resultType string). - */ - servicePrincipalId?: any; - /** - * The Key of the application used to authenticate against the Azure Data Lake Store account. + * The non-access credential portion of the connection string as well as an optional encrypted + * credential. Type: string, SecureString or AzureKeyVaultSecretReference. */ - servicePrincipalKey?: SecretBaseUnion; + connectionString: any; /** - * The name or ID of the tenant to which the service principal belongs. Type: string (or - * Expression with resultType string). + * Type of authentication used to connect to the ODBC data store. Possible values are: Anonymous + * and Basic. Type: string (or Expression with resultType string). */ - tenant?: any; + authenticationType?: any; /** - * Data Lake Store account name. Type: string (or Expression with resultType string). + * The access credential portion of the connection string specified in driver-specific + * property-value format. */ - accountName?: any; + credential?: SecretBaseUnion; /** - * Data Lake Store account subscription ID (if different from Data Factory account). Type: string - * (or Expression with resultType string). + * User name for Basic authentication. Type: string (or Expression with resultType string). */ - subscriptionId?: any; + userName?: any; /** - * Data Lake Store account resource group name (if different from Data Factory account). Type: - * string (or Expression with resultType string). + * Password for Basic authentication. */ - resourceGroupName?: any; + password?: SecretBaseUnion; /** * The encrypted credential used for authentication. Credentials are encrypted using the * integration runtime credential manager. Type: string (or Expression with resultType string). @@ -6110,13 +6232,13 @@ export interface AzureDataLakeStoreLinkedService { } /** - * Linked service for CosmosDB (MongoDB API) data source. + * Azure ML Service linked service. */ -export interface CosmosDbMongoDbApiLinkedService { +export interface AzureMLServiceLinkedService { /** * Polymorphic Discriminator */ - type: "CosmosDbMongoDbApi"; + type: "AzureMLService"; /** * The integration runtime reference. */ @@ -6134,61 +6256,49 @@ export interface CosmosDbMongoDbApiLinkedService { */ annotations?: any[]; /** - * The CosmosDB (MongoDB API) connection string. Type: string, SecureString or - * AzureKeyVaultSecretReference. Type: string, SecureString or AzureKeyVaultSecretReference. - */ - connectionString: any; - /** - * The name of the CosmosDB (MongoDB API) database that you want to access. Type: string (or - * Expression with resultType string). - */ - database: any; -} - -/** - * Linked service for MongoDB data source. - */ -export interface MongoDbV2LinkedService { - /** - * Polymorphic Discriminator + * Azure ML Service workspace subscription ID. Type: string (or Expression with resultType + * string). */ - type: "MongoDbV2"; + subscriptionId: any; /** - * The integration runtime reference. + * Azure ML Service workspace resource group name. Type: string (or Expression with resultType + * string). */ - connectVia?: IntegrationRuntimeReference; + resourceGroupName: any; /** - * Linked service description. + * Azure ML Service workspace name. Type: string (or Expression with resultType string). */ - description?: string; + mlWorkspaceName: any; /** - * Parameters for linked service. + * The ID of the service principal used to authenticate against the endpoint of a published Azure + * ML Service pipeline. Type: string (or Expression with resultType string). */ - parameters?: { [propertyName: string]: ParameterSpecification }; + servicePrincipalId?: any; /** - * List of tags that can be used for describing the linked service. + * The key of the service principal used to authenticate against the endpoint of a published + * Azure ML Service pipeline. */ - annotations?: any[]; + servicePrincipalKey?: SecretBaseUnion; /** - * The MongoDB connection string. Type: string, SecureString or AzureKeyVaultSecretReference. - * Type: string, SecureString or AzureKeyVaultSecretReference. + * The name or ID of the tenant to which the service principal belongs. Type: string (or + * Expression with resultType string). */ - connectionString: any; + tenant?: any; /** - * The name of the MongoDB database that you want to access. Type: string (or Expression with - * resultType string). + * The encrypted credential used for authentication. Credentials are encrypted using the + * integration runtime credential manager. Type: string (or Expression with resultType string). */ - database: any; + encryptedCredential?: any; } /** - * Linked service for MongoDb data source. + * Azure ML Studio Web Service linked service. */ -export interface MongoDbLinkedService { +export interface AzureMLLinkedService { /** * Polymorphic Discriminator */ - type: "MongoDb"; + type: "AzureML"; /** * The integration runtime reference. */ @@ -6206,48 +6316,35 @@ export interface MongoDbLinkedService { */ annotations?: any[]; /** - * The IP address or server name of the MongoDB server. Type: string (or Expression with - * resultType string). - */ - server: any; - /** - * The authentication type to be used to connect to the MongoDB database. Possible values - * include: 'Basic', 'Anonymous' - */ - authenticationType?: MongoDbAuthenticationType; - /** - * The name of the MongoDB database that you want to access. Type: string (or Expression with - * resultType string). - */ - databaseName: any; - /** - * Username for authentication. Type: string (or Expression with resultType string). + * The Batch Execution REST URL for an Azure ML Studio Web Service endpoint. Type: string (or + * Expression with resultType string). */ - username?: any; + mlEndpoint: any; /** - * Password for authentication. + * The API key for accessing the Azure ML model endpoint. */ - password?: SecretBaseUnion; + apiKey: SecretBaseUnion; /** - * Database to verify the username and password. Type: string (or Expression with resultType - * string). + * The Update Resource REST URL for an Azure ML Studio Web Service endpoint. Type: string (or + * Expression with resultType string). */ - authSource?: any; + updateResourceEndpoint?: any; /** - * The TCP port number that the MongoDB server uses to listen for client connections. The default - * value is 27017. Type: integer (or Expression with resultType integer), minimum: 0. + * The ID of the service principal used to authenticate against the ARM-based + * updateResourceEndpoint of an Azure ML Studio web service. Type: string (or Expression with + * resultType string). */ - port?: any; + servicePrincipalId?: any; /** - * Specifies whether the connections to the server are encrypted using SSL. The default value is - * false. Type: boolean (or Expression with resultType boolean). + * The key of the service principal used to authenticate against the ARM-based + * updateResourceEndpoint of an Azure ML Studio web service. */ - enableSsl?: any; + servicePrincipalKey?: SecretBaseUnion; /** - * Specifies whether to allow self-signed certificates from the server. The default value is - * false. Type: boolean (or Expression with resultType boolean). + * The name or ID of the tenant to which the service principal belongs. Type: string (or + * Expression with resultType string). */ - allowSelfSignedServerCert?: any; + tenant?: any; /** * The encrypted credential used for authentication. Credentials are encrypted using the * integration runtime credential manager. Type: string (or Expression with resultType string). @@ -6256,13 +6353,13 @@ export interface MongoDbLinkedService { } /** - * Linked service for Cassandra data source. + * Linked service for Teradata data source. */ -export interface CassandraLinkedService { +export interface TeradataLinkedService { /** * Polymorphic Discriminator */ - type: "Cassandra"; + type: "Teradata"; /** * The integration runtime reference. */ @@ -6280,18 +6377,17 @@ export interface CassandraLinkedService { */ annotations?: any[]; /** - * Host name for connection. Type: string (or Expression with resultType string). + * Teradata ODBC connection string. Type: string, SecureString or AzureKeyVaultSecretReference. */ - host: any; + connectionString?: any; /** - * AuthenticationType to be used for connection. Type: string (or Expression with resultType - * string). + * Server name for connection. Type: string (or Expression with resultType string). */ - authenticationType?: any; + server?: any; /** - * The port for the connection. Type: integer (or Expression with resultType integer). + * AuthenticationType to be used for connection. Possible values include: 'Basic', 'Windows' */ - port?: any; + authenticationType?: TeradataAuthenticationType; /** * Username for authentication. Type: string (or Expression with resultType string). */ @@ -6308,97 +6404,85 @@ export interface CassandraLinkedService { } /** - * Contains the possible cases for WebLinkedServiceTypeProperties. - */ -export type WebLinkedServiceTypePropertiesUnion = WebLinkedServiceTypeProperties | WebClientCertificateAuthentication | WebBasicAuthentication | WebAnonymousAuthentication; - -/** - * Base definition of WebLinkedServiceTypeProperties, this typeProperties is polymorphic based on - * authenticationType, so not flattened in SDK models. + * Linked service for DB2 data source. */ -export interface WebLinkedServiceTypeProperties { +export interface Db2LinkedService { /** * Polymorphic Discriminator */ - authenticationType: "WebLinkedServiceTypeProperties"; + type: "Db2"; /** - * The URL of the web service endpoint, e.g. http://www.microsoft.com . Type: string (or - * Expression with resultType string). + * The integration runtime reference. */ - url: any; -} - -/** - * A WebLinkedService that uses client certificate based authentication to communicate with an HTTP - * endpoint. This scheme follows mutual authentication; the server must also provide valid - * credentials to the client. - */ -export interface WebClientCertificateAuthentication { + connectVia?: IntegrationRuntimeReference; /** - * Polymorphic Discriminator + * Linked service description. */ - authenticationType: "ClientCertificate"; + description?: string; /** - * The URL of the web service endpoint, e.g. http://www.microsoft.com . Type: string (or - * Expression with resultType string). + * Parameters for linked service. */ - url: any; + parameters?: { [propertyName: string]: ParameterSpecification }; /** - * Base64-encoded contents of a PFX file. + * List of tags that can be used for describing the linked service. */ - pfx: SecretBaseUnion; + annotations?: any[]; /** - * Password for the PFX file. + * The connection string. It is mutually exclusive with server, database, authenticationType, + * userName, packageCollection and certificateCommonName property. Type: string, SecureString or + * AzureKeyVaultSecretReference. */ - password: SecretBaseUnion; -} - -/** - * A WebLinkedService that uses basic authentication to communicate with an HTTP endpoint. - */ -export interface WebBasicAuthentication { + connectionString?: any; /** - * Polymorphic Discriminator + * Server name for connection. It is mutually exclusive with connectionString property. Type: + * string (or Expression with resultType string). */ - authenticationType: "Basic"; + server?: any; /** - * The URL of the web service endpoint, e.g. http://www.microsoft.com . Type: string (or - * Expression with resultType string). + * Database name for connection. It is mutually exclusive with connectionString property. Type: + * string (or Expression with resultType string). */ - url: any; + database?: any; /** - * User name for Basic authentication. Type: string (or Expression with resultType string). + * AuthenticationType to be used for connection. It is mutually exclusive with connectionString + * property. Possible values include: 'Basic' */ - username: any; + authenticationType?: Db2AuthenticationType; /** - * The password for Basic authentication. + * Username for authentication. It is mutually exclusive with connectionString property. Type: + * string (or Expression with resultType string). */ - password: SecretBaseUnion; -} - -/** - * A WebLinkedService that uses anonymous authentication to communicate with an HTTP endpoint. - */ -export interface WebAnonymousAuthentication { + username?: any; /** - * Polymorphic Discriminator + * Password for authentication. */ - authenticationType: "Anonymous"; + password?: SecretBaseUnion; /** - * The URL of the web service endpoint, e.g. http://www.microsoft.com . Type: string (or - * Expression with resultType string). + * Under where packages are created when querying database. It is mutually exclusive with + * connectionString property. Type: string (or Expression with resultType string). */ - url: any; + packageCollection?: any; + /** + * Certificate Common Name when TLS is enabled. It is mutually exclusive with connectionString + * property. Type: string (or Expression with resultType string). + */ + certificateCommonName?: any; + /** + * The encrypted credential used for authentication. Credentials are encrypted using the + * integration runtime credential manager. It is mutually exclusive with connectionString + * property. Type: string (or Expression with resultType string). + */ + encryptedCredential?: any; } /** - * Web linked service. + * Linked service for Sybase data source. */ -export interface WebLinkedService { +export interface SybaseLinkedService { /** * Polymorphic Discriminator */ - type: "Web"; + type: "Sybase"; /** * The integration runtime reference. */ @@ -6416,19 +6500,44 @@ export interface WebLinkedService { */ annotations?: any[]; /** - * Web linked service properties. + * Server name for connection. Type: string (or Expression with resultType string). */ - typeProperties: WebLinkedServiceTypePropertiesUnion; + server: any; + /** + * Database name for connection. Type: string (or Expression with resultType string). + */ + database: any; + /** + * Schema name for connection. Type: string (or Expression with resultType string). + */ + schema?: any; + /** + * AuthenticationType to be used for connection. Possible values include: 'Basic', 'Windows' + */ + authenticationType?: SybaseAuthenticationType; + /** + * Username for authentication. Type: string (or Expression with resultType string). + */ + username?: any; + /** + * Password for authentication. + */ + password?: SecretBaseUnion; + /** + * The encrypted credential used for authentication. Credentials are encrypted using the + * integration runtime credential manager. Type: string (or Expression with resultType string). + */ + encryptedCredential?: any; } /** - * Open Data Protocol (OData) linked service. + * Linked service for PostgreSQL data source. */ -export interface ODataLinkedService { +export interface PostgreSqlLinkedService { /** * Polymorphic Discriminator */ - type: "OData"; + type: "PostgreSql"; /** * The integration runtime reference. */ @@ -6446,57 +6555,13 @@ export interface ODataLinkedService { */ annotations?: any[]; /** - * The URL of the OData service endpoint. Type: string (or Expression with resultType string). - */ - url: any; - /** - * Type of authentication used to connect to the OData service. Possible values include: 'Basic', - * 'Anonymous', 'Windows', 'AadServicePrincipal', 'ManagedServiceIdentity' - */ - authenticationType?: ODataAuthenticationType; - /** - * User name of the OData service. Type: string (or Expression with resultType string). - */ - userName?: any; - /** - * Password of the OData service. - */ - password?: SecretBaseUnion; - /** - * Specify the tenant information (domain name or tenant ID) under which your application - * resides. Type: string (or Expression with resultType string). - */ - tenant?: any; - /** - * Specify the application id of your application registered in Azure Active Directory. Type: - * string (or Expression with resultType string). - */ - servicePrincipalId?: any; - /** - * Specify the resource you are requesting authorization to use Directory. Type: string (or - * Expression with resultType string). - */ - aadResourceId?: any; - /** - * Specify the credential type (key or cert) is used for service principal. Possible values - * include: 'ServicePrincipalKey', 'ServicePrincipalCert' - */ - aadServicePrincipalCredentialType?: ODataAadServicePrincipalCredentialType; - /** - * Specify the secret of your application registered in Azure Active Directory. Type: string (or - * Expression with resultType string). - */ - servicePrincipalKey?: SecretBaseUnion; - /** - * Specify the base64 encoded certificate of your application registered in Azure Active - * Directory. Type: string (or Expression with resultType string). + * The connection string. */ - servicePrincipalEmbeddedCert?: SecretBaseUnion; + connectionString: any; /** - * Specify the password of your certificate if your certificate has a password and you are using - * AadServicePrincipal authentication. Type: string (or Expression with resultType string). + * The Azure key vault secret reference of password in connection string. */ - servicePrincipalEmbeddedCertPassword?: SecretBaseUnion; + password?: AzureKeyVaultSecretReference; /** * The encrypted credential used for authentication. Credentials are encrypted using the * integration runtime credential manager. Type: string (or Expression with resultType string). @@ -6505,13 +6570,13 @@ export interface ODataLinkedService { } /** - * Hadoop Distributed File System (HDFS) linked service. + * Linked service for MySQL data source. */ -export interface HdfsLinkedService { +export interface MySqlLinkedService { /** * Polymorphic Discriminator */ - type: "Hdfs"; + type: "MySql"; /** * The integration runtime reference. */ @@ -6529,38 +6594,28 @@ export interface HdfsLinkedService { */ annotations?: any[]; /** - * The URL of the HDFS service endpoint, e.g. http://myhostname:50070/webhdfs/v1 . Type: string - * (or Expression with resultType string). + * The connection string. */ - url: any; + connectionString: any; /** - * Type of authentication used to connect to the HDFS. Possible values are: Anonymous and - * Windows. Type: string (or Expression with resultType string). + * The Azure key vault secret reference of password in connection string. */ - authenticationType?: any; + password?: AzureKeyVaultSecretReference; /** * The encrypted credential used for authentication. Credentials are encrypted using the * integration runtime credential manager. Type: string (or Expression with resultType string). */ encryptedCredential?: any; - /** - * User name for Windows authentication. Type: string (or Expression with resultType string). - */ - userName?: any; - /** - * Password for Windows authentication. - */ - password?: SecretBaseUnion; } /** - * Microsoft Access linked service. + * Azure MySQL database linked service. */ -export interface MicrosoftAccessLinkedService { +export interface AzureMySqlLinkedService { /** * Polymorphic Discriminator */ - type: "MicrosoftAccess"; + type: "AzureMySql"; /** * The integration runtime reference. */ @@ -6578,28 +6633,13 @@ export interface MicrosoftAccessLinkedService { */ annotations?: any[]; /** - * The non-access credential portion of the connection string as well as an optional encrypted - * credential. Type: string, SecureString or AzureKeyVaultSecretReference. + * The connection string. Type: string, SecureString or AzureKeyVaultSecretReference. */ connectionString: any; /** - * Type of authentication used to connect to the Microsoft Access as ODBC data store. Possible - * values are: Anonymous and Basic. Type: string (or Expression with resultType string). - */ - authenticationType?: any; - /** - * The access credential portion of the connection string specified in driver-specific - * property-value format. - */ - credential?: SecretBaseUnion; - /** - * User name for Basic authentication. Type: string (or Expression with resultType string). - */ - userName?: any; - /** - * Password for Basic authentication. + * The Azure key vault secret reference of password in connection string. */ - password?: SecretBaseUnion; + password?: AzureKeyVaultSecretReference; /** * The encrypted credential used for authentication. Credentials are encrypted using the * integration runtime credential manager. Type: string (or Expression with resultType string). @@ -6608,13 +6648,13 @@ export interface MicrosoftAccessLinkedService { } /** - * Informix linked service. + * Oracle database. */ -export interface InformixLinkedService { +export interface OracleLinkedService { /** * Polymorphic Discriminator */ - type: "Informix"; + type: "Oracle"; /** * The integration runtime reference. */ @@ -6632,28 +6672,13 @@ export interface InformixLinkedService { */ annotations?: any[]; /** - * The non-access credential portion of the connection string as well as an optional encrypted - * credential. Type: string, SecureString or AzureKeyVaultSecretReference. + * The connection string. Type: string, SecureString or AzureKeyVaultSecretReference. */ connectionString: any; /** - * Type of authentication used to connect to the Informix as ODBC data store. Possible values - * are: Anonymous and Basic. Type: string (or Expression with resultType string). - */ - authenticationType?: any; - /** - * The access credential portion of the connection string specified in driver-specific - * property-value format. - */ - credential?: SecretBaseUnion; - /** - * User name for Basic authentication. Type: string (or Expression with resultType string). - */ - userName?: any; - /** - * Password for Basic authentication. + * The Azure key vault secret reference of password in connection string. */ - password?: SecretBaseUnion; + password?: AzureKeyVaultSecretReference; /** * The encrypted credential used for authentication. Credentials are encrypted using the * integration runtime credential manager. Type: string (or Expression with resultType string). @@ -6662,13 +6687,13 @@ export interface InformixLinkedService { } /** - * Open Database Connectivity (ODBC) linked service. + * Linked service for Google Cloud Storage. */ -export interface OdbcLinkedService { +export interface GoogleCloudStorageLinkedService { /** * Polymorphic Discriminator */ - type: "Odbc"; + type: "GoogleCloudStorage"; /** * The integration runtime reference. */ @@ -6686,28 +6711,20 @@ export interface OdbcLinkedService { */ annotations?: any[]; /** - * The non-access credential portion of the connection string as well as an optional encrypted - * credential. Type: string, SecureString or AzureKeyVaultSecretReference. - */ - connectionString: any; - /** - * Type of authentication used to connect to the ODBC data store. Possible values are: Anonymous - * and Basic. Type: string (or Expression with resultType string). - */ - authenticationType?: any; - /** - * The access credential portion of the connection string specified in driver-specific - * property-value format. + * The access key identifier of the Google Cloud Storage Identity and Access Management (IAM) + * user. Type: string (or Expression with resultType string). */ - credential?: SecretBaseUnion; + accessKeyId?: any; /** - * User name for Basic authentication. Type: string (or Expression with resultType string). + * The secret access key of the Google Cloud Storage Identity and Access Management (IAM) user. */ - userName?: any; + secretAccessKey?: SecretBaseUnion; /** - * Password for Basic authentication. + * This value specifies the endpoint to access with the Google Cloud Storage Connector. This is + * an optional property; change it only if you want to try a different service endpoint or want + * to switch between https and http. Type: string (or Expression with resultType string). */ - password?: SecretBaseUnion; + serviceUrl?: any; /** * The encrypted credential used for authentication. Credentials are encrypted using the * integration runtime credential manager. Type: string (or Expression with resultType string). @@ -6716,13 +6733,13 @@ export interface OdbcLinkedService { } /** - * Azure ML Service linked service. + * Azure File Storage linked service. */ -export interface AzureMLServiceLinkedService { +export interface AzureFileStorageLinkedService { /** * Polymorphic Discriminator */ - type: "AzureMLService"; + type: "AzureFileStorage"; /** * The integration runtime reference. */ @@ -6740,34 +6757,17 @@ export interface AzureMLServiceLinkedService { */ annotations?: any[]; /** - * Azure ML Service workspace subscription ID. Type: string (or Expression with resultType - * string). - */ - subscriptionId: any; - /** - * Azure ML Service workspace resource group name. Type: string (or Expression with resultType - * string). - */ - resourceGroupName: any; - /** - * Azure ML Service workspace name. Type: string (or Expression with resultType string). - */ - mlWorkspaceName: any; - /** - * The ID of the service principal used to authenticate against the endpoint of a published Azure - * ML Service pipeline. Type: string (or Expression with resultType string). + * Host name of the server. Type: string (or Expression with resultType string). */ - servicePrincipalId?: any; + host: any; /** - * The key of the service principal used to authenticate against the endpoint of a published - * Azure ML Service pipeline. + * User ID to logon the server. Type: string (or Expression with resultType string). */ - servicePrincipalKey?: SecretBaseUnion; + userId?: any; /** - * The name or ID of the tenant to which the service principal belongs. Type: string (or - * Expression with resultType string). + * Password to logon the server. */ - tenant?: any; + password?: SecretBaseUnion; /** * The encrypted credential used for authentication. Credentials are encrypted using the * integration runtime credential manager. Type: string (or Expression with resultType string). @@ -6776,13 +6776,13 @@ export interface AzureMLServiceLinkedService { } /** - * Azure ML Studio Web Service linked service. + * File system linked service. */ -export interface AzureMLLinkedService { +export interface FileServerLinkedService { /** * Polymorphic Discriminator */ - type: "AzureML"; + type: "FileServer"; /** * The integration runtime reference. */ @@ -6800,35 +6800,17 @@ export interface AzureMLLinkedService { */ annotations?: any[]; /** - * The Batch Execution REST URL for an Azure ML Studio Web Service endpoint. Type: string (or - * Expression with resultType string). - */ - mlEndpoint: any; - /** - * The API key for accessing the Azure ML model endpoint. - */ - apiKey: SecretBaseUnion; - /** - * The Update Resource REST URL for an Azure ML Studio Web Service endpoint. Type: string (or - * Expression with resultType string). - */ - updateResourceEndpoint?: any; - /** - * The ID of the service principal used to authenticate against the ARM-based - * updateResourceEndpoint of an Azure ML Studio web service. Type: string (or Expression with - * resultType string). + * Host name of the server. Type: string (or Expression with resultType string). */ - servicePrincipalId?: any; + host: any; /** - * The key of the service principal used to authenticate against the ARM-based - * updateResourceEndpoint of an Azure ML Studio web service. + * User ID to logon the server. Type: string (or Expression with resultType string). */ - servicePrincipalKey?: SecretBaseUnion; + userId?: any; /** - * The name or ID of the tenant to which the service principal belongs. Type: string (or - * Expression with resultType string). + * Password to logon the server. */ - tenant?: any; + password?: SecretBaseUnion; /** * The encrypted credential used for authentication. Credentials are encrypted using the * integration runtime credential manager. Type: string (or Expression with resultType string). @@ -6837,13 +6819,13 @@ export interface AzureMLLinkedService { } /** - * Linked service for Teradata data source. + * HDInsight linked service. */ -export interface TeradataLinkedService { +export interface HDInsightLinkedService { /** * Polymorphic Discriminator */ - type: "Teradata"; + type: "HDInsight"; /** * The integration runtime reference. */ @@ -6861,40 +6843,49 @@ export interface TeradataLinkedService { */ annotations?: any[]; /** - * Teradata ODBC connection string. Type: string, SecureString or AzureKeyVaultSecretReference. + * HDInsight cluster URI. Type: string (or Expression with resultType string). */ - connectionString?: any; + clusterUri: any; /** - * Server name for connection. Type: string (or Expression with resultType string). + * HDInsight cluster user name. Type: string (or Expression with resultType string). */ - server?: any; + userName?: any; /** - * AuthenticationType to be used for connection. Possible values include: 'Basic', 'Windows' + * HDInsight cluster password. */ - authenticationType?: TeradataAuthenticationType; + password?: SecretBaseUnion; /** - * Username for authentication. Type: string (or Expression with resultType string). + * The Azure Storage linked service reference. */ - username?: any; + linkedServiceName?: LinkedServiceReference; /** - * Password for authentication. + * A reference to the Azure SQL linked service that points to the HCatalog database. */ - password?: SecretBaseUnion; + hcatalogLinkedServiceName?: LinkedServiceReference; /** * The encrypted credential used for authentication. Credentials are encrypted using the * integration runtime credential manager. Type: string (or Expression with resultType string). */ encryptedCredential?: any; + /** + * Specify if the HDInsight is created with ESP (Enterprise Security Package). Type: Boolean. + */ + isEspEnabled?: any; + /** + * Specify the FileSystem if the main storage for the HDInsight is ADLS Gen2. Type: string (or + * Expression with resultType string). + */ + fileSystem?: any; } /** - * Linked service for DB2 data source. + * Common Data Service for Apps linked service. */ -export interface Db2LinkedService { +export interface CommonDataServiceForAppsLinkedService { /** * Polymorphic Discriminator */ - type: "Db2"; + type: "CommonDataServiceForApps"; /** * The integration runtime reference. */ @@ -6912,35 +6903,68 @@ export interface Db2LinkedService { */ annotations?: any[]; /** - * Server name for connection. Type: string (or Expression with resultType string). + * The deployment type of the Common Data Service for Apps instance. 'Online' for Common Data + * Service for Apps Online and 'OnPremisesWithIfd' for Common Data Service for Apps on-premises + * with Ifd. Type: string (or Expression with resultType string). Possible values include: + * 'Online', 'OnPremisesWithIfd' */ - server: any; + deploymentType: DynamicsDeploymentType; /** - * Database name for connection. Type: string (or Expression with resultType string). + * The host name of the on-premises Common Data Service for Apps server. The property is required + * for on-prem and not allowed for online. Type: string (or Expression with resultType string). */ - database: any; + hostName?: any; /** - * AuthenticationType to be used for connection. Possible values include: 'Basic' + * The port of on-premises Common Data Service for Apps server. The property is required for + * on-prem and not allowed for online. Default is 443. Type: integer (or Expression with + * resultType integer), minimum: 0. */ - authenticationType?: Db2AuthenticationType; + port?: any; /** - * Username for authentication. Type: string (or Expression with resultType string). + * The URL to the Microsoft Common Data Service for Apps server. The property is required for + * on-line and not allowed for on-prem. Type: string (or Expression with resultType string). + */ + serviceUri?: any; + /** + * The organization name of the Common Data Service for Apps instance. The property is required + * for on-prem and required for online when there are more than one Common Data Service for Apps + * instances associated with the user. Type: string (or Expression with resultType string). + */ + organizationName?: any; + /** + * The authentication type to connect to Common Data Service for Apps server. 'Office365' for + * online scenario, 'Ifd' for on-premises with Ifd scenario. 'AADServicePrincipal' for + * Server-To-Server authentication in online scenario. Type: string (or Expression with + * resultType string). Possible values include: 'Office365', 'Ifd', 'AADServicePrincipal' + */ + authenticationType: DynamicsAuthenticationType; + /** + * User name to access the Common Data Service for Apps instance. Type: string (or Expression + * with resultType string). */ username?: any; /** - * Password for authentication. + * Password to access the Common Data Service for Apps instance. */ password?: SecretBaseUnion; /** - * Under where packages are created when querying database. Type: string (or Expression with - * resultType string). + * The client ID of the application in Azure Active Directory used for Server-To-Server + * authentication. Type: string (or Expression with resultType string). */ - packageCollection?: any; + servicePrincipalId?: any; /** - * Certificate Common Name when TLS is enabled. Type: string (or Expression with resultType - * string). + * The service principal credential type to use in Server-To-Server authentication. + * 'ServicePrincipalKey' for key/secret, 'ServicePrincipalCert' for certificate. Type: string (or + * Expression with resultType string). */ - certificateCommonName?: any; + servicePrincipalCredentialType?: any; + /** + * The credential of the service principal object in Azure Active Directory. If + * servicePrincipalCredentialType is 'ServicePrincipalKey', servicePrincipalCredential can be + * SecureString or AzureKeyVaultSecretReference. If servicePrincipalCredentialType is + * 'ServicePrincipalCert', servicePrincipalCredential can only be AzureKeyVaultSecretReference. + */ + servicePrincipalCredential?: SecretBaseUnion; /** * The encrypted credential used for authentication. Credentials are encrypted using the * integration runtime credential manager. Type: string (or Expression with resultType string). @@ -6949,53 +6973,91 @@ export interface Db2LinkedService { } /** - * Linked service for Sybase data source. + * Dynamics CRM linked service. */ -export interface SybaseLinkedService { +export interface DynamicsCrmLinkedService { /** * Polymorphic Discriminator */ - type: "Sybase"; + type: "DynamicsCrm"; /** * The integration runtime reference. */ - connectVia?: IntegrationRuntimeReference; + connectVia?: IntegrationRuntimeReference; + /** + * Linked service description. + */ + description?: string; + /** + * Parameters for linked service. + */ + parameters?: { [propertyName: string]: ParameterSpecification }; + /** + * List of tags that can be used for describing the linked service. + */ + annotations?: any[]; + /** + * The deployment type of the Dynamics CRM instance. 'Online' for Dynamics CRM Online and + * 'OnPremisesWithIfd' for Dynamics CRM on-premises with Ifd. Type: string (or Expression with + * resultType string). Possible values include: 'Online', 'OnPremisesWithIfd' + */ + deploymentType: DynamicsDeploymentType; + /** + * The host name of the on-premises Dynamics CRM server. The property is required for on-prem and + * not allowed for online. Type: string (or Expression with resultType string). + */ + hostName?: any; /** - * Linked service description. + * The port of on-premises Dynamics CRM server. The property is required for on-prem and not + * allowed for online. Default is 443. Type: integer (or Expression with resultType integer), + * minimum: 0. */ - description?: string; + port?: any; /** - * Parameters for linked service. + * The URL to the Microsoft Dynamics CRM server. The property is required for on-line and not + * allowed for on-prem. Type: string (or Expression with resultType string). */ - parameters?: { [propertyName: string]: ParameterSpecification }; + serviceUri?: any; /** - * List of tags that can be used for describing the linked service. + * The organization name of the Dynamics CRM instance. The property is required for on-prem and + * required for online when there are more than one Dynamics CRM instances associated with the + * user. Type: string (or Expression with resultType string). */ - annotations?: any[]; + organizationName?: any; /** - * Server name for connection. Type: string (or Expression with resultType string). + * The authentication type to connect to Dynamics CRM server. 'Office365' for online scenario, + * 'Ifd' for on-premises with Ifd scenario, 'AADServicePrincipal' for Server-To-Server + * authentication in online scenario. Type: string (or Expression with resultType string). + * Possible values include: 'Office365', 'Ifd', 'AADServicePrincipal' */ - server: any; + authenticationType: DynamicsAuthenticationType; /** - * Database name for connection. Type: string (or Expression with resultType string). + * User name to access the Dynamics CRM instance. Type: string (or Expression with resultType + * string). */ - database: any; + username?: any; /** - * Schema name for connection. Type: string (or Expression with resultType string). + * Password to access the Dynamics CRM instance. */ - schema?: any; + password?: SecretBaseUnion; /** - * AuthenticationType to be used for connection. Possible values include: 'Basic', 'Windows' + * The client ID of the application in Azure Active Directory used for Server-To-Server + * authentication. Type: string (or Expression with resultType string). */ - authenticationType?: SybaseAuthenticationType; + servicePrincipalId?: any; /** - * Username for authentication. Type: string (or Expression with resultType string). + * The service principal credential type to use in Server-To-Server authentication. + * 'ServicePrincipalKey' for key/secret, 'ServicePrincipalCert' for certificate. Type: string (or + * Expression with resultType string). */ - username?: any; + servicePrincipalCredentialType?: any; /** - * Password for authentication. + * The credential of the service principal object in Azure Active Directory. If + * servicePrincipalCredentialType is 'ServicePrincipalKey', servicePrincipalCredential can be + * SecureString or AzureKeyVaultSecretReference. If servicePrincipalCredentialType is + * 'ServicePrincipalCert', servicePrincipalCredential can only be AzureKeyVaultSecretReference. */ - password?: SecretBaseUnion; + servicePrincipalCredential?: SecretBaseUnion; /** * The encrypted credential used for authentication. Credentials are encrypted using the * integration runtime credential manager. Type: string (or Expression with resultType string). @@ -7004,13 +7066,13 @@ export interface SybaseLinkedService { } /** - * Linked service for PostgreSQL data source. + * Dynamics linked service. */ -export interface PostgreSqlLinkedService { +export interface DynamicsLinkedService { /** * Polymorphic Discriminator */ - type: "PostgreSql"; + type: "Dynamics"; /** * The integration runtime reference. */ @@ -7028,52 +7090,65 @@ export interface PostgreSqlLinkedService { */ annotations?: any[]; /** - * The connection string. + * The deployment type of the Dynamics instance. 'Online' for Dynamics Online and + * 'OnPremisesWithIfd' for Dynamics on-premises with Ifd. Type: string (or Expression with + * resultType string). */ - connectionString: any; + deploymentType: any; /** - * The Azure key vault secret reference of password in connection string. + * The host name of the on-premises Dynamics server. The property is required for on-prem and not + * allowed for online. Type: string (or Expression with resultType string). */ - password?: AzureKeyVaultSecretReference; + hostName?: any; /** - * The encrypted credential used for authentication. Credentials are encrypted using the - * integration runtime credential manager. Type: string (or Expression with resultType string). + * The port of on-premises Dynamics server. The property is required for on-prem and not allowed + * for online. Default is 443. Type: integer (or Expression with resultType integer), minimum: 0. */ - encryptedCredential?: any; -} - -/** - * Linked service for MySQL data source. - */ -export interface MySqlLinkedService { + port?: any; /** - * Polymorphic Discriminator + * The URL to the Microsoft Dynamics server. The property is required for on-line and not allowed + * for on-prem. Type: string (or Expression with resultType string). */ - type: "MySql"; + serviceUri?: any; /** - * The integration runtime reference. + * The organization name of the Dynamics instance. The property is required for on-prem and + * required for online when there are more than one Dynamics instances associated with the user. + * Type: string (or Expression with resultType string). */ - connectVia?: IntegrationRuntimeReference; + organizationName?: any; /** - * Linked service description. + * The authentication type to connect to Dynamics server. 'Office365' for online scenario, 'Ifd' + * for on-premises with Ifd scenario, 'AADServicePrincipal' for Server-To-Server authentication + * in online scenario. Type: string (or Expression with resultType string). */ - description?: string; + authenticationType: any; /** - * Parameters for linked service. + * User name to access the Dynamics instance. Type: string (or Expression with resultType + * string). */ - parameters?: { [propertyName: string]: ParameterSpecification }; + username?: any; /** - * List of tags that can be used for describing the linked service. + * Password to access the Dynamics instance. */ - annotations?: any[]; + password?: SecretBaseUnion; /** - * The connection string. + * The client ID of the application in Azure Active Directory used for Server-To-Server + * authentication. Type: string (or Expression with resultType string). */ - connectionString: any; + servicePrincipalId?: any; /** - * The Azure key vault secret reference of password in connection string. + * The service principal credential type to use in Server-To-Server authentication. + * 'ServicePrincipalKey' for key/secret, 'ServicePrincipalCert' for certificate. Type: string (or + * Expression with resultType string). */ - password?: AzureKeyVaultSecretReference; + servicePrincipalCredentialType?: any; + /** + * The credential of the service principal object in Azure Active Directory. If + * servicePrincipalCredentialType is 'ServicePrincipalKey', servicePrincipalCredential can be + * SecureString or AzureKeyVaultSecretReference. If servicePrincipalCredentialType is + * 'ServicePrincipalCert', servicePrincipalCredential can only be AzureKeyVaultSecretReference. + */ + servicePrincipalCredential?: SecretBaseUnion; /** * The encrypted credential used for authentication. Credentials are encrypted using the * integration runtime credential manager. Type: string (or Expression with resultType string). @@ -7082,13 +7157,13 @@ export interface MySqlLinkedService { } /** - * Azure MySQL database linked service. + * Microsoft Azure Cosmos Database (CosmosDB) linked service. */ -export interface AzureMySqlLinkedService { +export interface CosmosDbLinkedService { /** * Polymorphic Discriminator */ - type: "AzureMySql"; + type: "CosmosDb"; /** * The integration runtime reference. */ @@ -7108,11 +7183,21 @@ export interface AzureMySqlLinkedService { /** * The connection string. Type: string, SecureString or AzureKeyVaultSecretReference. */ - connectionString: any; + connectionString?: any; /** - * The Azure key vault secret reference of password in connection string. + * The endpoint of the Azure CosmosDB account. Type: string (or Expression with resultType + * string) */ - password?: AzureKeyVaultSecretReference; + accountEndpoint?: any; + /** + * The name of the database. Type: string (or Expression with resultType string) + */ + database?: any; + /** + * The account key of the Azure CosmosDB account. Type: SecureString or + * AzureKeyVaultSecretReference. + */ + accountKey?: SecretBaseUnion; /** * The encrypted credential used for authentication. Credentials are encrypted using the * integration runtime credential manager. Type: string (or Expression with resultType string). @@ -7121,13 +7206,13 @@ export interface AzureMySqlLinkedService { } /** - * Oracle database. + * Azure Key Vault linked service. */ -export interface OracleLinkedService { +export interface AzureKeyVaultLinkedService { /** * Polymorphic Discriminator */ - type: "Oracle"; + type: "AzureKeyVault"; /** * The integration runtime reference. */ @@ -7145,28 +7230,20 @@ export interface OracleLinkedService { */ annotations?: any[]; /** - * The connection string. Type: string, SecureString or AzureKeyVaultSecretReference. - */ - connectionString: any; - /** - * The Azure key vault secret reference of password in connection string. - */ - password?: AzureKeyVaultSecretReference; - /** - * The encrypted credential used for authentication. Credentials are encrypted using the - * integration runtime credential manager. Type: string (or Expression with resultType string). + * The base URL of the Azure Key Vault. e.g. https://myakv.vault.azure.net Type: string (or + * Expression with resultType string). */ - encryptedCredential?: any; + baseUrl: any; } /** - * Linked service for Google Cloud Storage. + * Azure Batch linked service. */ -export interface GoogleCloudStorageLinkedService { +export interface AzureBatchLinkedService { /** * Polymorphic Discriminator */ - type: "GoogleCloudStorage"; + type: "AzureBatch"; /** * The integration runtime reference. */ @@ -7184,20 +7261,25 @@ export interface GoogleCloudStorageLinkedService { */ annotations?: any[]; /** - * The access key identifier of the Google Cloud Storage Identity and Access Management (IAM) - * user. Type: string (or Expression with resultType string). + * The Azure Batch account name. Type: string (or Expression with resultType string). */ - accessKeyId?: any; + accountName: any; /** - * The secret access key of the Google Cloud Storage Identity and Access Management (IAM) user. + * The Azure Batch account access key. */ - secretAccessKey?: SecretBaseUnion; + accessKey?: SecretBaseUnion; /** - * This value specifies the endpoint to access with the Google Cloud Storage Connector. This is - * an optional property; change it only if you want to try a different service endpoint or want - * to switch between https and http. Type: string (or Expression with resultType string). + * The Azure Batch URI. Type: string (or Expression with resultType string). */ - serviceUrl?: any; + batchUri: any; + /** + * The Azure Batch pool name. Type: string (or Expression with resultType string). + */ + poolName: any; + /** + * The Azure Storage linked service reference. + */ + linkedServiceName: LinkedServiceReference; /** * The encrypted credential used for authentication. Credentials are encrypted using the * integration runtime credential manager. Type: string (or Expression with resultType string). @@ -7206,13 +7288,13 @@ export interface GoogleCloudStorageLinkedService { } /** - * Azure File Storage linked service. + * Azure SQL Managed Instance linked service. */ -export interface AzureFileStorageLinkedService { +export interface AzureSqlMILinkedService { /** * Polymorphic Discriminator */ - type: "AzureFileStorage"; + type: "AzureSqlMI"; /** * The integration runtime reference. */ @@ -7230,17 +7312,27 @@ export interface AzureFileStorageLinkedService { */ annotations?: any[]; /** - * Host name of the server. Type: string (or Expression with resultType string). + * The connection string. Type: string, SecureString or AzureKeyVaultSecretReference. */ - host: any; + connectionString: any; /** - * User ID to logon the server. Type: string (or Expression with resultType string). + * The Azure key vault secret reference of password in connection string. */ - userId?: any; + password?: AzureKeyVaultSecretReference; /** - * Password to logon the server. + * The ID of the service principal used to authenticate against Azure SQL Managed Instance. Type: + * string (or Expression with resultType string). */ - password?: SecretBaseUnion; + servicePrincipalId?: any; + /** + * The key of the service principal used to authenticate against Azure SQL Managed Instance. + */ + servicePrincipalKey?: SecretBaseUnion; + /** + * The name or ID of the tenant to which the service principal belongs. Type: string (or + * Expression with resultType string). + */ + tenant?: any; /** * The encrypted credential used for authentication. Credentials are encrypted using the * integration runtime credential manager. Type: string (or Expression with resultType string). @@ -7249,13 +7341,13 @@ export interface AzureFileStorageLinkedService { } /** - * File system linked service. + * Microsoft Azure SQL Database linked service. */ -export interface FileServerLinkedService { +export interface AzureSqlDatabaseLinkedService { /** * Polymorphic Discriminator */ - type: "FileServer"; + type: "AzureSqlDatabase"; /** * The integration runtime reference. */ @@ -7273,17 +7365,27 @@ export interface FileServerLinkedService { */ annotations?: any[]; /** - * Host name of the server. Type: string (or Expression with resultType string). + * The connection string. Type: string, SecureString or AzureKeyVaultSecretReference. */ - host: any; + connectionString: any; /** - * User ID to logon the server. Type: string (or Expression with resultType string). + * The Azure key vault secret reference of password in connection string. */ - userId?: any; + password?: AzureKeyVaultSecretReference; /** - * Password to logon the server. + * The ID of the service principal used to authenticate against Azure SQL Database. Type: string + * (or Expression with resultType string). */ - password?: SecretBaseUnion; + servicePrincipalId?: any; + /** + * The key of the service principal used to authenticate against Azure SQL Database. + */ + servicePrincipalKey?: SecretBaseUnion; + /** + * The name or ID of the tenant to which the service principal belongs. Type: string (or + * Expression with resultType string). + */ + tenant?: any; /** * The encrypted credential used for authentication. Credentials are encrypted using the * integration runtime credential manager. Type: string (or Expression with resultType string). @@ -7292,13 +7394,13 @@ export interface FileServerLinkedService { } /** - * HDInsight linked service. + * SQL Server linked service. */ -export interface HDInsightLinkedService { +export interface SqlServerLinkedService { /** * Polymorphic Discriminator */ - type: "HDInsight"; + type: "SqlServer"; /** * The integration runtime reference. */ @@ -7316,49 +7418,33 @@ export interface HDInsightLinkedService { */ annotations?: any[]; /** - * HDInsight cluster URI. Type: string (or Expression with resultType string). + * The connection string. Type: string, SecureString or AzureKeyVaultSecretReference. */ - clusterUri: any; + connectionString: any; /** - * HDInsight cluster user name. Type: string (or Expression with resultType string). + * The on-premises Windows authentication user name. Type: string (or Expression with resultType + * string). */ userName?: any; /** - * HDInsight cluster password. + * The on-premises Windows authentication password. */ password?: SecretBaseUnion; - /** - * The Azure Storage linked service reference. - */ - linkedServiceName?: LinkedServiceReference; - /** - * A reference to the Azure SQL linked service that points to the HCatalog database. - */ - hcatalogLinkedServiceName?: LinkedServiceReference; /** * The encrypted credential used for authentication. Credentials are encrypted using the * integration runtime credential manager. Type: string (or Expression with resultType string). */ encryptedCredential?: any; - /** - * Specify if the HDInsight is created with ESP (Enterprise Security Package). Type: Boolean. - */ - isEspEnabled?: any; - /** - * Specify the FileSystem if the main storage for the HDInsight is ADLS Gen2. Type: string (or - * Expression with resultType string). - */ - fileSystem?: any; } /** - * Common Data Service for Apps linked service. + * Azure SQL Data Warehouse linked service. */ -export interface CommonDataServiceForAppsLinkedService { +export interface AzureSqlDWLinkedService { /** * Polymorphic Discriminator */ - type: "CommonDataServiceForApps"; + type: "AzureSqlDW"; /** * The integration runtime reference. */ @@ -7376,68 +7462,28 @@ export interface CommonDataServiceForAppsLinkedService { */ annotations?: any[]; /** - * The deployment type of the Common Data Service for Apps instance. 'Online' for Common Data - * Service for Apps Online and 'OnPremisesWithIfd' for Common Data Service for Apps on-premises - * with Ifd. Type: string (or Expression with resultType string). Possible values include: - * 'Online', 'OnPremisesWithIfd' - */ - deploymentType: DynamicsDeploymentType; - /** - * The host name of the on-premises Common Data Service for Apps server. The property is required - * for on-prem and not allowed for online. Type: string (or Expression with resultType string). - */ - hostName?: any; - /** - * The port of on-premises Common Data Service for Apps server. The property is required for - * on-prem and not allowed for online. Default is 443. Type: integer (or Expression with - * resultType integer), minimum: 0. - */ - port?: any; - /** - * The URL to the Microsoft Common Data Service for Apps server. The property is required for - * on-line and not allowed for on-prem. Type: string (or Expression with resultType string). - */ - serviceUri?: any; - /** - * The organization name of the Common Data Service for Apps instance. The property is required - * for on-prem and required for online when there are more than one Common Data Service for Apps - * instances associated with the user. Type: string (or Expression with resultType string). - */ - organizationName?: any; - /** - * The authentication type to connect to Common Data Service for Apps server. 'Office365' for - * online scenario, 'Ifd' for on-premises with Ifd scenario. 'AADServicePrincipal' for - * Server-To-Server authentication in online scenario. Type: string (or Expression with - * resultType string). Possible values include: 'Office365', 'Ifd', 'AADServicePrincipal' - */ - authenticationType: DynamicsAuthenticationType; - /** - * User name to access the Common Data Service for Apps instance. Type: string (or Expression - * with resultType string). + * The connection string. Type: string, SecureString or AzureKeyVaultSecretReference. Type: + * string, SecureString or AzureKeyVaultSecretReference. */ - username?: any; + connectionString: any; /** - * Password to access the Common Data Service for Apps instance. + * The Azure key vault secret reference of password in connection string. */ - password?: SecretBaseUnion; + password?: AzureKeyVaultSecretReference; /** - * The client ID of the application in Azure Active Directory used for Server-To-Server - * authentication. Type: string (or Expression with resultType string). + * The ID of the service principal used to authenticate against Azure SQL Data Warehouse. Type: + * string (or Expression with resultType string). */ servicePrincipalId?: any; /** - * The service principal credential type to use in Server-To-Server authentication. - * 'ServicePrincipalKey' for key/secret, 'ServicePrincipalCert' for certificate. Type: string (or - * Expression with resultType string). + * The key of the service principal used to authenticate against Azure SQL Data Warehouse. */ - servicePrincipalCredentialType?: any; + servicePrincipalKey?: SecretBaseUnion; /** - * The credential of the service principal object in Azure Active Directory. If - * servicePrincipalCredentialType is 'ServicePrincipalKey', servicePrincipalCredential can be - * SecureString or AzureKeyVaultSecretReference. If servicePrincipalCredentialType is - * 'ServicePrincipalCert', servicePrincipalCredential can only be AzureKeyVaultSecretReference. + * The name or ID of the tenant to which the service principal belongs. Type: string (or + * Expression with resultType string). */ - servicePrincipalCredential?: SecretBaseUnion; + tenant?: any; /** * The encrypted credential used for authentication. Credentials are encrypted using the * integration runtime credential manager. Type: string (or Expression with resultType string). @@ -7446,13 +7492,13 @@ export interface CommonDataServiceForAppsLinkedService { } /** - * Dynamics CRM linked service. + * The azure table storage linked service. */ -export interface DynamicsCrmLinkedService { +export interface AzureTableStorageLinkedService { /** * Polymorphic Discriminator */ - type: "DynamicsCrm"; + type: "AzureTableStorage"; /** * The integration runtime reference. */ @@ -7470,82 +7516,38 @@ export interface DynamicsCrmLinkedService { */ annotations?: any[]; /** - * The deployment type of the Dynamics CRM instance. 'Online' for Dynamics CRM Online and - * 'OnPremisesWithIfd' for Dynamics CRM on-premises with Ifd. Type: string (or Expression with - * resultType string). Possible values include: 'Online', 'OnPremisesWithIfd' - */ - deploymentType: DynamicsDeploymentType; - /** - * The host name of the on-premises Dynamics CRM server. The property is required for on-prem and - * not allowed for online. Type: string (or Expression with resultType string). - */ - hostName?: any; - /** - * The port of on-premises Dynamics CRM server. The property is required for on-prem and not - * allowed for online. Default is 443. Type: integer (or Expression with resultType integer), - * minimum: 0. - */ - port?: any; - /** - * The URL to the Microsoft Dynamics CRM server. The property is required for on-line and not - * allowed for on-prem. Type: string (or Expression with resultType string). - */ - serviceUri?: any; - /** - * The organization name of the Dynamics CRM instance. The property is required for on-prem and - * required for online when there are more than one Dynamics CRM instances associated with the - * user. Type: string (or Expression with resultType string). - */ - organizationName?: any; - /** - * The authentication type to connect to Dynamics CRM server. 'Office365' for online scenario, - * 'Ifd' for on-premises with Ifd scenario, 'AADServicePrincipal' for Server-To-Server - * authentication in online scenario. Type: string (or Expression with resultType string). - * Possible values include: 'Office365', 'Ifd', 'AADServicePrincipal' - */ - authenticationType: DynamicsAuthenticationType; - /** - * User name to access the Dynamics CRM instance. Type: string (or Expression with resultType - * string). - */ - username?: any; - /** - * Password to access the Dynamics CRM instance. + * The connection string. It is mutually exclusive with sasUri property. Type: string, + * SecureString or AzureKeyVaultSecretReference. */ - password?: SecretBaseUnion; + connectionString?: any; /** - * The client ID of the application in Azure Active Directory used for Server-To-Server - * authentication. Type: string (or Expression with resultType string). + * The Azure key vault secret reference of accountKey in connection string. */ - servicePrincipalId?: any; + accountKey?: AzureKeyVaultSecretReference; /** - * The service principal credential type to use in Server-To-Server authentication. - * 'ServicePrincipalKey' for key/secret, 'ServicePrincipalCert' for certificate. Type: string (or - * Expression with resultType string). + * SAS URI of the Azure Storage resource. It is mutually exclusive with connectionString + * property. Type: string, SecureString or AzureKeyVaultSecretReference. */ - servicePrincipalCredentialType?: any; + sasUri?: any; /** - * The credential of the service principal object in Azure Active Directory. If - * servicePrincipalCredentialType is 'ServicePrincipalKey', servicePrincipalCredential can be - * SecureString or AzureKeyVaultSecretReference. If servicePrincipalCredentialType is - * 'ServicePrincipalCert', servicePrincipalCredential can only be AzureKeyVaultSecretReference. + * The Azure key vault secret reference of sasToken in sas uri. */ - servicePrincipalCredential?: SecretBaseUnion; + sasToken?: AzureKeyVaultSecretReference; /** * The encrypted credential used for authentication. Credentials are encrypted using the * integration runtime credential manager. Type: string (or Expression with resultType string). */ - encryptedCredential?: any; + encryptedCredential?: string; } /** - * Dynamics linked service. + * The azure blob storage linked service. */ -export interface DynamicsLinkedService { +export interface AzureBlobStorageLinkedService { /** * Polymorphic Discriminator */ - type: "Dynamics"; + type: "AzureBlobStorage"; /** * The integration runtime reference. */ @@ -7563,80 +7565,57 @@ export interface DynamicsLinkedService { */ annotations?: any[]; /** - * The deployment type of the Dynamics instance. 'Online' for Dynamics Online and - * 'OnPremisesWithIfd' for Dynamics on-premises with Ifd. Type: string (or Expression with - * resultType string). - */ - deploymentType: any; - /** - * The host name of the on-premises Dynamics server. The property is required for on-prem and not - * allowed for online. Type: string (or Expression with resultType string). - */ - hostName?: any; - /** - * The port of on-premises Dynamics server. The property is required for on-prem and not allowed - * for online. Default is 443. Type: integer (or Expression with resultType integer), minimum: 0. - */ - port?: any; - /** - * The URL to the Microsoft Dynamics server. The property is required for on-line and not allowed - * for on-prem. Type: string (or Expression with resultType string). + * The connection string. It is mutually exclusive with sasUri, serviceEndpoint property. Type: + * string, SecureString or AzureKeyVaultSecretReference. */ - serviceUri?: any; + connectionString?: any; /** - * The organization name of the Dynamics instance. The property is required for on-prem and - * required for online when there are more than one Dynamics instances associated with the user. - * Type: string (or Expression with resultType string). + * The Azure key vault secret reference of accountKey in connection string. */ - organizationName?: any; + accountKey?: AzureKeyVaultSecretReference; /** - * The authentication type to connect to Dynamics server. 'Office365' for online scenario, 'Ifd' - * for on-premises with Ifd scenario, 'AADServicePrincipal' for Server-To-Server authentication - * in online scenario. Type: string (or Expression with resultType string). + * SAS URI of the Azure Blob Storage resource. It is mutually exclusive with connectionString, + * serviceEndpoint property. Type: string, SecureString or AzureKeyVaultSecretReference. */ - authenticationType: any; + sasUri?: any; /** - * User name to access the Dynamics instance. Type: string (or Expression with resultType - * string). + * The Azure key vault secret reference of sasToken in sas uri. */ - username?: any; + sasToken?: AzureKeyVaultSecretReference; /** - * Password to access the Dynamics instance. + * Blob service endpoint of the Azure Blob Storage resource. It is mutually exclusive with + * connectionString, sasUri property. */ - password?: SecretBaseUnion; + serviceEndpoint?: string; /** - * The client ID of the application in Azure Active Directory used for Server-To-Server - * authentication. Type: string (or Expression with resultType string). + * The ID of the service principal used to authenticate against Azure SQL Data Warehouse. Type: + * string (or Expression with resultType string). */ servicePrincipalId?: any; /** - * The service principal credential type to use in Server-To-Server authentication. - * 'ServicePrincipalKey' for key/secret, 'ServicePrincipalCert' for certificate. Type: string (or - * Expression with resultType string). + * The key of the service principal used to authenticate against Azure SQL Data Warehouse. */ - servicePrincipalCredentialType?: any; + servicePrincipalKey?: SecretBaseUnion; /** - * The credential of the service principal object in Azure Active Directory. If - * servicePrincipalCredentialType is 'ServicePrincipalKey', servicePrincipalCredential can be - * SecureString or AzureKeyVaultSecretReference. If servicePrincipalCredentialType is - * 'ServicePrincipalCert', servicePrincipalCredential can only be AzureKeyVaultSecretReference. + * The name or ID of the tenant to which the service principal belongs. Type: string (or + * Expression with resultType string). */ - servicePrincipalCredential?: SecretBaseUnion; + tenant?: any; /** * The encrypted credential used for authentication. Credentials are encrypted using the * integration runtime credential manager. Type: string (or Expression with resultType string). */ - encryptedCredential?: any; + encryptedCredential?: string; } /** - * Microsoft Azure Cosmos Database (CosmosDB) linked service. + * The storage account linked service. */ -export interface CosmosDbLinkedService { +export interface AzureStorageLinkedService { /** * Polymorphic Discriminator */ - type: "CosmosDb"; + type: "AzureStorage"; /** * The integration runtime reference. */ @@ -7654,490 +7633,451 @@ export interface CosmosDbLinkedService { */ annotations?: any[]; /** - * The connection string. Type: string, SecureString or AzureKeyVaultSecretReference. + * The connection string. It is mutually exclusive with sasUri property. Type: string, + * SecureString or AzureKeyVaultSecretReference. */ connectionString?: any; - /** - * The endpoint of the Azure CosmosDB account. Type: string (or Expression with resultType - * string) + /** + * The Azure key vault secret reference of accountKey in connection string. */ - accountEndpoint?: any; + accountKey?: AzureKeyVaultSecretReference; /** - * The name of the database. Type: string (or Expression with resultType string) + * SAS URI of the Azure Storage resource. It is mutually exclusive with connectionString + * property. Type: string, SecureString or AzureKeyVaultSecretReference. */ - database?: any; + sasUri?: any; /** - * The account key of the Azure CosmosDB account. Type: SecureString or - * AzureKeyVaultSecretReference. + * The Azure key vault secret reference of sasToken in sas uri. */ - accountKey?: SecretBaseUnion; + sasToken?: AzureKeyVaultSecretReference; /** * The encrypted credential used for authentication. Credentials are encrypted using the * integration runtime credential manager. Type: string (or Expression with resultType string). */ - encryptedCredential?: any; + encryptedCredential?: string; } /** - * Azure Key Vault linked service. + * Google AdWords service dataset. */ -export interface AzureKeyVaultLinkedService { +export interface GoogleAdWordsObjectDataset { /** * Polymorphic Discriminator */ - type: "AzureKeyVault"; + type: "GoogleAdWordsObject"; /** - * The integration runtime reference. + * Dataset description. */ - connectVia?: IntegrationRuntimeReference; + description?: string; /** - * Linked service description. + * Columns that define the structure of the dataset. Type: array (or Expression with resultType + * array), itemType: DatasetDataElement. */ - description?: string; + structure?: any; /** - * Parameters for linked service. + * Columns that define the physical type schema of the dataset. Type: array (or Expression with + * resultType array), itemType: DatasetSchemaDataElement. + */ + schema?: any; + /** + * Linked service reference. + */ + linkedServiceName: LinkedServiceReference; + /** + * Parameters for dataset. */ parameters?: { [propertyName: string]: ParameterSpecification }; /** - * List of tags that can be used for describing the linked service. + * List of tags that can be used for describing the Dataset. */ annotations?: any[]; /** - * The base URL of the Azure Key Vault. e.g. https://myakv.vault.azure.net Type: string (or - * Expression with resultType string). + * The folder that this Dataset is in. If not specified, Dataset will appear at the root level. */ - baseUrl: any; + folder?: DatasetFolder; + /** + * The table name. Type: string (or Expression with resultType string). + */ + tableName?: any; } /** - * Azure Batch linked service. + * The Azure Data Explorer (Kusto) dataset. */ -export interface AzureBatchLinkedService { +export interface AzureDataExplorerTableDataset { /** * Polymorphic Discriminator */ - type: "AzureBatch"; - /** - * The integration runtime reference. - */ - connectVia?: IntegrationRuntimeReference; + type: "AzureDataExplorerTable"; /** - * Linked service description. + * Dataset description. */ description?: string; /** - * Parameters for linked service. - */ - parameters?: { [propertyName: string]: ParameterSpecification }; - /** - * List of tags that can be used for describing the linked service. + * Columns that define the structure of the dataset. Type: array (or Expression with resultType + * array), itemType: DatasetDataElement. */ - annotations?: any[]; + structure?: any; /** - * The Azure Batch account name. Type: string (or Expression with resultType string). + * Columns that define the physical type schema of the dataset. Type: array (or Expression with + * resultType array), itemType: DatasetSchemaDataElement. */ - accountName: any; + schema?: any; /** - * The Azure Batch account access key. + * Linked service reference. */ - accessKey?: SecretBaseUnion; + linkedServiceName: LinkedServiceReference; /** - * The Azure Batch URI. Type: string (or Expression with resultType string). + * Parameters for dataset. */ - batchUri: any; + parameters?: { [propertyName: string]: ParameterSpecification }; /** - * The Azure Batch pool name. Type: string (or Expression with resultType string). + * List of tags that can be used for describing the Dataset. */ - poolName: any; + annotations?: any[]; /** - * The Azure Storage linked service reference. + * The folder that this Dataset is in. If not specified, Dataset will appear at the root level. */ - linkedServiceName: LinkedServiceReference; + folder?: DatasetFolder; /** - * The encrypted credential used for authentication. Credentials are encrypted using the - * integration runtime credential manager. Type: string (or Expression with resultType string). + * The table name of the Azure Data Explorer database. Type: string (or Expression with + * resultType string). */ - encryptedCredential?: any; + table?: any; } /** - * Azure SQL Managed Instance linked service. + * Oracle Service Cloud dataset. */ -export interface AzureSqlMILinkedService { +export interface OracleServiceCloudObjectDataset { /** * Polymorphic Discriminator */ - type: "AzureSqlMI"; - /** - * The integration runtime reference. - */ - connectVia?: IntegrationRuntimeReference; + type: "OracleServiceCloudObject"; /** - * Linked service description. + * Dataset description. */ description?: string; /** - * Parameters for linked service. - */ - parameters?: { [propertyName: string]: ParameterSpecification }; - /** - * List of tags that can be used for describing the linked service. + * Columns that define the structure of the dataset. Type: array (or Expression with resultType + * array), itemType: DatasetDataElement. */ - annotations?: any[]; + structure?: any; /** - * The connection string. Type: string, SecureString or AzureKeyVaultSecretReference. + * Columns that define the physical type schema of the dataset. Type: array (or Expression with + * resultType array), itemType: DatasetSchemaDataElement. */ - connectionString: any; + schema?: any; /** - * The Azure key vault secret reference of password in connection string. + * Linked service reference. */ - password?: AzureKeyVaultSecretReference; + linkedServiceName: LinkedServiceReference; /** - * The ID of the service principal used to authenticate against Azure SQL Managed Instance. Type: - * string (or Expression with resultType string). + * Parameters for dataset. */ - servicePrincipalId?: any; + parameters?: { [propertyName: string]: ParameterSpecification }; /** - * The key of the service principal used to authenticate against Azure SQL Managed Instance. + * List of tags that can be used for describing the Dataset. */ - servicePrincipalKey?: SecretBaseUnion; + annotations?: any[]; /** - * The name or ID of the tenant to which the service principal belongs. Type: string (or - * Expression with resultType string). + * The folder that this Dataset is in. If not specified, Dataset will appear at the root level. */ - tenant?: any; + folder?: DatasetFolder; /** - * The encrypted credential used for authentication. Credentials are encrypted using the - * integration runtime credential manager. Type: string (or Expression with resultType string). + * The table name. Type: string (or Expression with resultType string). */ - encryptedCredential?: any; + tableName?: any; } /** - * Microsoft Azure SQL Database linked service. + * The path of the Dynamics AX OData entity. */ -export interface AzureSqlDatabaseLinkedService { +export interface DynamicsAXResourceDataset { /** * Polymorphic Discriminator */ - type: "AzureSqlDatabase"; - /** - * The integration runtime reference. - */ - connectVia?: IntegrationRuntimeReference; + type: "DynamicsAXResource"; /** - * Linked service description. + * Dataset description. */ description?: string; /** - * Parameters for linked service. - */ - parameters?: { [propertyName: string]: ParameterSpecification }; - /** - * List of tags that can be used for describing the linked service. + * Columns that define the structure of the dataset. Type: array (or Expression with resultType + * array), itemType: DatasetDataElement. */ - annotations?: any[]; + structure?: any; /** - * The connection string. Type: string, SecureString or AzureKeyVaultSecretReference. + * Columns that define the physical type schema of the dataset. Type: array (or Expression with + * resultType array), itemType: DatasetSchemaDataElement. */ - connectionString: any; + schema?: any; /** - * The Azure key vault secret reference of password in connection string. + * Linked service reference. */ - password?: AzureKeyVaultSecretReference; + linkedServiceName: LinkedServiceReference; /** - * The ID of the service principal used to authenticate against Azure SQL Database. Type: string - * (or Expression with resultType string). + * Parameters for dataset. */ - servicePrincipalId?: any; + parameters?: { [propertyName: string]: ParameterSpecification }; /** - * The key of the service principal used to authenticate against Azure SQL Database. + * List of tags that can be used for describing the Dataset. */ - servicePrincipalKey?: SecretBaseUnion; + annotations?: any[]; /** - * The name or ID of the tenant to which the service principal belongs. Type: string (or - * Expression with resultType string). + * The folder that this Dataset is in. If not specified, Dataset will appear at the root level. */ - tenant?: any; + folder?: DatasetFolder; /** - * The encrypted credential used for authentication. Credentials are encrypted using the - * integration runtime credential manager. Type: string (or Expression with resultType string). + * The path of the Dynamics AX OData entity. Type: string (or Expression with resultType string). */ - encryptedCredential?: any; + path: any; } /** - * SQL Server linked service. + * Responsys dataset. */ -export interface SqlServerLinkedService { +export interface ResponsysObjectDataset { /** * Polymorphic Discriminator */ - type: "SqlServer"; + type: "ResponsysObject"; /** - * The integration runtime reference. + * Dataset description. */ - connectVia?: IntegrationRuntimeReference; + description?: string; /** - * Linked service description. + * Columns that define the structure of the dataset. Type: array (or Expression with resultType + * array), itemType: DatasetDataElement. */ - description?: string; + structure?: any; /** - * Parameters for linked service. + * Columns that define the physical type schema of the dataset. Type: array (or Expression with + * resultType array), itemType: DatasetSchemaDataElement. */ - parameters?: { [propertyName: string]: ParameterSpecification }; + schema?: any; /** - * List of tags that can be used for describing the linked service. + * Linked service reference. */ - annotations?: any[]; + linkedServiceName: LinkedServiceReference; /** - * The connection string. Type: string, SecureString or AzureKeyVaultSecretReference. + * Parameters for dataset. */ - connectionString: any; + parameters?: { [propertyName: string]: ParameterSpecification }; /** - * The on-premises Windows authentication user name. Type: string (or Expression with resultType - * string). + * List of tags that can be used for describing the Dataset. */ - userName?: any; + annotations?: any[]; /** - * The on-premises Windows authentication password. + * The folder that this Dataset is in. If not specified, Dataset will appear at the root level. */ - password?: SecretBaseUnion; + folder?: DatasetFolder; /** - * The encrypted credential used for authentication. Credentials are encrypted using the - * integration runtime credential manager. Type: string (or Expression with resultType string). + * The table name. Type: string (or Expression with resultType string). */ - encryptedCredential?: any; + tableName?: any; } /** - * Azure SQL Data Warehouse linked service. + * Salesforce Marketing Cloud dataset. */ -export interface AzureSqlDWLinkedService { +export interface SalesforceMarketingCloudObjectDataset { /** * Polymorphic Discriminator */ - type: "AzureSqlDW"; - /** - * The integration runtime reference. - */ - connectVia?: IntegrationRuntimeReference; - /** - * Linked service description. - */ - description?: string; + type: "SalesforceMarketingCloudObject"; /** - * Parameters for linked service. + * Dataset description. */ - parameters?: { [propertyName: string]: ParameterSpecification }; + description?: string; /** - * List of tags that can be used for describing the linked service. + * Columns that define the structure of the dataset. Type: array (or Expression with resultType + * array), itemType: DatasetDataElement. */ - annotations?: any[]; + structure?: any; /** - * The connection string. Type: string, SecureString or AzureKeyVaultSecretReference. Type: - * string, SecureString or AzureKeyVaultSecretReference. + * Columns that define the physical type schema of the dataset. Type: array (or Expression with + * resultType array), itemType: DatasetSchemaDataElement. */ - connectionString: any; + schema?: any; /** - * The Azure key vault secret reference of password in connection string. + * Linked service reference. */ - password?: AzureKeyVaultSecretReference; + linkedServiceName: LinkedServiceReference; /** - * The ID of the service principal used to authenticate against Azure SQL Data Warehouse. Type: - * string (or Expression with resultType string). + * Parameters for dataset. */ - servicePrincipalId?: any; + parameters?: { [propertyName: string]: ParameterSpecification }; /** - * The key of the service principal used to authenticate against Azure SQL Data Warehouse. + * List of tags that can be used for describing the Dataset. */ - servicePrincipalKey?: SecretBaseUnion; + annotations?: any[]; /** - * The name or ID of the tenant to which the service principal belongs. Type: string (or - * Expression with resultType string). + * The folder that this Dataset is in. If not specified, Dataset will appear at the root level. */ - tenant?: any; + folder?: DatasetFolder; /** - * The encrypted credential used for authentication. Credentials are encrypted using the - * integration runtime credential manager. Type: string (or Expression with resultType string). + * The table name. Type: string (or Expression with resultType string). */ - encryptedCredential?: any; + tableName?: any; } /** - * The azure table storage linked service. + * Vertica dataset. */ -export interface AzureTableStorageLinkedService { +export interface VerticaTableDataset { /** * Polymorphic Discriminator */ - type: "AzureTableStorage"; + type: "VerticaTable"; /** - * The integration runtime reference. + * Dataset description. */ - connectVia?: IntegrationRuntimeReference; + description?: string; /** - * Linked service description. + * Columns that define the structure of the dataset. Type: array (or Expression with resultType + * array), itemType: DatasetDataElement. */ - description?: string; + structure?: any; /** - * Parameters for linked service. + * Columns that define the physical type schema of the dataset. Type: array (or Expression with + * resultType array), itemType: DatasetSchemaDataElement. */ - parameters?: { [propertyName: string]: ParameterSpecification }; + schema?: any; /** - * List of tags that can be used for describing the linked service. + * Linked service reference. */ - annotations?: any[]; + linkedServiceName: LinkedServiceReference; /** - * The connection string. It is mutually exclusive with sasUri property. Type: string, - * SecureString or AzureKeyVaultSecretReference. + * Parameters for dataset. */ - connectionString?: any; + parameters?: { [propertyName: string]: ParameterSpecification }; /** - * The Azure key vault secret reference of accountKey in connection string. + * List of tags that can be used for describing the Dataset. */ - accountKey?: AzureKeyVaultSecretReference; + annotations?: any[]; /** - * SAS URI of the Azure Storage resource. It is mutually exclusive with connectionString - * property. Type: string, SecureString or AzureKeyVaultSecretReference. + * The folder that this Dataset is in. If not specified, Dataset will appear at the root level. */ - sasUri?: any; + folder?: DatasetFolder; /** - * The Azure key vault secret reference of sasToken in sas uri. + * This property will be retired. Please consider using schema + table properties instead. */ - sasToken?: AzureKeyVaultSecretReference; + tableName?: any; /** - * The encrypted credential used for authentication. Credentials are encrypted using the - * integration runtime credential manager. Type: string (or Expression with resultType string). + * The table name of the Vertica. Type: string (or Expression with resultType string). */ - encryptedCredential?: string; + table?: any; + /** + * The schema name of the Vertica. Type: string (or Expression with resultType string). + */ + verticaTableDatasetSchema?: any; } /** - * The azure blob storage linked service. + * Netezza dataset. */ -export interface AzureBlobStorageLinkedService { +export interface NetezzaTableDataset { /** * Polymorphic Discriminator */ - type: "AzureBlobStorage"; - /** - * The integration runtime reference. - */ - connectVia?: IntegrationRuntimeReference; + type: "NetezzaTable"; /** - * Linked service description. + * Dataset description. */ description?: string; /** - * Parameters for linked service. - */ - parameters?: { [propertyName: string]: ParameterSpecification }; - /** - * List of tags that can be used for describing the linked service. - */ - annotations?: any[]; - /** - * The connection string. It is mutually exclusive with sasUri, serviceEndpoint property. Type: - * string, SecureString or AzureKeyVaultSecretReference. + * Columns that define the structure of the dataset. Type: array (or Expression with resultType + * array), itemType: DatasetDataElement. */ - connectionString?: any; + structure?: any; /** - * The Azure key vault secret reference of accountKey in connection string. + * Columns that define the physical type schema of the dataset. Type: array (or Expression with + * resultType array), itemType: DatasetSchemaDataElement. */ - accountKey?: AzureKeyVaultSecretReference; + schema?: any; /** - * SAS URI of the Azure Blob Storage resource. It is mutually exclusive with connectionString, - * serviceEndpoint property. Type: string, SecureString or AzureKeyVaultSecretReference. + * Linked service reference. */ - sasUri?: any; + linkedServiceName: LinkedServiceReference; /** - * The Azure key vault secret reference of sasToken in sas uri. + * Parameters for dataset. */ - sasToken?: AzureKeyVaultSecretReference; + parameters?: { [propertyName: string]: ParameterSpecification }; /** - * Blob service endpoint of the Azure Blob Storage resource. It is mutually exclusive with - * connectionString, sasUri property. + * List of tags that can be used for describing the Dataset. */ - serviceEndpoint?: string; + annotations?: any[]; /** - * The ID of the service principal used to authenticate against Azure SQL Data Warehouse. Type: - * string (or Expression with resultType string). + * The folder that this Dataset is in. If not specified, Dataset will appear at the root level. */ - servicePrincipalId?: any; + folder?: DatasetFolder; /** - * The key of the service principal used to authenticate against Azure SQL Data Warehouse. + * This property will be retired. Please consider using schema + table properties instead. */ - servicePrincipalKey?: SecretBaseUnion; + tableName?: any; /** - * The name or ID of the tenant to which the service principal belongs. Type: string (or - * Expression with resultType string). + * The table name of the Netezza. Type: string (or Expression with resultType string). */ - tenant?: any; + table?: any; /** - * The encrypted credential used for authentication. Credentials are encrypted using the - * integration runtime credential manager. Type: string (or Expression with resultType string). + * The schema name of the Netezza. Type: string (or Expression with resultType string). */ - encryptedCredential?: string; + netezzaTableDatasetSchema?: any; } /** - * The storage account linked service. + * Zoho server dataset. */ -export interface AzureStorageLinkedService { +export interface ZohoObjectDataset { /** * Polymorphic Discriminator */ - type: "AzureStorage"; - /** - * The integration runtime reference. - */ - connectVia?: IntegrationRuntimeReference; + type: "ZohoObject"; /** - * Linked service description. + * Dataset description. */ description?: string; /** - * Parameters for linked service. + * Columns that define the structure of the dataset. Type: array (or Expression with resultType + * array), itemType: DatasetDataElement. */ - parameters?: { [propertyName: string]: ParameterSpecification }; + structure?: any; /** - * List of tags that can be used for describing the linked service. + * Columns that define the physical type schema of the dataset. Type: array (or Expression with + * resultType array), itemType: DatasetSchemaDataElement. */ - annotations?: any[]; + schema?: any; /** - * The connection string. It is mutually exclusive with sasUri property. Type: string, - * SecureString or AzureKeyVaultSecretReference. + * Linked service reference. */ - connectionString?: any; + linkedServiceName: LinkedServiceReference; /** - * The Azure key vault secret reference of accountKey in connection string. + * Parameters for dataset. */ - accountKey?: AzureKeyVaultSecretReference; + parameters?: { [propertyName: string]: ParameterSpecification }; /** - * SAS URI of the Azure Storage resource. It is mutually exclusive with connectionString - * property. Type: string, SecureString or AzureKeyVaultSecretReference. + * List of tags that can be used for describing the Dataset. */ - sasUri?: any; + annotations?: any[]; /** - * The Azure key vault secret reference of sasToken in sas uri. + * The folder that this Dataset is in. If not specified, Dataset will appear at the root level. */ - sasToken?: AzureKeyVaultSecretReference; + folder?: DatasetFolder; /** - * The encrypted credential used for authentication. Credentials are encrypted using the - * integration runtime credential manager. Type: string (or Expression with resultType string). + * The table name. Type: string (or Expression with resultType string). */ - encryptedCredential?: string; + tableName?: any; } /** - * Google AdWords service dataset. + * Xero Service dataset. */ -export interface GoogleAdWordsObjectDataset { +export interface XeroObjectDataset { /** * Polymorphic Discriminator */ - type: "GoogleAdWordsObject"; + type: "XeroObject"; /** * Dataset description. */ @@ -8175,13 +8115,13 @@ export interface GoogleAdWordsObjectDataset { } /** - * The Azure Data Explorer (Kusto) dataset. + * Square Service dataset. */ -export interface AzureDataExplorerTableDataset { +export interface SquareObjectDataset { /** * Polymorphic Discriminator */ - type: "AzureDataExplorerTable"; + type: "SquareObject"; /** * Dataset description. */ @@ -8213,20 +8153,19 @@ export interface AzureDataExplorerTableDataset { */ folder?: DatasetFolder; /** - * The table name of the Azure Data Explorer database. Type: string (or Expression with - * resultType string). + * The table name. Type: string (or Expression with resultType string). */ - table?: any; + tableName?: any; } /** - * Oracle Service Cloud dataset. + * Spark Server dataset. */ -export interface OracleServiceCloudObjectDataset { +export interface SparkObjectDataset { /** * Polymorphic Discriminator */ - type: "OracleServiceCloudObject"; + type: "SparkObject"; /** * Dataset description. */ @@ -8258,19 +8197,27 @@ export interface OracleServiceCloudObjectDataset { */ folder?: DatasetFolder; /** - * The table name. Type: string (or Expression with resultType string). + * This property will be retired. Please consider using schema + table properties instead. */ tableName?: any; + /** + * The table name of the Spark. Type: string (or Expression with resultType string). + */ + table?: any; + /** + * The schema name of the Spark. Type: string (or Expression with resultType string). + */ + sparkObjectDatasetSchema?: any; } /** - * The path of the Dynamics AX OData entity. + * Shopify Service dataset. */ -export interface DynamicsAXResourceDataset { +export interface ShopifyObjectDataset { /** * Polymorphic Discriminator */ - type: "DynamicsAXResource"; + type: "ShopifyObject"; /** * Dataset description. */ @@ -8302,19 +8249,19 @@ export interface DynamicsAXResourceDataset { */ folder?: DatasetFolder; /** - * The path of the Dynamics AX OData entity. Type: string (or Expression with resultType string). + * The table name. Type: string (or Expression with resultType string). */ - path: any; + tableName?: any; } /** - * Responsys dataset. + * ServiceNow server dataset. */ -export interface ResponsysObjectDataset { +export interface ServiceNowObjectDataset { /** * Polymorphic Discriminator */ - type: "ResponsysObject"; + type: "ServiceNowObject"; /** * Dataset description. */ @@ -8352,13 +8299,13 @@ export interface ResponsysObjectDataset { } /** - * Salesforce Marketing Cloud dataset. + * QuickBooks server dataset. */ -export interface SalesforceMarketingCloudObjectDataset { +export interface QuickBooksObjectDataset { /** * Polymorphic Discriminator */ - type: "SalesforceMarketingCloudObject"; + type: "QuickBooksObject"; /** * Dataset description. */ @@ -8396,13 +8343,13 @@ export interface SalesforceMarketingCloudObjectDataset { } /** - * Vertica dataset. + * Presto server dataset. */ -export interface VerticaTableDataset { +export interface PrestoObjectDataset { /** * Polymorphic Discriminator */ - type: "VerticaTable"; + type: "PrestoObject"; /** * Dataset description. */ @@ -8438,23 +8385,23 @@ export interface VerticaTableDataset { */ tableName?: any; /** - * The table name of the Vertica. Type: string (or Expression with resultType string). + * The table name of the Presto. Type: string (or Expression with resultType string). */ table?: any; /** - * The schema name of the Vertica. Type: string (or Expression with resultType string). + * The schema name of the Presto. Type: string (or Expression with resultType string). */ - verticaTableDatasetSchema?: any; + prestoObjectDatasetSchema?: any; } /** - * Netezza dataset. + * Phoenix server dataset. */ -export interface NetezzaTableDataset { +export interface PhoenixObjectDataset { /** * Polymorphic Discriminator */ - type: "NetezzaTable"; + type: "PhoenixObject"; /** * Dataset description. */ @@ -8490,23 +8437,23 @@ export interface NetezzaTableDataset { */ tableName?: any; /** - * The table name of the Netezza. Type: string (or Expression with resultType string). + * The table name of the Phoenix. Type: string (or Expression with resultType string). */ table?: any; /** - * The schema name of the Netezza. Type: string (or Expression with resultType string). + * The schema name of the Phoenix. Type: string (or Expression with resultType string). */ - netezzaTableDatasetSchema?: any; + phoenixObjectDatasetSchema?: any; } /** - * Zoho server dataset. + * Paypal Service dataset. */ -export interface ZohoObjectDataset { +export interface PaypalObjectDataset { /** * Polymorphic Discriminator */ - type: "ZohoObject"; + type: "PaypalObject"; /** * Dataset description. */ @@ -8544,13 +8491,13 @@ export interface ZohoObjectDataset { } /** - * Xero Service dataset. + * Marketo server dataset. */ -export interface XeroObjectDataset { +export interface MarketoObjectDataset { /** * Polymorphic Discriminator */ - type: "XeroObject"; + type: "MarketoObject"; /** * Dataset description. */ @@ -8588,13 +8535,13 @@ export interface XeroObjectDataset { } /** - * Square Service dataset. + * Azure Database for MariaDB dataset. */ -export interface SquareObjectDataset { +export interface AzureMariaDBTableDataset { /** * Polymorphic Discriminator */ - type: "SquareObject"; + type: "AzureMariaDBTable"; /** * Dataset description. */ @@ -8632,13 +8579,13 @@ export interface SquareObjectDataset { } /** - * Spark Server dataset. + * MariaDB server dataset. */ -export interface SparkObjectDataset { +export interface MariaDBTableDataset { /** * Polymorphic Discriminator */ - type: "SparkObject"; + type: "MariaDBTable"; /** * Dataset description. */ @@ -8670,27 +8617,19 @@ export interface SparkObjectDataset { */ folder?: DatasetFolder; /** - * This property will be retired. Please consider using schema + table properties instead. + * The table name. Type: string (or Expression with resultType string). */ tableName?: any; - /** - * The table name of the Spark. Type: string (or Expression with resultType string). - */ - table?: any; - /** - * The schema name of the Spark. Type: string (or Expression with resultType string). - */ - sparkObjectDatasetSchema?: any; } /** - * Shopify Service dataset. + * Magento server dataset. */ -export interface ShopifyObjectDataset { +export interface MagentoObjectDataset { /** * Polymorphic Discriminator */ - type: "ShopifyObject"; + type: "MagentoObject"; /** * Dataset description. */ @@ -8728,13 +8667,13 @@ export interface ShopifyObjectDataset { } /** - * ServiceNow server dataset. + * Jira Service dataset. */ -export interface ServiceNowObjectDataset { +export interface JiraObjectDataset { /** * Polymorphic Discriminator */ - type: "ServiceNowObject"; + type: "JiraObject"; /** * Dataset description. */ @@ -8772,13 +8711,13 @@ export interface ServiceNowObjectDataset { } /** - * QuickBooks server dataset. + * Impala server dataset. */ -export interface QuickBooksObjectDataset { +export interface ImpalaObjectDataset { /** * Polymorphic Discriminator */ - type: "QuickBooksObject"; + type: "ImpalaObject"; /** * Dataset description. */ @@ -8810,19 +8749,27 @@ export interface QuickBooksObjectDataset { */ folder?: DatasetFolder; /** - * The table name. Type: string (or Expression with resultType string). + * This property will be retired. Please consider using schema + table properties instead. */ tableName?: any; + /** + * The table name of the Impala. Type: string (or Expression with resultType string). + */ + table?: any; + /** + * The schema name of the Impala. Type: string (or Expression with resultType string). + */ + impalaObjectDatasetSchema?: any; } /** - * Presto server dataset. + * Hubspot Service dataset. */ -export interface PrestoObjectDataset { +export interface HubspotObjectDataset { /** * Polymorphic Discriminator */ - type: "PrestoObject"; + type: "HubspotObject"; /** * Dataset description. */ @@ -8854,27 +8801,19 @@ export interface PrestoObjectDataset { */ folder?: DatasetFolder; /** - * This property will be retired. Please consider using schema + table properties instead. + * The table name. Type: string (or Expression with resultType string). */ tableName?: any; - /** - * The table name of the Presto. Type: string (or Expression with resultType string). - */ - table?: any; - /** - * The schema name of the Presto. Type: string (or Expression with resultType string). - */ - prestoObjectDatasetSchema?: any; } /** - * Phoenix server dataset. + * Hive Server dataset. */ -export interface PhoenixObjectDataset { +export interface HiveObjectDataset { /** * Polymorphic Discriminator */ - type: "PhoenixObject"; + type: "HiveObject"; /** * Dataset description. */ @@ -8910,23 +8849,23 @@ export interface PhoenixObjectDataset { */ tableName?: any; /** - * The table name of the Phoenix. Type: string (or Expression with resultType string). + * The table name of the Hive. Type: string (or Expression with resultType string). */ table?: any; /** - * The schema name of the Phoenix. Type: string (or Expression with resultType string). + * The schema name of the Hive. Type: string (or Expression with resultType string). */ - phoenixObjectDatasetSchema?: any; + hiveObjectDatasetSchema?: any; } /** - * Paypal Service dataset. + * HBase server dataset. */ -export interface PaypalObjectDataset { +export interface HBaseObjectDataset { /** * Polymorphic Discriminator */ - type: "PaypalObject"; + type: "HBaseObject"; /** * Dataset description. */ @@ -8964,13 +8903,13 @@ export interface PaypalObjectDataset { } /** - * Marketo server dataset. + * Greenplum Database dataset. */ -export interface MarketoObjectDataset { +export interface GreenplumTableDataset { /** * Polymorphic Discriminator */ - type: "MarketoObject"; + type: "GreenplumTable"; /** * Dataset description. */ @@ -9002,19 +8941,27 @@ export interface MarketoObjectDataset { */ folder?: DatasetFolder; /** - * The table name. Type: string (or Expression with resultType string). + * This property will be retired. Please consider using schema + table properties instead. */ tableName?: any; + /** + * The table name of Greenplum. Type: string (or Expression with resultType string). + */ + table?: any; + /** + * The schema name of Greenplum. Type: string (or Expression with resultType string). + */ + greenplumTableDatasetSchema?: any; } /** - * Azure Database for MariaDB dataset. + * Google BigQuery service dataset. */ -export interface AzureMariaDBTableDataset { +export interface GoogleBigQueryObjectDataset { /** * Polymorphic Discriminator */ - type: "AzureMariaDBTable"; + type: "GoogleBigQueryObject"; /** * Dataset description. */ @@ -9046,19 +8993,27 @@ export interface AzureMariaDBTableDataset { */ folder?: DatasetFolder; /** - * The table name. Type: string (or Expression with resultType string). + * This property will be retired. Please consider using database + table properties instead. */ tableName?: any; + /** + * The table name of the Google BigQuery. Type: string (or Expression with resultType string). + */ + table?: any; + /** + * The database name of the Google BigQuery. Type: string (or Expression with resultType string). + */ + dataset?: any; } /** - * MariaDB server dataset. + * Eloqua server dataset. */ -export interface MariaDBTableDataset { +export interface EloquaObjectDataset { /** * Polymorphic Discriminator */ - type: "MariaDBTable"; + type: "EloquaObject"; /** * Dataset description. */ @@ -9096,13 +9051,13 @@ export interface MariaDBTableDataset { } /** - * Magento server dataset. + * Drill server dataset. */ -export interface MagentoObjectDataset { +export interface DrillTableDataset { /** * Polymorphic Discriminator */ - type: "MagentoObject"; + type: "DrillTable"; /** * Dataset description. */ @@ -9134,19 +9089,27 @@ export interface MagentoObjectDataset { */ folder?: DatasetFolder; /** - * The table name. Type: string (or Expression with resultType string). + * This property will be retired. Please consider using schema + table properties instead. */ tableName?: any; + /** + * The table name of the Drill. Type: string (or Expression with resultType string). + */ + table?: any; + /** + * The schema name of the Drill. Type: string (or Expression with resultType string). + */ + drillTableDatasetSchema?: any; } /** - * Jira Service dataset. + * Couchbase server dataset. */ -export interface JiraObjectDataset { +export interface CouchbaseTableDataset { /** * Polymorphic Discriminator */ - type: "JiraObject"; + type: "CouchbaseTable"; /** * Dataset description. */ @@ -9184,13 +9147,13 @@ export interface JiraObjectDataset { } /** - * Impala server dataset. + * Concur Service dataset. */ -export interface ImpalaObjectDataset { +export interface ConcurObjectDataset { /** * Polymorphic Discriminator */ - type: "ImpalaObject"; + type: "ConcurObject"; /** * Dataset description. */ @@ -9222,27 +9185,19 @@ export interface ImpalaObjectDataset { */ folder?: DatasetFolder; /** - * This property will be retired. Please consider using schema + table properties instead. + * The table name. Type: string (or Expression with resultType string). */ tableName?: any; - /** - * The table name of the Impala. Type: string (or Expression with resultType string). - */ - table?: any; - /** - * The schema name of the Impala. Type: string (or Expression with resultType string). - */ - impalaObjectDatasetSchema?: any; } /** - * Hubspot Service dataset. + * Azure PostgreSQL dataset. */ -export interface HubspotObjectDataset { +export interface AzurePostgreSqlTableDataset { /** * Polymorphic Discriminator */ - type: "HubspotObject"; + type: "AzurePostgreSqlTable"; /** * Dataset description. */ @@ -9274,19 +9229,30 @@ export interface HubspotObjectDataset { */ folder?: DatasetFolder; /** - * The table name. Type: string (or Expression with resultType string). + * The table name of the Azure PostgreSQL database which includes both schema and table. Type: + * string (or Expression with resultType string). */ tableName?: any; + /** + * The table name of the Azure PostgreSQL database. Type: string (or Expression with resultType + * string). + */ + table?: any; + /** + * The schema name of the Azure PostgreSQL database. Type: string (or Expression with resultType + * string). + */ + azurePostgreSqlTableDatasetSchema?: any; } /** - * Hive Server dataset. + * Amazon Marketplace Web Service dataset. */ -export interface HiveObjectDataset { +export interface AmazonMWSObjectDataset { /** * Polymorphic Discriminator */ - type: "HiveObject"; + type: "AmazonMWSObject"; /** * Dataset description. */ @@ -9318,219 +9284,282 @@ export interface HiveObjectDataset { */ folder?: DatasetFolder; /** - * This property will be retired. Please consider using schema + table properties instead. + * The table name. Type: string (or Expression with resultType string). */ tableName?: any; - /** - * The table name of the Hive. Type: string (or Expression with resultType string). - */ - table?: any; - /** - * The schema name of the Hive. Type: string (or Expression with resultType string). - */ - hiveObjectDatasetSchema?: any; } /** - * HBase server dataset. + * Contains the possible cases for DatasetCompression. */ -export interface HBaseObjectDataset { +export type DatasetCompressionUnion = DatasetCompression | DatasetZipDeflateCompression | DatasetDeflateCompression | DatasetGZipCompression | DatasetBZip2Compression; + +/** + * The compression method used on a dataset. + */ +export interface DatasetCompression { /** * Polymorphic Discriminator */ - type: "HBaseObject"; - /** - * Dataset description. - */ - description?: string; + type: "DatasetCompression"; /** - * Columns that define the structure of the dataset. Type: array (or Expression with resultType - * array), itemType: DatasetDataElement. + * Describes unknown properties. The value of an unknown property can be of "any" type. */ - structure?: any; - /** - * Columns that define the physical type schema of the dataset. Type: array (or Expression with - * resultType array), itemType: DatasetSchemaDataElement. + [property: string]: any; +} + +/** + * The ZipDeflate compression method used on a dataset. + */ +export interface DatasetZipDeflateCompression { + /** + * Polymorphic Discriminator */ - schema?: any; + type: "ZipDeflate"; /** - * Linked service reference. + * The ZipDeflate compression level. */ - linkedServiceName: LinkedServiceReference; + level?: any; +} + +/** + * The Deflate compression method used on a dataset. + */ +export interface DatasetDeflateCompression { /** - * Parameters for dataset. + * Polymorphic Discriminator */ - parameters?: { [propertyName: string]: ParameterSpecification }; + type: "Deflate"; /** - * List of tags that can be used for describing the Dataset. + * The Deflate compression level. */ - annotations?: any[]; + level?: any; +} + +/** + * The GZip compression method used on a dataset. + */ +export interface DatasetGZipCompression { /** - * The folder that this Dataset is in. If not specified, Dataset will appear at the root level. + * Polymorphic Discriminator */ - folder?: DatasetFolder; + type: "GZip"; /** - * The table name. Type: string (or Expression with resultType string). + * The GZip compression level. */ - tableName?: any; + level?: any; } /** - * Greenplum Database dataset. + * The BZip2 compression method used on a dataset. */ -export interface GreenplumTableDataset { +export interface DatasetBZip2Compression { /** * Polymorphic Discriminator */ - type: "GreenplumTable"; + type: "BZip2"; +} + +/** + * Contains the possible cases for DatasetStorageFormat. + */ +export type DatasetStorageFormatUnion = DatasetStorageFormat | ParquetFormat | OrcFormat | AvroFormat | JsonFormat | TextFormat; + +/** + * The format definition of a storage. + */ +export interface DatasetStorageFormat { /** - * Dataset description. + * Polymorphic Discriminator */ - description?: string; + type: "DatasetStorageFormat"; /** - * Columns that define the structure of the dataset. Type: array (or Expression with resultType - * array), itemType: DatasetDataElement. + * Serializer. Type: string (or Expression with resultType string). */ - structure?: any; + serializer?: any; /** - * Columns that define the physical type schema of the dataset. Type: array (or Expression with - * resultType array), itemType: DatasetSchemaDataElement. + * Deserializer. Type: string (or Expression with resultType string). */ - schema?: any; + deserializer?: any; /** - * Linked service reference. + * Describes unknown properties. The value of an unknown property can be of "any" type. */ - linkedServiceName: LinkedServiceReference; + [property: string]: any; +} + +/** + * The data stored in Parquet format. + */ +export interface ParquetFormat { /** - * Parameters for dataset. + * Polymorphic Discriminator */ - parameters?: { [propertyName: string]: ParameterSpecification }; + type: "ParquetFormat"; /** - * List of tags that can be used for describing the Dataset. + * Serializer. Type: string (or Expression with resultType string). */ - annotations?: any[]; + serializer?: any; /** - * The folder that this Dataset is in. If not specified, Dataset will appear at the root level. + * Deserializer. Type: string (or Expression with resultType string). */ - folder?: DatasetFolder; + deserializer?: any; +} + +/** + * The data stored in Optimized Row Columnar (ORC) format. + */ +export interface OrcFormat { /** - * This property will be retired. Please consider using schema + table properties instead. + * Polymorphic Discriminator */ - tableName?: any; + type: "OrcFormat"; /** - * The table name of Greenplum. Type: string (or Expression with resultType string). + * Serializer. Type: string (or Expression with resultType string). */ - table?: any; + serializer?: any; /** - * The schema name of Greenplum. Type: string (or Expression with resultType string). + * Deserializer. Type: string (or Expression with resultType string). */ - greenplumTableDatasetSchema?: any; + deserializer?: any; } /** - * Google BigQuery service dataset. + * The data stored in Avro format. */ -export interface GoogleBigQueryObjectDataset { +export interface AvroFormat { /** * Polymorphic Discriminator */ - type: "GoogleBigQueryObject"; + type: "AvroFormat"; /** - * Dataset description. + * Serializer. Type: string (or Expression with resultType string). */ - description?: string; + serializer?: any; /** - * Columns that define the structure of the dataset. Type: array (or Expression with resultType - * array), itemType: DatasetDataElement. + * Deserializer. Type: string (or Expression with resultType string). */ - structure?: any; + deserializer?: any; +} + +/** + * The data stored in JSON format. + */ +export interface JsonFormat { /** - * Columns that define the physical type schema of the dataset. Type: array (or Expression with - * resultType array), itemType: DatasetSchemaDataElement. + * Polymorphic Discriminator */ - schema?: any; + type: "JsonFormat"; /** - * Linked service reference. + * Serializer. Type: string (or Expression with resultType string). */ - linkedServiceName: LinkedServiceReference; + serializer?: any; /** - * Parameters for dataset. + * Deserializer. Type: string (or Expression with resultType string). */ - parameters?: { [propertyName: string]: ParameterSpecification }; + deserializer?: any; /** - * List of tags that can be used for describing the Dataset. + * File pattern of JSON. To be more specific, the way of separating a collection of JSON objects. + * The default value is 'setOfObjects'. It is case-sensitive. */ - annotations?: any[]; + filePattern?: any; /** - * The folder that this Dataset is in. If not specified, Dataset will appear at the root level. + * The character used to separate nesting levels. Default value is '.' (dot). Type: string (or + * Expression with resultType string). */ - folder?: DatasetFolder; + nestingSeparator?: any; /** - * This property will be retired. Please consider using database + table properties instead. + * The code page name of the preferred encoding. If not provided, the default value is 'utf-8', + * unless the byte order mark (BOM) denotes another Unicode encoding. The full list of supported + * values can be found in the 'Name' column of the table of encodings in the following reference: + * https://go.microsoft.com/fwlink/?linkid=861078. Type: string (or Expression with resultType + * string). */ - tableName?: any; + encodingName?: any; /** - * The table name of the Google BigQuery. Type: string (or Expression with resultType string). + * The JSONPath of the JSON array element to be flattened. Example: "$.ArrayPath". Type: string + * (or Expression with resultType string). */ - table?: any; + jsonNodeReference?: any; /** - * The database name of the Google BigQuery. Type: string (or Expression with resultType string). + * The JSONPath definition for each column mapping with a customized column name to extract data + * from JSON file. For fields under root object, start with "$"; for fields inside the array + * chosen by jsonNodeReference property, start from the array element. Example: {"Column1": + * "$.Column1Path", "Column2": "Column2PathInArray"}. Type: object (or Expression with resultType + * object). */ - dataset?: any; + jsonPathDefinition?: any; } /** - * Eloqua server dataset. + * The data stored in text format. */ -export interface EloquaObjectDataset { +export interface TextFormat { /** * Polymorphic Discriminator */ - type: "EloquaObject"; + type: "TextFormat"; /** - * Dataset description. + * Serializer. Type: string (or Expression with resultType string). */ - description?: string; + serializer?: any; /** - * Columns that define the structure of the dataset. Type: array (or Expression with resultType - * array), itemType: DatasetDataElement. + * Deserializer. Type: string (or Expression with resultType string). */ - structure?: any; + deserializer?: any; /** - * Columns that define the physical type schema of the dataset. Type: array (or Expression with - * resultType array), itemType: DatasetSchemaDataElement. + * The column delimiter. Type: string (or Expression with resultType string). */ - schema?: any; + columnDelimiter?: any; /** - * Linked service reference. + * The row delimiter. Type: string (or Expression with resultType string). */ - linkedServiceName: LinkedServiceReference; + rowDelimiter?: any; /** - * Parameters for dataset. + * The escape character. Type: string (or Expression with resultType string). */ - parameters?: { [propertyName: string]: ParameterSpecification }; + escapeChar?: any; /** - * List of tags that can be used for describing the Dataset. + * The quote character. Type: string (or Expression with resultType string). */ - annotations?: any[]; + quoteChar?: any; /** - * The folder that this Dataset is in. If not specified, Dataset will appear at the root level. + * The null value string. Type: string (or Expression with resultType string). */ - folder?: DatasetFolder; + nullValue?: any; /** - * The table name. Type: string (or Expression with resultType string). + * The code page name of the preferred encoding. If miss, the default value is ΓÇ£utf-8ΓÇ¥, + * unless BOM denotes another Unicode encoding. Refer to the ΓÇ£NameΓÇ¥ column of the table in + * the following link to set supported values: + * https://msdn.microsoft.com/library/system.text.encoding.aspx. Type: string (or Expression with + * resultType string). */ - tableName?: any; + encodingName?: any; + /** + * Treat empty column values in the text file as null. The default value is true. Type: boolean + * (or Expression with resultType boolean). + */ + treatEmptyAsNull?: any; + /** + * The number of lines/rows to be skipped when parsing text files. The default value is 0. Type: + * integer (or Expression with resultType integer). + */ + skipLineCount?: any; + /** + * When used as input, treat the first row of data as headers. When used as output,write the + * headers into the output as the first row of data. The default value is false. Type: boolean + * (or Expression with resultType boolean). + */ + firstRowAsHeader?: any; } /** - * Drill server dataset. + * A file in an HTTP web server. */ -export interface DrillTableDataset { +export interface HttpDataset { /** * Polymorphic Discriminator */ - type: "DrillTable"; + type: "HttpFile"; /** * Dataset description. */ @@ -9562,27 +9591,43 @@ export interface DrillTableDataset { */ folder?: DatasetFolder; /** - * This property will be retired. Please consider using schema + table properties instead. + * The relative URL based on the URL in the HttpLinkedService refers to an HTTP file Type: string + * (or Expression with resultType string). + */ + relativeUrl?: any; + /** + * The HTTP method for the HTTP request. Type: string (or Expression with resultType string). + */ + requestMethod?: any; + /** + * The body for the HTTP request. Type: string (or Expression with resultType string). + */ + requestBody?: any; + /** + * The headers for the HTTP Request. e.g. request-header-name-1:request-header-value-1 + * ... + * request-header-name-n:request-header-value-n Type: string (or Expression with resultType + * string). */ - tableName?: any; + additionalHeaders?: any; /** - * The table name of the Drill. Type: string (or Expression with resultType string). + * The format of files. */ - table?: any; + format?: DatasetStorageFormatUnion; /** - * The schema name of the Drill. Type: string (or Expression with resultType string). + * The data compression method used on files. */ - drillTableDatasetSchema?: any; + compression?: DatasetCompressionUnion; } /** - * Couchbase server dataset. + * The Azure Search Index. */ -export interface CouchbaseTableDataset { +export interface AzureSearchIndexDataset { /** * Polymorphic Discriminator */ - type: "CouchbaseTable"; + type: "AzureSearchIndex"; /** * Dataset description. */ @@ -9614,19 +9659,19 @@ export interface CouchbaseTableDataset { */ folder?: DatasetFolder; /** - * The table name. Type: string (or Expression with resultType string). + * The name of the Azure Search Index. Type: string (or Expression with resultType string). */ - tableName?: any; + indexName: any; } /** - * Concur Service dataset. + * The dataset points to a HTML table in the web page. */ -export interface ConcurObjectDataset { +export interface WebTableDataset { /** * Polymorphic Discriminator */ - type: "ConcurObject"; + type: "WebTable"; /** * Dataset description. */ @@ -9658,19 +9703,25 @@ export interface ConcurObjectDataset { */ folder?: DatasetFolder; /** - * The table name. Type: string (or Expression with resultType string). + * The zero-based index of the table in the web page. Type: integer (or Expression with + * resultType integer), minimum: 0. */ - tableName?: any; + index: any; + /** + * The relative URL to the web page from the linked service URL. Type: string (or Expression with + * resultType string). + */ + path?: any; } /** - * Azure PostgreSQL dataset. + * SAP Table Resource properties. */ -export interface AzurePostgreSqlTableDataset { +export interface SapTableResourceDataset { /** * Polymorphic Discriminator */ - type: "AzurePostgreSqlTable"; + type: "SapTableResource"; /** * Dataset description. */ @@ -9702,30 +9753,19 @@ export interface AzurePostgreSqlTableDataset { */ folder?: DatasetFolder; /** - * The table name of the Azure PostgreSQL database which includes both schema and table. Type: - * string (or Expression with resultType string). - */ - tableName?: any; - /** - * The table name of the Azure PostgreSQL database. Type: string (or Expression with resultType - * string). - */ - table?: any; - /** - * The schema name of the Azure PostgreSQL database. Type: string (or Expression with resultType - * string). + * The name of the SAP Table. Type: string (or Expression with resultType string). */ - azurePostgreSqlTableDatasetSchema?: any; + tableName: any; } /** - * Amazon Marketplace Web Service dataset. + * A Rest service dataset. */ -export interface AmazonMWSObjectDataset { +export interface RestResourceDataset { /** * Polymorphic Discriminator */ - type: "AmazonMWSObject"; + type: "RestResource"; /** * Dataset description. */ @@ -9757,282 +9797,149 @@ export interface AmazonMWSObjectDataset { */ folder?: DatasetFolder; /** - * The table name. Type: string (or Expression with resultType string). - */ - tableName?: any; -} - -/** - * Contains the possible cases for DatasetCompression. - */ -export type DatasetCompressionUnion = DatasetCompression | DatasetZipDeflateCompression | DatasetDeflateCompression | DatasetGZipCompression | DatasetBZip2Compression; - -/** - * The compression method used on a dataset. - */ -export interface DatasetCompression { - /** - * Polymorphic Discriminator - */ - type: "DatasetCompression"; - /** - * Describes unknown properties. The value of an unknown property can be of "any" type. - */ - [property: string]: any; -} - -/** - * The ZipDeflate compression method used on a dataset. - */ -export interface DatasetZipDeflateCompression { - /** - * Polymorphic Discriminator - */ - type: "ZipDeflate"; - /** - * The ZipDeflate compression level. - */ - level?: any; -} - -/** - * The Deflate compression method used on a dataset. - */ -export interface DatasetDeflateCompression { - /** - * Polymorphic Discriminator - */ - type: "Deflate"; - /** - * The Deflate compression level. - */ - level?: any; -} - -/** - * The GZip compression method used on a dataset. - */ -export interface DatasetGZipCompression { - /** - * Polymorphic Discriminator - */ - type: "GZip"; - /** - * The GZip compression level. - */ - level?: any; -} - -/** - * The BZip2 compression method used on a dataset. - */ -export interface DatasetBZip2Compression { - /** - * Polymorphic Discriminator - */ - type: "BZip2"; -} - -/** - * Contains the possible cases for DatasetStorageFormat. - */ -export type DatasetStorageFormatUnion = DatasetStorageFormat | ParquetFormat | OrcFormat | AvroFormat | JsonFormat | TextFormat; - -/** - * The format definition of a storage. - */ -export interface DatasetStorageFormat { - /** - * Polymorphic Discriminator - */ - type: "DatasetStorageFormat"; - /** - * Serializer. Type: string (or Expression with resultType string). - */ - serializer?: any; - /** - * Deserializer. Type: string (or Expression with resultType string). - */ - deserializer?: any; - /** - * Describes unknown properties. The value of an unknown property can be of "any" type. - */ - [property: string]: any; -} - -/** - * The data stored in Parquet format. - */ -export interface ParquetFormat { - /** - * Polymorphic Discriminator - */ - type: "ParquetFormat"; - /** - * Serializer. Type: string (or Expression with resultType string). + * The relative URL to the resource that the RESTful API provides. Type: string (or Expression + * with resultType string). */ - serializer?: any; + relativeUrl?: any; /** - * Deserializer. Type: string (or Expression with resultType string). + * The HTTP method used to call the RESTful API. The default is GET. Type: string (or Expression + * with resultType string). */ - deserializer?: any; -} - -/** - * The data stored in Optimized Row Columnar (ORC) format. - */ -export interface OrcFormat { + requestMethod?: any; /** - * Polymorphic Discriminator + * The HTTP request body to the RESTful API if requestMethod is POST. Type: string (or Expression + * with resultType string). */ - type: "OrcFormat"; + requestBody?: any; /** - * Serializer. Type: string (or Expression with resultType string). + * The additional HTTP headers in the request to the RESTful API. Type: string (or Expression + * with resultType string). */ - serializer?: any; + additionalHeaders?: any; /** - * Deserializer. Type: string (or Expression with resultType string). + * The pagination rules to compose next page requests. Type: string (or Expression with + * resultType string). */ - deserializer?: any; + paginationRules?: any; } /** - * The data stored in Avro format. + * The on-premises SQL Server dataset. */ -export interface AvroFormat { +export interface SqlServerTableDataset { /** * Polymorphic Discriminator */ - type: "AvroFormat"; + type: "SqlServerTable"; /** - * Serializer. Type: string (or Expression with resultType string). + * Dataset description. */ - serializer?: any; + description?: string; /** - * Deserializer. Type: string (or Expression with resultType string). + * Columns that define the structure of the dataset. Type: array (or Expression with resultType + * array), itemType: DatasetDataElement. */ - deserializer?: any; -} - -/** - * The data stored in JSON format. - */ -export interface JsonFormat { + structure?: any; /** - * Polymorphic Discriminator + * Columns that define the physical type schema of the dataset. Type: array (or Expression with + * resultType array), itemType: DatasetSchemaDataElement. */ - type: "JsonFormat"; + schema?: any; /** - * Serializer. Type: string (or Expression with resultType string). + * Linked service reference. */ - serializer?: any; + linkedServiceName: LinkedServiceReference; /** - * Deserializer. Type: string (or Expression with resultType string). + * Parameters for dataset. */ - deserializer?: any; + parameters?: { [propertyName: string]: ParameterSpecification }; /** - * File pattern of JSON. To be more specific, the way of separating a collection of JSON objects. - * The default value is 'setOfObjects'. It is case-sensitive. + * List of tags that can be used for describing the Dataset. */ - filePattern?: any; + annotations?: any[]; /** - * The character used to separate nesting levels. Default value is '.' (dot). Type: string (or - * Expression with resultType string). + * The folder that this Dataset is in. If not specified, Dataset will appear at the root level. */ - nestingSeparator?: any; - /** - * The code page name of the preferred encoding. If not provided, the default value is 'utf-8', - * unless the byte order mark (BOM) denotes another Unicode encoding. The full list of supported - * values can be found in the 'Name' column of the table of encodings in the following reference: - * https://go.microsoft.com/fwlink/?linkid=861078. Type: string (or Expression with resultType - * string). + folder?: DatasetFolder; + /** + * This property will be retired. Please consider using schema + table properties instead. */ - encodingName?: any; + tableName?: any; /** - * The JSONPath of the JSON array element to be flattened. Example: "$.ArrayPath". Type: string - * (or Expression with resultType string). + * The schema name of the SQL Server dataset. Type: string (or Expression with resultType + * string). */ - jsonNodeReference?: any; + sqlServerTableDatasetSchema?: any; /** - * The JSONPath definition for each column mapping with a customized column name to extract data - * from JSON file. For fields under root object, start with "$"; for fields inside the array - * chosen by jsonNodeReference property, start from the array element. Example: {"Column1": - * "$.Column1Path", "Column2": "Column2PathInArray"}. Type: object (or Expression with resultType - * object). + * The table name of the SQL Server dataset. Type: string (or Expression with resultType string). */ - jsonPathDefinition?: any; + table?: any; } /** - * The data stored in text format. + * Sap Business Warehouse Open Hub Destination Table properties. */ -export interface TextFormat { +export interface SapOpenHubTableDataset { /** * Polymorphic Discriminator */ - type: "TextFormat"; + type: "SapOpenHubTable"; /** - * Serializer. Type: string (or Expression with resultType string). + * Dataset description. */ - serializer?: any; + description?: string; /** - * Deserializer. Type: string (or Expression with resultType string). + * Columns that define the structure of the dataset. Type: array (or Expression with resultType + * array), itemType: DatasetDataElement. */ - deserializer?: any; + structure?: any; /** - * The column delimiter. Type: string (or Expression with resultType string). + * Columns that define the physical type schema of the dataset. Type: array (or Expression with + * resultType array), itemType: DatasetSchemaDataElement. */ - columnDelimiter?: any; + schema?: any; /** - * The row delimiter. Type: string (or Expression with resultType string). + * Linked service reference. */ - rowDelimiter?: any; + linkedServiceName: LinkedServiceReference; /** - * The escape character. Type: string (or Expression with resultType string). + * Parameters for dataset. */ - escapeChar?: any; + parameters?: { [propertyName: string]: ParameterSpecification }; /** - * The quote character. Type: string (or Expression with resultType string). + * List of tags that can be used for describing the Dataset. */ - quoteChar?: any; + annotations?: any[]; /** - * The null value string. Type: string (or Expression with resultType string). + * The folder that this Dataset is in. If not specified, Dataset will appear at the root level. */ - nullValue?: any; + folder?: DatasetFolder; /** - * The code page name of the preferred encoding. If miss, the default value is ΓÇ£utf-8ΓÇ¥, - * unless BOM denotes another Unicode encoding. Refer to the ΓÇ£NameΓÇ¥ column of the table in - * the following link to set supported values: - * https://msdn.microsoft.com/library/system.text.encoding.aspx. Type: string (or Expression with - * resultType string). + * The name of the Open Hub Destination with destination type as Database Table. Type: string (or + * Expression with resultType string). */ - encodingName?: any; + openHubDestinationName: any; /** - * Treat empty column values in the text file as null. The default value is true. Type: boolean + * Whether to exclude the records of the last request. The default value is true. Type: boolean * (or Expression with resultType boolean). */ - treatEmptyAsNull?: any; - /** - * The number of lines/rows to be skipped when parsing text files. The default value is 0. Type: - * integer (or Expression with resultType integer). - */ - skipLineCount?: any; + excludeLastRequest?: any; /** - * When used as input, treat the first row of data as headers. When used as output,write the - * headers into the output as the first row of data. The default value is false. Type: boolean - * (or Expression with resultType boolean). + * The ID of request for delta loading. Once it is set, only data with requestId larger than the + * value of this property will be retrieved. The default value is 0. Type: integer (or Expression + * with resultType integer ). */ - firstRowAsHeader?: any; + baseRequestId?: any; } /** - * A file in an HTTP web server. + * SAP HANA Table properties. */ -export interface HttpDataset { +export interface SapHanaTableDataset { /** * Polymorphic Discriminator */ - type: "HttpFile"; + type: "SapHanaTable"; /** * Dataset description. */ @@ -10064,43 +9971,23 @@ export interface HttpDataset { */ folder?: DatasetFolder; /** - * The relative URL based on the URL in the HttpLinkedService refers to an HTTP file Type: string - * (or Expression with resultType string). - */ - relativeUrl?: any; - /** - * The HTTP method for the HTTP request. Type: string (or Expression with resultType string). - */ - requestMethod?: any; - /** - * The body for the HTTP request. Type: string (or Expression with resultType string). - */ - requestBody?: any; - /** - * The headers for the HTTP Request. e.g. request-header-name-1:request-header-value-1 - * ... - * request-header-name-n:request-header-value-n Type: string (or Expression with resultType - * string). - */ - additionalHeaders?: any; - /** - * The format of files. + * The schema name of SAP HANA. Type: string (or Expression with resultType string). */ - format?: DatasetStorageFormatUnion; + sapHanaTableDatasetSchema?: any; /** - * The data compression method used on files. + * The table name of SAP HANA. Type: string (or Expression with resultType string). */ - compression?: DatasetCompressionUnion; + table?: any; } /** - * The Azure Search Index. + * The path of the SAP ECC OData entity. */ -export interface AzureSearchIndexDataset { +export interface SapEccResourceDataset { /** * Polymorphic Discriminator */ - type: "AzureSearchIndex"; + type: "SapEccResource"; /** * Dataset description. */ @@ -10132,19 +10019,19 @@ export interface AzureSearchIndexDataset { */ folder?: DatasetFolder; /** - * The name of the Azure Search Index. Type: string (or Expression with resultType string). + * The path of the SAP ECC OData entity. Type: string (or Expression with resultType string). */ - indexName: any; + path: any; } /** - * The dataset points to a HTML table in the web page. + * The path of the SAP Cloud for Customer OData entity. */ -export interface WebTableDataset { +export interface SapCloudForCustomerResourceDataset { /** * Polymorphic Discriminator */ - type: "WebTable"; + type: "SapCloudForCustomerResource"; /** * Dataset description. */ @@ -10176,25 +10063,20 @@ export interface WebTableDataset { */ folder?: DatasetFolder; /** - * The zero-based index of the table in the web page. Type: integer (or Expression with - * resultType integer), minimum: 0. - */ - index: any; - /** - * The relative URL to the web page from the linked service URL. Type: string (or Expression with + * The path of the SAP Cloud for Customer OData entity. Type: string (or Expression with * resultType string). */ - path?: any; + path: any; } /** - * SAP Table Resource properties. + * The SAP BW cube dataset. */ -export interface SapTableResourceDataset { +export interface SapBwCubeDataset { /** * Polymorphic Discriminator */ - type: "SapTableResource"; + type: "SapBwCube"; /** * Dataset description. */ @@ -10225,20 +10107,16 @@ export interface SapTableResourceDataset { * The folder that this Dataset is in. If not specified, Dataset will appear at the root level. */ folder?: DatasetFolder; - /** - * The name of the SAP Table. Type: string (or Expression with resultType string). - */ - tableName: any; } /** - * A Rest service dataset. + * The Sybase table dataset. */ -export interface RestResourceDataset { +export interface SybaseTableDataset { /** * Polymorphic Discriminator */ - type: "RestResource"; + type: "SybaseTable"; /** * Dataset description. */ @@ -10270,40 +10148,64 @@ export interface RestResourceDataset { */ folder?: DatasetFolder; /** - * The relative URL to the resource that the RESTful API provides. Type: string (or Expression - * with resultType string). + * The Sybase table name. Type: string (or Expression with resultType string). */ - relativeUrl?: any; + tableName?: any; +} + +/** + * The Salesforce Service Cloud object dataset. + */ +export interface SalesforceServiceCloudObjectDataset { /** - * The HTTP method used to call the RESTful API. The default is GET. Type: string (or Expression - * with resultType string). + * Polymorphic Discriminator */ - requestMethod?: any; + type: "SalesforceServiceCloudObject"; /** - * The HTTP request body to the RESTful API if requestMethod is POST. Type: string (or Expression - * with resultType string). + * Dataset description. */ - requestBody?: any; + description?: string; /** - * The additional HTTP headers in the request to the RESTful API. Type: string (or Expression - * with resultType string). + * Columns that define the structure of the dataset. Type: array (or Expression with resultType + * array), itemType: DatasetDataElement. */ - additionalHeaders?: any; + structure?: any; /** - * The pagination rules to compose next page requests. Type: string (or Expression with - * resultType string). + * Columns that define the physical type schema of the dataset. Type: array (or Expression with + * resultType array), itemType: DatasetSchemaDataElement. */ - paginationRules?: any; + schema?: any; + /** + * Linked service reference. + */ + linkedServiceName: LinkedServiceReference; + /** + * Parameters for dataset. + */ + parameters?: { [propertyName: string]: ParameterSpecification }; + /** + * List of tags that can be used for describing the Dataset. + */ + annotations?: any[]; + /** + * The folder that this Dataset is in. If not specified, Dataset will appear at the root level. + */ + folder?: DatasetFolder; + /** + * The Salesforce Service Cloud object API name. Type: string (or Expression with resultType + * string). + */ + objectApiName?: any; } /** - * The on-premises SQL Server dataset. + * The Salesforce object dataset. */ -export interface SqlServerTableDataset { +export interface SalesforceObjectDataset { /** * Polymorphic Discriminator */ - type: "SqlServerTable"; + type: "SalesforceObject"; /** * Dataset description. */ @@ -10335,28 +10237,19 @@ export interface SqlServerTableDataset { */ folder?: DatasetFolder; /** - * This property will be retired. Please consider using schema + table properties instead. - */ - tableName?: any; - /** - * The schema name of the SQL Server dataset. Type: string (or Expression with resultType - * string). - */ - sqlServerTableDatasetSchema?: any; - /** - * The table name of the SQL Server dataset. Type: string (or Expression with resultType string). + * The Salesforce object API name. Type: string (or Expression with resultType string). */ - table?: any; + objectApiName?: any; } /** - * Sap Business Warehouse Open Hub Destination Table properties. + * The Microsoft Access table dataset. */ -export interface SapOpenHubTableDataset { +export interface MicrosoftAccessTableDataset { /** * Polymorphic Discriminator */ - type: "SapOpenHubTable"; + type: "MicrosoftAccessTable"; /** * Dataset description. */ @@ -10388,31 +10281,19 @@ export interface SapOpenHubTableDataset { */ folder?: DatasetFolder; /** - * The name of the Open Hub Destination with destination type as Database Table. Type: string (or - * Expression with resultType string). - */ - openHubDestinationName: any; - /** - * Whether to exclude the records of the last request. The default value is true. Type: boolean - * (or Expression with resultType boolean). - */ - excludeLastRequest?: any; - /** - * The ID of request for delta loading. Once it is set, only data with requestId larger than the - * value of this property will be retrieved. The default value is 0. Type: integer (or Expression - * with resultType integer ). + * The Microsoft Access table name. Type: string (or Expression with resultType string). */ - baseRequestId?: any; + tableName?: any; } /** - * SAP HANA Table properties. + * The PostgreSQL table dataset. */ -export interface SapHanaTableDataset { +export interface PostgreSqlTableDataset { /** * Polymorphic Discriminator */ - type: "SapHanaTable"; + type: "PostgreSqlTable"; /** * Dataset description. */ @@ -10444,23 +10325,27 @@ export interface SapHanaTableDataset { */ folder?: DatasetFolder; /** - * The schema name of SAP HANA. Type: string (or Expression with resultType string). + * This property will be retired. Please consider using schema + table properties instead. */ - sapHanaTableDatasetSchema?: any; + tableName?: any; /** - * The table name of SAP HANA. Type: string (or Expression with resultType string). + * The PostgreSQL table name. Type: string (or Expression with resultType string). */ table?: any; + /** + * The PostgreSQL schema name. Type: string (or Expression with resultType string). + */ + postgreSqlTableDatasetSchema?: any; } /** - * The path of the SAP ECC OData entity. + * The MySQL table dataset. */ -export interface SapEccResourceDataset { +export interface MySqlTableDataset { /** * Polymorphic Discriminator */ - type: "SapEccResource"; + type: "MySqlTable"; /** * Dataset description. */ @@ -10492,19 +10377,19 @@ export interface SapEccResourceDataset { */ folder?: DatasetFolder; /** - * The path of the SAP ECC OData entity. Type: string (or Expression with resultType string). + * The MySQL table name. Type: string (or Expression with resultType string). */ - path: any; + tableName?: any; } /** - * The path of the SAP Cloud for Customer OData entity. + * The ODBC table dataset. */ -export interface SapCloudForCustomerResourceDataset { +export interface OdbcTableDataset { /** * Polymorphic Discriminator */ - type: "SapCloudForCustomerResource"; + type: "OdbcTable"; /** * Dataset description. */ @@ -10536,20 +10421,19 @@ export interface SapCloudForCustomerResourceDataset { */ folder?: DatasetFolder; /** - * The path of the SAP Cloud for Customer OData entity. Type: string (or Expression with - * resultType string). + * The ODBC table name. Type: string (or Expression with resultType string). */ - path: any; + tableName?: any; } /** - * The SAP BW cube dataset. + * The Informix table dataset. */ -export interface SapBwCubeDataset { +export interface InformixTableDataset { /** * Polymorphic Discriminator */ - type: "SapBwCube"; + type: "InformixTable"; /** * Dataset description. */ @@ -10580,16 +10464,20 @@ export interface SapBwCubeDataset { * The folder that this Dataset is in. If not specified, Dataset will appear at the root level. */ folder?: DatasetFolder; + /** + * The Informix table name. Type: string (or Expression with resultType string). + */ + tableName?: any; } /** - * The Sybase table dataset. + * The relational table dataset. */ -export interface SybaseTableDataset { +export interface RelationalTableDataset { /** * Polymorphic Discriminator */ - type: "SybaseTable"; + type: "RelationalTable"; /** * Dataset description. */ @@ -10621,19 +10509,19 @@ export interface SybaseTableDataset { */ folder?: DatasetFolder; /** - * The Sybase table name. Type: string (or Expression with resultType string). + * The relational table name. Type: string (or Expression with resultType string). */ tableName?: any; } /** - * The Salesforce Service Cloud object dataset. + * The Db2 table dataset. */ -export interface SalesforceServiceCloudObjectDataset { +export interface Db2TableDataset { /** * Polymorphic Discriminator */ - type: "SalesforceServiceCloudObject"; + type: "Db2Table"; /** * Dataset description. */ @@ -10665,20 +10553,27 @@ export interface SalesforceServiceCloudObjectDataset { */ folder?: DatasetFolder; /** - * The Salesforce Service Cloud object API name. Type: string (or Expression with resultType - * string). + * This property will be retired. Please consider using schema + table properties instead. */ - objectApiName?: any; + tableName?: any; + /** + * The Db2 schema name. Type: string (or Expression with resultType string). + */ + db2TableDatasetSchema?: any; + /** + * The Db2 table name. Type: string (or Expression with resultType string). + */ + table?: any; } /** - * The Salesforce object dataset. + * The Amazon Redshift table dataset. */ -export interface SalesforceObjectDataset { +export interface AmazonRedshiftTableDataset { /** * Polymorphic Discriminator */ - type: "SalesforceObject"; + type: "AmazonRedshiftTable"; /** * Dataset description. */ @@ -10710,19 +10605,27 @@ export interface SalesforceObjectDataset { */ folder?: DatasetFolder; /** - * The Salesforce object API name. Type: string (or Expression with resultType string). + * This property will be retired. Please consider using schema + table properties instead. */ - objectApiName?: any; + tableName?: any; + /** + * The Amazon Redshift table name. Type: string (or Expression with resultType string). + */ + table?: any; + /** + * The Amazon Redshift schema name. Type: string (or Expression with resultType string). + */ + amazonRedshiftTableDatasetSchema?: any; } /** - * The Microsoft Access table dataset. + * The Azure MySQL database dataset. */ -export interface MicrosoftAccessTableDataset { +export interface AzureMySqlTableDataset { /** * Polymorphic Discriminator */ - type: "MicrosoftAccessTable"; + type: "AzureMySqlTable"; /** * Dataset description. */ @@ -10754,19 +10657,23 @@ export interface MicrosoftAccessTableDataset { */ folder?: DatasetFolder; /** - * The Microsoft Access table name. Type: string (or Expression with resultType string). + * The Azure MySQL database table name. Type: string (or Expression with resultType string). */ tableName?: any; + /** + * The name of Azure MySQL database table. Type: string (or Expression with resultType string). + */ + table?: any; } /** - * The PostgreSQL table dataset. + * The Teradata database dataset. */ -export interface PostgreSqlTableDataset { +export interface TeradataTableDataset { /** * Polymorphic Discriminator */ - type: "PostgreSqlTable"; + type: "TeradataTable"; /** * Dataset description. */ @@ -10798,27 +10705,23 @@ export interface PostgreSqlTableDataset { */ folder?: DatasetFolder; /** - * This property will be retired. Please consider using schema + table properties instead. + * The database name of Teradata. Type: string (or Expression with resultType string). */ - tableName?: any; + database?: any; /** - * The PostgreSQL table name. Type: string (or Expression with resultType string). + * The table name of Teradata. Type: string (or Expression with resultType string). */ table?: any; - /** - * The PostgreSQL schema name. Type: string (or Expression with resultType string). - */ - postgreSqlTableDatasetSchema?: any; } /** - * The MySQL table dataset. + * The on-premises Oracle database dataset. */ -export interface MySqlTableDataset { +export interface OracleTableDataset { /** * Polymorphic Discriminator */ - type: "MySqlTable"; + type: "OracleTable"; /** * Dataset description. */ @@ -10850,19 +10753,29 @@ export interface MySqlTableDataset { */ folder?: DatasetFolder; /** - * The MySQL table name. Type: string (or Expression with resultType string). + * This property will be retired. Please consider using schema + table properties instead. */ tableName?: any; + /** + * The schema name of the on-premises Oracle database. Type: string (or Expression with + * resultType string). + */ + oracleTableDatasetSchema?: any; + /** + * The table name of the on-premises Oracle database. Type: string (or Expression with resultType + * string). + */ + table?: any; } /** - * The ODBC table dataset. + * The Open Data Protocol (OData) resource dataset. */ -export interface OdbcTableDataset { +export interface ODataResourceDataset { /** * Polymorphic Discriminator */ - type: "OdbcTable"; + type: "ODataResource"; /** * Dataset description. */ @@ -10894,19 +10807,19 @@ export interface OdbcTableDataset { */ folder?: DatasetFolder; /** - * The ODBC table name. Type: string (or Expression with resultType string). + * The OData resource path. Type: string (or Expression with resultType string). */ - tableName?: any; + path?: any; } /** - * The Informix table dataset. + * The CosmosDB (MongoDB API) database dataset. */ -export interface InformixTableDataset { +export interface CosmosDbMongoDbApiCollectionDataset { /** * Polymorphic Discriminator */ - type: "InformixTable"; + type: "CosmosDbMongoDbApiCollection"; /** * Dataset description. */ @@ -10938,19 +10851,20 @@ export interface InformixTableDataset { */ folder?: DatasetFolder; /** - * The Informix table name. Type: string (or Expression with resultType string). + * The collection name of the CosmosDB (MongoDB API) database. Type: string (or Expression with + * resultType string). */ - tableName?: any; + collection: any; } /** - * The relational table dataset. + * The MongoDB database dataset. */ -export interface RelationalTableDataset { +export interface MongoDbV2CollectionDataset { /** * Polymorphic Discriminator */ - type: "RelationalTable"; + type: "MongoDbV2Collection"; /** * Dataset description. */ @@ -10982,19 +10896,20 @@ export interface RelationalTableDataset { */ folder?: DatasetFolder; /** - * The relational table name. Type: string (or Expression with resultType string). + * The collection name of the MongoDB database. Type: string (or Expression with resultType + * string). */ - tableName?: any; + collection: any; } /** - * The Db2 table dataset. + * The MongoDB database dataset. */ -export interface Db2TableDataset { +export interface MongoDbCollectionDataset { /** * Polymorphic Discriminator */ - type: "Db2Table"; + type: "MongoDbCollection"; /** * Dataset description. */ @@ -11026,27 +10941,19 @@ export interface Db2TableDataset { */ folder?: DatasetFolder; /** - * This property will be retired. Please consider using schema + table properties instead. - */ - tableName?: any; - /** - * The Db2 schema name. Type: string (or Expression with resultType string). - */ - db2TableDatasetSchema?: any; - /** - * The Db2 table name. Type: string (or Expression with resultType string). + * The table name of the MongoDB database. Type: string (or Expression with resultType string). */ - table?: any; + collectionName: any; } /** - * The Amazon Redshift table dataset. + * An on-premises file system dataset. */ -export interface AmazonRedshiftTableDataset { +export interface FileShareDataset { /** * Polymorphic Discriminator */ - type: "AmazonRedshiftTable"; + type: "FileShare"; /** * Dataset description. */ @@ -11078,27 +10985,44 @@ export interface AmazonRedshiftTableDataset { */ folder?: DatasetFolder; /** - * This property will be retired. Please consider using schema + table properties instead. + * The path of the on-premises file system. Type: string (or Expression with resultType string). */ - tableName?: any; + folderPath?: any; /** - * The Amazon Redshift table name. Type: string (or Expression with resultType string). + * The name of the on-premises file system. Type: string (or Expression with resultType string). */ - table?: any; + fileName?: any; /** - * The Amazon Redshift schema name. Type: string (or Expression with resultType string). + * The start of file's modified datetime. Type: string (or Expression with resultType string). */ - amazonRedshiftTableDatasetSchema?: any; + modifiedDatetimeStart?: any; + /** + * The end of file's modified datetime. Type: string (or Expression with resultType string). + */ + modifiedDatetimeEnd?: any; + /** + * The format of the files. + */ + format?: DatasetStorageFormatUnion; + /** + * Specify a filter to be used to select a subset of files in the folderPath rather than all + * files. Type: string (or Expression with resultType string). + */ + fileFilter?: any; + /** + * The data compression method used for the file system. + */ + compression?: DatasetCompressionUnion; } /** - * The Azure MySQL database dataset. + * The Office365 account. */ -export interface AzureMySqlTableDataset { +export interface Office365Dataset { /** * Polymorphic Discriminator */ - type: "AzureMySqlTable"; + type: "Office365Table"; /** * Dataset description. */ @@ -11130,23 +11054,25 @@ export interface AzureMySqlTableDataset { */ folder?: DatasetFolder; /** - * The Azure MySQL database table name. Type: string (or Expression with resultType string). + * Name of the dataset to extract from Office 365. Type: string (or Expression with resultType + * string). */ - tableName?: any; + tableName: any; /** - * The name of Azure MySQL database table. Type: string (or Expression with resultType string). + * A predicate expression that can be used to filter the specific rows to extract from Office + * 365. Type: string (or Expression with resultType string). */ - table?: any; + predicate?: any; } /** - * The Teradata database dataset. + * The Azure Data Lake Storage Gen2 storage. */ -export interface TeradataTableDataset { +export interface AzureBlobFSDataset { /** * Polymorphic Discriminator */ - type: "TeradataTable"; + type: "AzureBlobFSFile"; /** * Dataset description. */ @@ -11178,23 +11104,33 @@ export interface TeradataTableDataset { */ folder?: DatasetFolder; /** - * The database name of Teradata. Type: string (or Expression with resultType string). + * The path of the Azure Data Lake Storage Gen2 storage. Type: string (or Expression with + * resultType string). */ - database?: any; + folderPath?: any; /** - * The table name of Teradata. Type: string (or Expression with resultType string). + * The name of the Azure Data Lake Storage Gen2. Type: string (or Expression with resultType + * string). */ - table?: any; + fileName?: any; + /** + * The format of the Azure Data Lake Storage Gen2 storage. + */ + format?: DatasetStorageFormatUnion; + /** + * The data compression method used for the blob storage. + */ + compression?: DatasetCompressionUnion; } /** - * The on-premises Oracle database dataset. + * Azure Data Lake Store dataset. */ -export interface OracleTableDataset { +export interface AzureDataLakeStoreDataset { /** * Polymorphic Discriminator */ - type: "OracleTable"; + type: "AzureDataLakeStoreFile"; /** * Dataset description. */ @@ -11226,29 +11162,33 @@ export interface OracleTableDataset { */ folder?: DatasetFolder; /** - * This property will be retired. Please consider using schema + table properties instead. + * Path to the folder in the Azure Data Lake Store. Type: string (or Expression with resultType + * string). */ - tableName?: any; + folderPath?: any; /** - * The schema name of the on-premises Oracle database. Type: string (or Expression with - * resultType string). + * The name of the file in the Azure Data Lake Store. Type: string (or Expression with resultType + * string). */ - oracleTableDatasetSchema?: any; + fileName?: any; /** - * The table name of the on-premises Oracle database. Type: string (or Expression with resultType - * string). + * The format of the Data Lake Store. */ - table?: any; + format?: DatasetStorageFormatUnion; + /** + * The data compression method used for the item(s) in the Azure Data Lake Store. + */ + compression?: DatasetCompressionUnion; } /** - * The Open Data Protocol (OData) resource dataset. + * The Common Data Service for Apps entity dataset. */ -export interface ODataResourceDataset { +export interface CommonDataServiceForAppsEntityDataset { /** * Polymorphic Discriminator */ - type: "ODataResource"; + type: "CommonDataServiceForAppsEntity"; /** * Dataset description. */ @@ -11280,19 +11220,19 @@ export interface ODataResourceDataset { */ folder?: DatasetFolder; /** - * The OData resource path. Type: string (or Expression with resultType string). + * The logical name of the entity. Type: string (or Expression with resultType string). */ - path?: any; + entityName?: any; } /** - * The CosmosDB (MongoDB API) database dataset. + * The Dynamics CRM entity dataset. */ -export interface CosmosDbMongoDbApiCollectionDataset { +export interface DynamicsCrmEntityDataset { /** * Polymorphic Discriminator */ - type: "CosmosDbMongoDbApiCollection"; + type: "DynamicsCrmEntity"; /** * Dataset description. */ @@ -11324,20 +11264,19 @@ export interface CosmosDbMongoDbApiCollectionDataset { */ folder?: DatasetFolder; /** - * The collection name of the CosmosDB (MongoDB API) database. Type: string (or Expression with - * resultType string). + * The logical name of the entity. Type: string (or Expression with resultType string). */ - collection: any; + entityName?: any; } /** - * The MongoDB database dataset. + * The Dynamics entity dataset. */ -export interface MongoDbV2CollectionDataset { +export interface DynamicsEntityDataset { /** * Polymorphic Discriminator */ - type: "MongoDbV2Collection"; + type: "DynamicsEntity"; /** * Dataset description. */ @@ -11369,20 +11308,19 @@ export interface MongoDbV2CollectionDataset { */ folder?: DatasetFolder; /** - * The collection name of the MongoDB database. Type: string (or Expression with resultType - * string). + * The logical name of the entity. Type: string (or Expression with resultType string). */ - collection: any; + entityName?: any; } /** - * The MongoDB database dataset. + * Microsoft Azure Document Database Collection dataset. */ -export interface MongoDbCollectionDataset { +export interface DocumentDbCollectionDataset { /** * Polymorphic Discriminator */ - type: "MongoDbCollection"; + type: "DocumentDbCollection"; /** * Dataset description. */ @@ -11414,19 +11352,19 @@ export interface MongoDbCollectionDataset { */ folder?: DatasetFolder; /** - * The table name of the MongoDB database. Type: string (or Expression with resultType string). + * Document Database collection name. Type: string (or Expression with resultType string). */ collectionName: any; } /** - * An on-premises file system dataset. + * Microsoft Azure CosmosDB (SQL API) Collection dataset. */ -export interface FileShareDataset { +export interface CosmosDbSqlApiCollectionDataset { /** * Polymorphic Discriminator */ - type: "FileShare"; + type: "CosmosDbSqlApiCollection"; /** * Dataset description. */ @@ -11458,44 +11396,19 @@ export interface FileShareDataset { */ folder?: DatasetFolder; /** - * The path of the on-premises file system. Type: string (or Expression with resultType string). - */ - folderPath?: any; - /** - * The name of the on-premises file system. Type: string (or Expression with resultType string). - */ - fileName?: any; - /** - * The start of file's modified datetime. Type: string (or Expression with resultType string). - */ - modifiedDatetimeStart?: any; - /** - * The end of file's modified datetime. Type: string (or Expression with resultType string). - */ - modifiedDatetimeEnd?: any; - /** - * The format of the files. - */ - format?: DatasetStorageFormatUnion; - /** - * Specify a filter to be used to select a subset of files in the folderPath rather than all - * files. Type: string (or Expression with resultType string). - */ - fileFilter?: any; - /** - * The data compression method used for the file system. + * CosmosDB (SQL API) collection name. Type: string (or Expression with resultType string). */ - compression?: DatasetCompressionUnion; + collectionName: any; } /** - * The Office365 account. + * The custom dataset. */ -export interface Office365Dataset { +export interface CustomDataset { /** * Polymorphic Discriminator */ - type: "Office365Table"; + type: "CustomDataset"; /** * Dataset description. */ @@ -11527,25 +11440,19 @@ export interface Office365Dataset { */ folder?: DatasetFolder; /** - * Name of the dataset to extract from Office 365. Type: string (or Expression with resultType - * string). - */ - tableName: any; - /** - * A predicate expression that can be used to filter the specific rows to extract from Office - * 365. Type: string (or Expression with resultType string). + * Custom dataset properties. */ - predicate?: any; + typeProperties?: any; } /** - * The Azure Data Lake Storage Gen2 storage. + * The Cassandra database dataset. */ -export interface AzureBlobFSDataset { +export interface CassandraTableDataset { /** * Polymorphic Discriminator */ - type: "AzureBlobFSFile"; + type: "CassandraTable"; /** * Dataset description. */ @@ -11577,33 +11484,23 @@ export interface AzureBlobFSDataset { */ folder?: DatasetFolder; /** - * The path of the Azure Data Lake Storage Gen2 storage. Type: string (or Expression with - * resultType string). - */ - folderPath?: any; - /** - * The name of the Azure Data Lake Storage Gen2. Type: string (or Expression with resultType - * string). - */ - fileName?: any; - /** - * The format of the Azure Data Lake Storage Gen2 storage. + * The table name of the Cassandra database. Type: string (or Expression with resultType string). */ - format?: DatasetStorageFormatUnion; + tableName?: any; /** - * The data compression method used for the blob storage. + * The keyspace of the Cassandra database. Type: string (or Expression with resultType string). */ - compression?: DatasetCompressionUnion; + keyspace?: any; } /** - * Azure Data Lake Store dataset. + * The Azure SQL Data Warehouse dataset. */ -export interface AzureDataLakeStoreDataset { +export interface AzureSqlDWTableDataset { /** * Polymorphic Discriminator */ - type: "AzureDataLakeStoreFile"; + type: "AzureSqlDWTable"; /** * Dataset description. */ @@ -11635,33 +11532,29 @@ export interface AzureDataLakeStoreDataset { */ folder?: DatasetFolder; /** - * Path to the folder in the Azure Data Lake Store. Type: string (or Expression with resultType - * string). + * This property will be retired. Please consider using schema + table properties instead. */ - folderPath?: any; + tableName?: any; /** - * The name of the file in the Azure Data Lake Store. Type: string (or Expression with resultType + * The schema name of the Azure SQL Data Warehouse. Type: string (or Expression with resultType * string). */ - fileName?: any; - /** - * The format of the Data Lake Store. - */ - format?: DatasetStorageFormatUnion; + azureSqlDWTableDatasetSchema?: any; /** - * The data compression method used for the item(s) in the Azure Data Lake Store. + * The table name of the Azure SQL Data Warehouse. Type: string (or Expression with resultType + * string). */ - compression?: DatasetCompressionUnion; + table?: any; } /** - * The Common Data Service for Apps entity dataset. + * The Azure SQL Managed Instance dataset. */ -export interface CommonDataServiceForAppsEntityDataset { +export interface AzureSqlMITableDataset { /** * Polymorphic Discriminator */ - type: "CommonDataServiceForAppsEntity"; + type: "AzureSqlMITable"; /** * Dataset description. */ @@ -11693,19 +11586,29 @@ export interface CommonDataServiceForAppsEntityDataset { */ folder?: DatasetFolder; /** - * The logical name of the entity. Type: string (or Expression with resultType string). + * This property will be retired. Please consider using schema + table properties instead. */ - entityName?: any; + tableName?: any; + /** + * The schema name of the Azure SQL Managed Instance. Type: string (or Expression with resultType + * string). + */ + azureSqlMITableDatasetSchema?: any; + /** + * The table name of the Azure SQL Managed Instance dataset. Type: string (or Expression with + * resultType string). + */ + table?: any; } /** - * The Dynamics CRM entity dataset. + * The Azure SQL Server database dataset. */ -export interface DynamicsCrmEntityDataset { +export interface AzureSqlTableDataset { /** * Polymorphic Discriminator */ - type: "DynamicsCrmEntity"; + type: "AzureSqlTable"; /** * Dataset description. */ @@ -11737,19 +11640,28 @@ export interface DynamicsCrmEntityDataset { */ folder?: DatasetFolder; /** - * The logical name of the entity. Type: string (or Expression with resultType string). + * This property will be retired. Please consider using schema + table properties instead. */ - entityName?: any; + tableName?: any; + /** + * The schema name of the Azure SQL database. Type: string (or Expression with resultType + * string). + */ + azureSqlTableDatasetSchema?: any; + /** + * The table name of the Azure SQL database. Type: string (or Expression with resultType string). + */ + table?: any; } /** - * The Dynamics entity dataset. + * The Azure Table storage dataset. */ -export interface DynamicsEntityDataset { +export interface AzureTableDataset { /** * Polymorphic Discriminator */ - type: "DynamicsEntity"; + type: "AzureTable"; /** * Dataset description. */ @@ -11781,19 +11693,20 @@ export interface DynamicsEntityDataset { */ folder?: DatasetFolder; /** - * The logical name of the entity. Type: string (or Expression with resultType string). + * The table name of the Azure Table storage. Type: string (or Expression with resultType + * string). */ - entityName?: any; + tableName: any; } /** - * Microsoft Azure Document Database Collection dataset. + * The Azure Blob storage. */ -export interface DocumentDbCollectionDataset { +export interface AzureBlobDataset { /** * Polymorphic Discriminator */ - type: "DocumentDbCollection"; + type: "AzureBlob"; /** * Dataset description. */ @@ -11825,263 +11738,300 @@ export interface DocumentDbCollectionDataset { */ folder?: DatasetFolder; /** - * Document Database collection name. Type: string (or Expression with resultType string). + * The path of the Azure Blob storage. Type: string (or Expression with resultType string). */ - collectionName: any; + folderPath?: any; + /** + * The root of blob path. Type: string (or Expression with resultType string). + */ + tableRootLocation?: any; + /** + * The name of the Azure Blob. Type: string (or Expression with resultType string). + */ + fileName?: any; + /** + * The start of Azure Blob's modified datetime. Type: string (or Expression with resultType + * string). + */ + modifiedDatetimeStart?: any; + /** + * The end of Azure Blob's modified datetime. Type: string (or Expression with resultType + * string). + */ + modifiedDatetimeEnd?: any; + /** + * The format of the Azure Blob storage. + */ + format?: DatasetStorageFormatUnion; + /** + * The data compression method used for the blob storage. + */ + compression?: DatasetCompressionUnion; } /** - * Microsoft Azure CosmosDB (SQL API) Collection dataset. + * Contains the possible cases for DatasetLocation. + */ +export type DatasetLocationUnion = DatasetLocation | HdfsLocation | HttpServerLocation | SftpLocation | FtpServerLocation | GoogleCloudStorageLocation | AzureFileStorageLocation | FileServerLocation | AmazonS3Location | AzureDataLakeStoreLocation | AzureBlobFSLocation | AzureBlobStorageLocation; + +/** + * Dataset location. */ -export interface CosmosDbSqlApiCollectionDataset { +export interface DatasetLocation { /** * Polymorphic Discriminator */ - type: "CosmosDbSqlApiCollection"; - /** - * Dataset description. - */ - description?: string; - /** - * Columns that define the structure of the dataset. Type: array (or Expression with resultType - * array), itemType: DatasetDataElement. - */ - structure?: any; + type: "DatasetLocation"; /** - * Columns that define the physical type schema of the dataset. Type: array (or Expression with - * resultType array), itemType: DatasetSchemaDataElement. + * Specify the folder path of dataset. Type: string (or Expression with resultType string) */ - schema?: any; + folderPath?: any; /** - * Linked service reference. + * Specify the file name of dataset. Type: string (or Expression with resultType string). */ - linkedServiceName: LinkedServiceReference; + fileName?: any; /** - * Parameters for dataset. + * Describes unknown properties. The value of an unknown property can be of "any" type. */ - parameters?: { [propertyName: string]: ParameterSpecification }; + [property: string]: any; +} + +/** + * The location of HDFS. + */ +export interface HdfsLocation { /** - * List of tags that can be used for describing the Dataset. + * Polymorphic Discriminator */ - annotations?: any[]; + type: "HdfsLocation"; /** - * The folder that this Dataset is in. If not specified, Dataset will appear at the root level. + * Specify the folder path of dataset. Type: string (or Expression with resultType string) */ - folder?: DatasetFolder; + folderPath?: any; /** - * CosmosDB (SQL API) collection name. Type: string (or Expression with resultType string). + * Specify the file name of dataset. Type: string (or Expression with resultType string). */ - collectionName: any; + fileName?: any; } /** - * The custom dataset. + * The location of http server. */ -export interface CustomDataset { +export interface HttpServerLocation { /** * Polymorphic Discriminator */ - type: "CustomDataset"; - /** - * Dataset description. - */ - description?: string; - /** - * Columns that define the structure of the dataset. Type: array (or Expression with resultType - * array), itemType: DatasetDataElement. - */ - structure?: any; + type: "HttpServerLocation"; /** - * Columns that define the physical type schema of the dataset. Type: array (or Expression with - * resultType array), itemType: DatasetSchemaDataElement. + * Specify the folder path of dataset. Type: string (or Expression with resultType string) */ - schema?: any; + folderPath?: any; /** - * Linked service reference. + * Specify the file name of dataset. Type: string (or Expression with resultType string). */ - linkedServiceName: LinkedServiceReference; + fileName?: any; /** - * Parameters for dataset. + * Specify the relativeUrl of http server. Type: string (or Expression with resultType string) */ - parameters?: { [propertyName: string]: ParameterSpecification }; + relativeUrl?: any; +} + +/** + * The location of SFTP dataset. + */ +export interface SftpLocation { /** - * List of tags that can be used for describing the Dataset. + * Polymorphic Discriminator */ - annotations?: any[]; + type: "SftpLocation"; /** - * The folder that this Dataset is in. If not specified, Dataset will appear at the root level. + * Specify the folder path of dataset. Type: string (or Expression with resultType string) */ - folder?: DatasetFolder; + folderPath?: any; /** - * Custom dataset properties. + * Specify the file name of dataset. Type: string (or Expression with resultType string). */ - typeProperties?: any; + fileName?: any; } /** - * The Cassandra database dataset. + * The location of ftp server dataset. */ -export interface CassandraTableDataset { +export interface FtpServerLocation { /** * Polymorphic Discriminator */ - type: "CassandraTable"; - /** - * Dataset description. - */ - description?: string; - /** - * Columns that define the structure of the dataset. Type: array (or Expression with resultType - * array), itemType: DatasetDataElement. - */ - structure?: any; + type: "FtpServerLocation"; /** - * Columns that define the physical type schema of the dataset. Type: array (or Expression with - * resultType array), itemType: DatasetSchemaDataElement. + * Specify the folder path of dataset. Type: string (or Expression with resultType string) */ - schema?: any; + folderPath?: any; /** - * Linked service reference. + * Specify the file name of dataset. Type: string (or Expression with resultType string). */ - linkedServiceName: LinkedServiceReference; + fileName?: any; +} + +/** + * The location of Google Cloud Storage dataset. + */ +export interface GoogleCloudStorageLocation { /** - * Parameters for dataset. + * Polymorphic Discriminator */ - parameters?: { [propertyName: string]: ParameterSpecification }; + type: "GoogleCloudStorageLocation"; /** - * List of tags that can be used for describing the Dataset. + * Specify the folder path of dataset. Type: string (or Expression with resultType string) */ - annotations?: any[]; + folderPath?: any; /** - * The folder that this Dataset is in. If not specified, Dataset will appear at the root level. + * Specify the file name of dataset. Type: string (or Expression with resultType string). */ - folder?: DatasetFolder; + fileName?: any; /** - * The table name of the Cassandra database. Type: string (or Expression with resultType string). + * Specify the bucketName of Google Cloud Storage. Type: string (or Expression with resultType + * string) */ - tableName?: any; + bucketName?: any; /** - * The keyspace of the Cassandra database. Type: string (or Expression with resultType string). + * Specify the version of Google Cloud Storage. Type: string (or Expression with resultType + * string). */ - keyspace?: any; + version?: any; } /** - * The Azure SQL Data Warehouse dataset. + * The location of file server dataset. */ -export interface AzureSqlDWTableDataset { +export interface AzureFileStorageLocation { /** * Polymorphic Discriminator */ - type: "AzureSqlDWTable"; + type: "AzureFileStorageLocation"; /** - * Dataset description. + * Specify the folder path of dataset. Type: string (or Expression with resultType string) */ - description?: string; + folderPath?: any; /** - * Columns that define the structure of the dataset. Type: array (or Expression with resultType - * array), itemType: DatasetDataElement. + * Specify the file name of dataset. Type: string (or Expression with resultType string). */ - structure?: any; + fileName?: any; +} + +/** + * The location of file server dataset. + */ +export interface FileServerLocation { /** - * Columns that define the physical type schema of the dataset. Type: array (or Expression with - * resultType array), itemType: DatasetSchemaDataElement. + * Polymorphic Discriminator */ - schema?: any; + type: "FileServerLocation"; /** - * Linked service reference. + * Specify the folder path of dataset. Type: string (or Expression with resultType string) */ - linkedServiceName: LinkedServiceReference; + folderPath?: any; /** - * Parameters for dataset. + * Specify the file name of dataset. Type: string (or Expression with resultType string). */ - parameters?: { [propertyName: string]: ParameterSpecification }; + fileName?: any; +} + +/** + * The location of amazon S3 dataset. + */ +export interface AmazonS3Location { /** - * List of tags that can be used for describing the Dataset. + * Polymorphic Discriminator */ - annotations?: any[]; + type: "AmazonS3Location"; /** - * The folder that this Dataset is in. If not specified, Dataset will appear at the root level. + * Specify the folder path of dataset. Type: string (or Expression with resultType string) */ - folder?: DatasetFolder; + folderPath?: any; /** - * This property will be retired. Please consider using schema + table properties instead. + * Specify the file name of dataset. Type: string (or Expression with resultType string). */ - tableName?: any; + fileName?: any; /** - * The schema name of the Azure SQL Data Warehouse. Type: string (or Expression with resultType - * string). + * Specify the bucketName of amazon S3. Type: string (or Expression with resultType string) */ - azureSqlDWTableDatasetSchema?: any; + bucketName?: any; /** - * The table name of the Azure SQL Data Warehouse. Type: string (or Expression with resultType - * string). + * Specify the version of amazon S3. Type: string (or Expression with resultType string). */ - table?: any; + version?: any; } /** - * The Azure SQL Managed Instance dataset. + * The location of azure data lake store dataset. */ -export interface AzureSqlMITableDataset { +export interface AzureDataLakeStoreLocation { /** * Polymorphic Discriminator */ - type: "AzureSqlMITable"; + type: "AzureDataLakeStoreLocation"; /** - * Dataset description. + * Specify the folder path of dataset. Type: string (or Expression with resultType string) */ - description?: string; + folderPath?: any; /** - * Columns that define the structure of the dataset. Type: array (or Expression with resultType - * array), itemType: DatasetDataElement. + * Specify the file name of dataset. Type: string (or Expression with resultType string). */ - structure?: any; + fileName?: any; +} + +/** + * The location of azure blobFS dataset. + */ +export interface AzureBlobFSLocation { /** - * Columns that define the physical type schema of the dataset. Type: array (or Expression with - * resultType array), itemType: DatasetSchemaDataElement. + * Polymorphic Discriminator */ - schema?: any; + type: "AzureBlobFSLocation"; /** - * Linked service reference. + * Specify the folder path of dataset. Type: string (or Expression with resultType string) */ - linkedServiceName: LinkedServiceReference; + folderPath?: any; /** - * Parameters for dataset. + * Specify the file name of dataset. Type: string (or Expression with resultType string). */ - parameters?: { [propertyName: string]: ParameterSpecification }; + fileName?: any; /** - * List of tags that can be used for describing the Dataset. + * Specify the fileSystem of azure blobFS. Type: string (or Expression with resultType string). */ - annotations?: any[]; + fileSystem?: any; +} + +/** + * The location of azure blob dataset. + */ +export interface AzureBlobStorageLocation { /** - * The folder that this Dataset is in. If not specified, Dataset will appear at the root level. + * Polymorphic Discriminator */ - folder?: DatasetFolder; + type: "AzureBlobStorageLocation"; /** - * This property will be retired. Please consider using schema + table properties instead. + * Specify the folder path of dataset. Type: string (or Expression with resultType string) */ - tableName?: any; + folderPath?: any; /** - * The schema name of the Azure SQL Managed Instance. Type: string (or Expression with resultType - * string). + * Specify the file name of dataset. Type: string (or Expression with resultType string). */ - azureSqlMITableDatasetSchema?: any; + fileName?: any; /** - * The table name of the Azure SQL Managed Instance dataset. Type: string (or Expression with - * resultType string). + * Specify the container of azure blob. Type: string (or Expression with resultType string). */ - table?: any; + container?: any; } /** - * The Azure SQL Server database dataset. + * Binary dataset. */ -export interface AzureSqlTableDataset { +export interface BinaryDataset { /** * Polymorphic Discriminator */ - type: "AzureSqlTable"; + type: "Binary"; /** * Dataset description. */ @@ -12113,28 +12063,23 @@ export interface AzureSqlTableDataset { */ folder?: DatasetFolder; /** - * This property will be retired. Please consider using schema + table properties instead. - */ - tableName?: any; - /** - * The schema name of the Azure SQL database. Type: string (or Expression with resultType - * string). + * The location of the Binary storage. */ - azureSqlTableDatasetSchema?: any; + location: DatasetLocationUnion; /** - * The table name of the Azure SQL database. Type: string (or Expression with resultType string). + * The data compression method used for the binary dataset. */ - table?: any; + compression?: DatasetCompressionUnion; } /** - * The Azure Table storage dataset. + * ORC dataset. */ -export interface AzureTableDataset { +export interface OrcDataset { /** * Polymorphic Discriminator */ - type: "AzureTable"; + type: "Orc"; /** * Dataset description. */ @@ -12166,20 +12111,23 @@ export interface AzureTableDataset { */ folder?: DatasetFolder; /** - * The table name of the Azure Table storage. Type: string (or Expression with resultType - * string). + * The location of the ORC data storage. */ - tableName: any; + location: DatasetLocationUnion; + /** + * Possible values include: 'none', 'zlib', 'snappy' + */ + orcCompressionCodec?: OrcCompressionCodec; } /** - * The Azure Blob storage. + * Json dataset. */ -export interface AzureBlobDataset { +export interface JsonDataset { /** * Polymorphic Discriminator */ - type: "AzureBlob"; + type: "Json"; /** * Dataset description. */ @@ -12211,300 +12159,158 @@ export interface AzureBlobDataset { */ folder?: DatasetFolder; /** - * The path of the Azure Blob storage. Type: string (or Expression with resultType string). - */ - folderPath?: any; - /** - * The root of blob path. Type: string (or Expression with resultType string). - */ - tableRootLocation?: any; - /** - * The name of the Azure Blob. Type: string (or Expression with resultType string). - */ - fileName?: any; - /** - * The start of Azure Blob's modified datetime. Type: string (or Expression with resultType - * string). - */ - modifiedDatetimeStart?: any; - /** - * The end of Azure Blob's modified datetime. Type: string (or Expression with resultType - * string). + * The location of the json data storage. */ - modifiedDatetimeEnd?: any; + location: DatasetLocationUnion; /** - * The format of the Azure Blob storage. + * The code page name of the preferred encoding. If not specified, the default value is UTF-8, + * unless BOM denotes another Unicode encoding. Refer to the name column of the table in the + * following link to set supported values: + * https://msdn.microsoft.com/library/system.text.encoding.aspx. Type: string (or Expression with + * resultType string). */ - format?: DatasetStorageFormatUnion; + encodingName?: any; /** - * The data compression method used for the blob storage. + * The data compression method used for the json dataset. */ compression?: DatasetCompressionUnion; } /** - * Contains the possible cases for DatasetLocation. - */ -export type DatasetLocationUnion = DatasetLocation | HdfsLocation | HttpServerLocation | SftpLocation | FtpServerLocation | GoogleCloudStorageLocation | AzureFileStorageLocation | FileServerLocation | AmazonS3Location | AzureDataLakeStoreLocation | AzureBlobFSLocation | AzureBlobStorageLocation; - -/** - * Dataset location. - */ -export interface DatasetLocation { - /** - * Polymorphic Discriminator - */ - type: "DatasetLocation"; - /** - * Specify the folder path of dataset. Type: string (or Expression with resultType string) - */ - folderPath?: any; - /** - * Specify the file name of dataset. Type: string (or Expression with resultType string). - */ - fileName?: any; - /** - * Describes unknown properties. The value of an unknown property can be of "any" type. - */ - [property: string]: any; -} - -/** - * The location of HDFS. - */ -export interface HdfsLocation { - /** - * Polymorphic Discriminator - */ - type: "HdfsLocation"; - /** - * Specify the folder path of dataset. Type: string (or Expression with resultType string) - */ - folderPath?: any; - /** - * Specify the file name of dataset. Type: string (or Expression with resultType string). - */ - fileName?: any; -} - -/** - * The location of http server. - */ -export interface HttpServerLocation { - /** - * Polymorphic Discriminator - */ - type: "HttpServerLocation"; - /** - * Specify the folder path of dataset. Type: string (or Expression with resultType string) - */ - folderPath?: any; - /** - * Specify the file name of dataset. Type: string (or Expression with resultType string). - */ - fileName?: any; - /** - * Specify the relativeUrl of http server. Type: string (or Expression with resultType string) - */ - relativeUrl?: any; -} - -/** - * The location of SFTP dataset. - */ -export interface SftpLocation { - /** - * Polymorphic Discriminator - */ - type: "SftpLocation"; - /** - * Specify the folder path of dataset. Type: string (or Expression with resultType string) - */ - folderPath?: any; - /** - * Specify the file name of dataset. Type: string (or Expression with resultType string). - */ - fileName?: any; -} - -/** - * The location of ftp server dataset. - */ -export interface FtpServerLocation { - /** - * Polymorphic Discriminator - */ - type: "FtpServerLocation"; - /** - * Specify the folder path of dataset. Type: string (or Expression with resultType string) - */ - folderPath?: any; - /** - * Specify the file name of dataset. Type: string (or Expression with resultType string). - */ - fileName?: any; -} - -/** - * The location of Google Cloud Storage dataset. + * Delimited text dataset. */ -export interface GoogleCloudStorageLocation { +export interface DelimitedTextDataset { /** * Polymorphic Discriminator */ - type: "GoogleCloudStorageLocation"; - /** - * Specify the folder path of dataset. Type: string (or Expression with resultType string) - */ - folderPath?: any; - /** - * Specify the file name of dataset. Type: string (or Expression with resultType string). - */ - fileName?: any; + type: "DelimitedText"; /** - * Specify the bucketName of Google Cloud Storage. Type: string (or Expression with resultType - * string) + * Dataset description. */ - bucketName?: any; + description?: string; /** - * Specify the version of Google Cloud Storage. Type: string (or Expression with resultType - * string). + * Columns that define the structure of the dataset. Type: array (or Expression with resultType + * array), itemType: DatasetDataElement. */ - version?: any; -} - -/** - * The location of file server dataset. - */ -export interface AzureFileStorageLocation { + structure?: any; /** - * Polymorphic Discriminator + * Columns that define the physical type schema of the dataset. Type: array (or Expression with + * resultType array), itemType: DatasetSchemaDataElement. */ - type: "AzureFileStorageLocation"; + schema?: any; /** - * Specify the folder path of dataset. Type: string (or Expression with resultType string) + * Linked service reference. */ - folderPath?: any; + linkedServiceName: LinkedServiceReference; /** - * Specify the file name of dataset. Type: string (or Expression with resultType string). + * Parameters for dataset. */ - fileName?: any; -} - -/** - * The location of file server dataset. - */ -export interface FileServerLocation { + parameters?: { [propertyName: string]: ParameterSpecification }; /** - * Polymorphic Discriminator + * List of tags that can be used for describing the Dataset. */ - type: "FileServerLocation"; + annotations?: any[]; /** - * Specify the folder path of dataset. Type: string (or Expression with resultType string) + * The folder that this Dataset is in. If not specified, Dataset will appear at the root level. */ - folderPath?: any; + folder?: DatasetFolder; /** - * Specify the file name of dataset. Type: string (or Expression with resultType string). - */ - fileName?: any; -} - -/** - * The location of amazon S3 dataset. - */ -export interface AmazonS3Location { + * The location of the delimited text storage. + */ + location: DatasetLocationUnion; /** - * Polymorphic Discriminator + * The column delimiter. Type: string (or Expression with resultType string). */ - type: "AmazonS3Location"; + columnDelimiter?: any; /** - * Specify the folder path of dataset. Type: string (or Expression with resultType string) + * The row delimiter. Type: string (or Expression with resultType string). */ - folderPath?: any; + rowDelimiter?: any; /** - * Specify the file name of dataset. Type: string (or Expression with resultType string). + * The code page name of the preferred encoding. If miss, the default value is UTF-8, unless BOM + * denotes another Unicode encoding. Refer to the name column of the table in the following link + * to set supported values: https://msdn.microsoft.com/library/system.text.encoding.aspx. Type: + * string (or Expression with resultType string). */ - fileName?: any; + encodingName?: any; + compressionCodec?: any; /** - * Specify the bucketName of amazon S3. Type: string (or Expression with resultType string) + * The data compression method used for DelimitedText. */ - bucketName?: any; + compressionLevel?: any; /** - * Specify the version of amazon S3. Type: string (or Expression with resultType string). + * The quote character. Type: string (or Expression with resultType string). */ - version?: any; -} - -/** - * The location of azure data lake store dataset. - */ -export interface AzureDataLakeStoreLocation { + quoteChar?: any; /** - * Polymorphic Discriminator + * The escape character. Type: string (or Expression with resultType string). */ - type: "AzureDataLakeStoreLocation"; + escapeChar?: any; /** - * Specify the folder path of dataset. Type: string (or Expression with resultType string) + * When used as input, treat the first row of data as headers. When used as output,write the + * headers into the output as the first row of data. The default value is false. Type: boolean + * (or Expression with resultType boolean). */ - folderPath?: any; + firstRowAsHeader?: any; /** - * Specify the file name of dataset. Type: string (or Expression with resultType string). + * The null value string. Type: string (or Expression with resultType string). */ - fileName?: any; + nullValue?: any; } /** - * The location of azure blobFS dataset. + * Parquet dataset. */ -export interface AzureBlobFSLocation { +export interface ParquetDataset { /** * Polymorphic Discriminator */ - type: "AzureBlobFSLocation"; + type: "Parquet"; /** - * Specify the folder path of dataset. Type: string (or Expression with resultType string) + * Dataset description. */ - folderPath?: any; + description?: string; /** - * Specify the file name of dataset. Type: string (or Expression with resultType string). + * Columns that define the structure of the dataset. Type: array (or Expression with resultType + * array), itemType: DatasetDataElement. */ - fileName?: any; + structure?: any; /** - * Specify the fileSystem of azure blobFS. Type: string (or Expression with resultType string). + * Columns that define the physical type schema of the dataset. Type: array (or Expression with + * resultType array), itemType: DatasetSchemaDataElement. */ - fileSystem?: any; -} - -/** - * The location of azure blob dataset. - */ -export interface AzureBlobStorageLocation { + schema?: any; /** - * Polymorphic Discriminator + * Linked service reference. */ - type: "AzureBlobStorageLocation"; + linkedServiceName: LinkedServiceReference; /** - * Specify the folder path of dataset. Type: string (or Expression with resultType string) + * Parameters for dataset. */ - folderPath?: any; + parameters?: { [propertyName: string]: ParameterSpecification }; /** - * Specify the file name of dataset. Type: string (or Expression with resultType string). + * List of tags that can be used for describing the Dataset. */ - fileName?: any; + annotations?: any[]; /** - * Specify the container of azure blob. Type: string (or Expression with resultType string). + * The folder that this Dataset is in. If not specified, Dataset will appear at the root level. */ - container?: any; + folder?: DatasetFolder; + /** + * The location of the parquet storage. + */ + location: DatasetLocationUnion; + compressionCodec?: any; } /** - * Binary dataset. + * Avro dataset. */ -export interface BinaryDataset { +export interface AvroDataset { /** * Polymorphic Discriminator */ - type: "Binary"; + type: "Avro"; /** * Dataset description. */ @@ -12536,23 +12342,24 @@ export interface BinaryDataset { */ folder?: DatasetFolder; /** - * The location of the Binary storage. + * The location of the avro storage. */ location: DatasetLocationUnion; /** - * The data compression method used for the binary dataset. + * Possible values include: 'none', 'deflate', 'snappy', 'xz', 'bzip2' */ - compression?: DatasetCompressionUnion; + avroCompressionCodec?: AvroCompressionCodec; + avroCompressionLevel?: number; } /** - * ORC dataset. + * A single Amazon Simple Storage Service (S3) object or a set of S3 objects. */ -export interface OrcDataset { +export interface AmazonS3Dataset { /** * Polymorphic Discriminator */ - type: "Orc"; + type: "AmazonS3Object"; /** * Dataset description. */ @@ -12584,318 +12391,529 @@ export interface OrcDataset { */ folder?: DatasetFolder; /** - * The location of the ORC data storage. + * The name of the Amazon S3 bucket. Type: string (or Expression with resultType string). */ - location: DatasetLocationUnion; + bucketName: any; /** - * Possible values include: 'none', 'zlib', 'snappy' + * The key of the Amazon S3 object. Type: string (or Expression with resultType string). + */ + key?: any; + /** + * The prefix filter for the S3 object name. Type: string (or Expression with resultType string). + */ + prefix?: any; + /** + * The version for the S3 object. Type: string (or Expression with resultType string). + */ + version?: any; + /** + * The start of S3 object's modified datetime. Type: string (or Expression with resultType + * string). + */ + modifiedDatetimeStart?: any; + /** + * The end of S3 object's modified datetime. Type: string (or Expression with resultType string). + */ + modifiedDatetimeEnd?: any; + /** + * The format of files. + */ + format?: DatasetStorageFormatUnion; + /** + * The data compression method used for the Amazon S3 object. + */ + compression?: DatasetCompressionUnion; +} + +/** + * Trigger that allows the referenced pipeline to depend on other pipeline runs based on + * runDimension Name/Value pairs. Upstream pipelines should declare the same runDimension Name and + * their runs should have the values for those runDimensions. The referenced pipeline run would be + * triggered if the values for the runDimension match for all upstream pipeline runs. + */ +export interface ChainingTrigger { + /** + * Polymorphic Discriminator + */ + type: "ChainingTrigger"; + /** + * Trigger description. + */ + description?: string; + /** + * Indicates if trigger is running or not. Updated when Start/Stop APIs are called on the + * Trigger. Possible values include: 'Started', 'Stopped', 'Disabled' + * **NOTE: This property will not be serialized. It can only be populated by the server.** + */ + readonly runtimeState?: TriggerRuntimeState; + /** + * List of tags that can be used for describing the trigger. + */ + annotations?: any[]; + /** + * Pipeline for which runs are created when all upstream pipelines complete successfully. + */ + pipelineProperty: TriggerPipelineReference; + /** + * Upstream Pipelines. + */ + dependsOn: PipelineReference[]; + /** + * Run Dimension property that needs to be emitted by upstream pipelines. + */ + runDimension: string; +} + +/** + * Trigger that schedules pipeline reruns for all fixed time interval windows from a requested + * start time to requested end time. + */ +export interface RerunTumblingWindowTrigger { + /** + * Polymorphic Discriminator + */ + type: "RerunTumblingWindowTrigger"; + /** + * Trigger description. + */ + description?: string; + /** + * Indicates if trigger is running or not. Updated when Start/Stop APIs are called on the + * Trigger. Possible values include: 'Started', 'Stopped', 'Disabled' + * **NOTE: This property will not be serialized. It can only be populated by the server.** + */ + readonly runtimeState?: TriggerRuntimeState; + /** + * List of tags that can be used for describing the trigger. + */ + annotations?: any[]; + /** + * The parent trigger reference. + */ + parentTrigger: any; + /** + * The start time for the time period for which restatement is initiated. Only UTC time is + * currently supported. + */ + requestedStartTime: Date; + /** + * The end time for the time period for which restatement is initiated. Only UTC time is + * currently supported. + */ + requestedEndTime: Date; + /** + * The max number of parallel time windows (ready for execution) for which a rerun is triggered. + */ + rerunConcurrency: number; +} + +/** + * Contains the possible cases for DependencyReference. + */ +export type DependencyReferenceUnion = DependencyReference | SelfDependencyTumblingWindowTriggerReference | TriggerDependencyReferenceUnion; + +/** + * Referenced dependency. + */ +export interface DependencyReference { + /** + * Polymorphic Discriminator + */ + type: "DependencyReference"; +} + +/** + * Self referenced tumbling window trigger dependency. + */ +export interface SelfDependencyTumblingWindowTriggerReference { + /** + * Polymorphic Discriminator + */ + type: "SelfDependencyTumblingWindowTriggerReference"; + /** + * Timespan applied to the start time of a tumbling window when evaluating dependency. + */ + offset: string; + /** + * The size of the window when evaluating the dependency. If undefined the frequency of the + * tumbling window will be used. */ - orcCompressionCodec?: OrcCompressionCodec; + size?: string; } /** - * Json dataset. + * Trigger reference type. */ -export interface JsonDataset { - /** - * Polymorphic Discriminator - */ - type: "Json"; - /** - * Dataset description. - */ - description?: string; +export interface TriggerReference { /** - * Columns that define the structure of the dataset. Type: array (or Expression with resultType - * array), itemType: DatasetDataElement. + * Reference trigger name. */ - structure?: any; + referenceName: string; +} + +/** + * Contains the possible cases for TriggerDependencyReference. + */ +export type TriggerDependencyReferenceUnion = TriggerDependencyReference | TumblingWindowTriggerDependencyReference; + +/** + * Trigger referenced dependency. + */ +export interface TriggerDependencyReference { /** - * Columns that define the physical type schema of the dataset. Type: array (or Expression with - * resultType array), itemType: DatasetSchemaDataElement. + * Polymorphic Discriminator */ - schema?: any; + type: "TriggerDependencyReference"; /** - * Linked service reference. + * Referenced trigger. */ - linkedServiceName: LinkedServiceReference; + referenceTrigger: TriggerReference; +} + +/** + * Referenced tumbling window trigger dependency. + */ +export interface TumblingWindowTriggerDependencyReference { /** - * Parameters for dataset. + * Polymorphic Discriminator */ - parameters?: { [propertyName: string]: ParameterSpecification }; + type: "TumblingWindowTriggerDependencyReference"; /** - * List of tags that can be used for describing the Dataset. + * Referenced trigger. */ - annotations?: any[]; + referenceTrigger: TriggerReference; /** - * The folder that this Dataset is in. If not specified, Dataset will appear at the root level. + * Timespan applied to the start time of a tumbling window when evaluating dependency. */ - folder?: DatasetFolder; + offset?: string; /** - * The location of the json data storage. + * The size of the window when evaluating the dependency. If undefined the frequency of the + * tumbling window will be used. */ - location: DatasetLocationUnion; + size?: string; +} + +/** + * Execution policy for an activity. + */ +export interface RetryPolicy { /** - * The code page name of the preferred encoding. If not specified, the default value is UTF-8, - * unless BOM denotes another Unicode encoding. Refer to the name column of the table in the - * following link to set supported values: - * https://msdn.microsoft.com/library/system.text.encoding.aspx. Type: string (or Expression with - * resultType string). + * Maximum ordinary retry attempts. Default is 0. Type: integer (or Expression with resultType + * integer), minimum: 0. */ - encodingName?: any; + count?: any; /** - * The data compression method used for the json dataset. + * Interval between retries in seconds. Default is 30. */ - compression?: DatasetCompressionUnion; + intervalInSeconds?: number; } /** - * Delimited text dataset. + * Trigger that schedules pipeline runs for all fixed time interval windows from a start time + * without gaps and also supports backfill scenarios (when start time is in the past). */ -export interface DelimitedTextDataset { +export interface TumblingWindowTrigger { /** * Polymorphic Discriminator */ - type: "DelimitedText"; + type: "TumblingWindowTrigger"; /** - * Dataset description. + * Trigger description. */ description?: string; /** - * Columns that define the structure of the dataset. Type: array (or Expression with resultType - * array), itemType: DatasetDataElement. + * Indicates if trigger is running or not. Updated when Start/Stop APIs are called on the + * Trigger. Possible values include: 'Started', 'Stopped', 'Disabled' + * **NOTE: This property will not be serialized. It can only be populated by the server.** */ - structure?: any; + readonly runtimeState?: TriggerRuntimeState; /** - * Columns that define the physical type schema of the dataset. Type: array (or Expression with - * resultType array), itemType: DatasetSchemaDataElement. + * List of tags that can be used for describing the trigger. */ - schema?: any; + annotations?: any[]; /** - * Linked service reference. + * Pipeline for which runs are created when an event is fired for trigger window that is ready. */ - linkedServiceName: LinkedServiceReference; + pipelineProperty: TriggerPipelineReference; /** - * Parameters for dataset. + * The frequency of the time windows. Possible values include: 'Minute', 'Hour' */ - parameters?: { [propertyName: string]: ParameterSpecification }; + frequency: TumblingWindowFrequency; /** - * List of tags that can be used for describing the Dataset. + * The interval of the time windows. The minimum interval allowed is 15 Minutes. */ - annotations?: any[]; + interval: number; /** - * The folder that this Dataset is in. If not specified, Dataset will appear at the root level. + * The start time for the time period for the trigger during which events are fired for windows + * that are ready. Only UTC time is currently supported. */ - folder?: DatasetFolder; + startTime: Date; /** - * The location of the delimited text storage. + * The end time for the time period for the trigger during which events are fired for windows + * that are ready. Only UTC time is currently supported. */ - location: DatasetLocationUnion; + endTime?: Date; /** - * The column delimiter. Type: string (or Expression with resultType string). + * Specifies how long the trigger waits past due time before triggering new run. It doesn't alter + * window start and end time. The default is 0. Type: string (or Expression with resultType + * string), pattern: ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). */ - columnDelimiter?: any; + delay?: any; /** - * The row delimiter. Type: string (or Expression with resultType string). + * The max number of parallel time windows (ready for execution) for which a new run is + * triggered. */ - rowDelimiter?: any; + maxConcurrency: number; /** - * The code page name of the preferred encoding. If miss, the default value is UTF-8, unless BOM - * denotes another Unicode encoding. Refer to the name column of the table in the following link - * to set supported values: https://msdn.microsoft.com/library/system.text.encoding.aspx. Type: - * string (or Expression with resultType string). + * Retry policy that will be applied for failed pipeline runs. */ - encodingName?: any; - compressionCodec?: any; + retryPolicy?: RetryPolicy; /** - * The data compression method used for DelimitedText. + * Triggers that this trigger depends on. Only tumbling window triggers are supported. */ - compressionLevel?: any; + dependsOn?: DependencyReferenceUnion[]; +} + +/** + * Contains the possible cases for MultiplePipelineTrigger. + */ +export type MultiplePipelineTriggerUnion = MultiplePipelineTrigger | BlobEventsTrigger | BlobTrigger | ScheduleTrigger; + +/** + * Base class for all triggers that support one to many model for trigger to pipeline. + */ +export interface MultiplePipelineTrigger { /** - * The quote character. Type: string (or Expression with resultType string). + * Polymorphic Discriminator */ - quoteChar?: any; + type: "MultiplePipelineTrigger"; /** - * The escape character. Type: string (or Expression with resultType string). + * Trigger description. */ - escapeChar?: any; + description?: string; /** - * When used as input, treat the first row of data as headers. When used as output,write the - * headers into the output as the first row of data. The default value is false. Type: boolean - * (or Expression with resultType boolean). + * Indicates if trigger is running or not. Updated when Start/Stop APIs are called on the + * Trigger. Possible values include: 'Started', 'Stopped', 'Disabled' + * **NOTE: This property will not be serialized. It can only be populated by the server.** */ - firstRowAsHeader?: any; + readonly runtimeState?: TriggerRuntimeState; /** - * The null value string. Type: string (or Expression with resultType string). + * List of tags that can be used for describing the trigger. */ - nullValue?: any; + annotations?: any[]; + /** + * Pipelines that need to be started. + */ + pipelines?: TriggerPipelineReference[]; } /** - * Parquet dataset. + * Trigger that runs every time a Blob event occurs. */ -export interface ParquetDataset { +export interface BlobEventsTrigger { /** * Polymorphic Discriminator */ - type: "Parquet"; + type: "BlobEventsTrigger"; /** - * Dataset description. + * Trigger description. */ description?: string; /** - * Columns that define the structure of the dataset. Type: array (or Expression with resultType - * array), itemType: DatasetDataElement. + * Indicates if trigger is running or not. Updated when Start/Stop APIs are called on the + * Trigger. Possible values include: 'Started', 'Stopped', 'Disabled' + * **NOTE: This property will not be serialized. It can only be populated by the server.** */ - structure?: any; + readonly runtimeState?: TriggerRuntimeState; /** - * Columns that define the physical type schema of the dataset. Type: array (or Expression with - * resultType array), itemType: DatasetSchemaDataElement. + * List of tags that can be used for describing the trigger. */ - schema?: any; + annotations?: any[]; /** - * Linked service reference. + * Pipelines that need to be started. */ - linkedServiceName: LinkedServiceReference; + pipelines?: TriggerPipelineReference[]; /** - * Parameters for dataset. + * The blob path must begin with the pattern provided for trigger to fire. For example, + * '/records/blobs/december/' will only fire the trigger for blobs in the december folder under + * the records container. At least one of these must be provided: blobPathBeginsWith, + * blobPathEndsWith. */ - parameters?: { [propertyName: string]: ParameterSpecification }; + blobPathBeginsWith?: string; /** - * List of tags that can be used for describing the Dataset. + * The blob path must end with the pattern provided for trigger to fire. For example, + * 'december/boxes.csv' will only fire the trigger for blobs named boxes in a december folder. At + * least one of these must be provided: blobPathBeginsWith, blobPathEndsWith. */ - annotations?: any[]; + blobPathEndsWith?: string; /** - * The folder that this Dataset is in. If not specified, Dataset will appear at the root level. + * If set to true, blobs with zero bytes will be ignored. */ - folder?: DatasetFolder; + ignoreEmptyBlobs?: boolean; /** - * The location of the parquet storage. + * The type of events that cause this trigger to fire. */ - location: DatasetLocationUnion; - compressionCodec?: any; + events: BlobEventTypes[]; + /** + * The ARM resource ID of the Storage Account. + */ + scope: string; } /** - * Avro dataset. + * Trigger that runs every time the selected Blob container changes. */ -export interface AvroDataset { +export interface BlobTrigger { /** * Polymorphic Discriminator */ - type: "Avro"; + type: "BlobTrigger"; /** - * Dataset description. + * Trigger description. */ description?: string; /** - * Columns that define the structure of the dataset. Type: array (or Expression with resultType - * array), itemType: DatasetDataElement. + * Indicates if trigger is running or not. Updated when Start/Stop APIs are called on the + * Trigger. Possible values include: 'Started', 'Stopped', 'Disabled' + * **NOTE: This property will not be serialized. It can only be populated by the server.** */ - structure?: any; + readonly runtimeState?: TriggerRuntimeState; /** - * Columns that define the physical type schema of the dataset. Type: array (or Expression with - * resultType array), itemType: DatasetSchemaDataElement. + * List of tags that can be used for describing the trigger. */ - schema?: any; + annotations?: any[]; + /** + * Pipelines that need to be started. + */ + pipelines?: TriggerPipelineReference[]; /** - * Linked service reference. + * The path of the container/folder that will trigger the pipeline. */ - linkedServiceName: LinkedServiceReference; + folderPath: string; /** - * Parameters for dataset. + * The max number of parallel files to handle when it is triggered. */ - parameters?: { [propertyName: string]: ParameterSpecification }; + maxConcurrency: number; /** - * List of tags that can be used for describing the Dataset. + * The Azure Storage linked service reference. */ - annotations?: any[]; + linkedService: LinkedServiceReference; +} + +/** + * The recurrence schedule occurrence. + */ +export interface RecurrenceScheduleOccurrence { /** - * The folder that this Dataset is in. If not specified, Dataset will appear at the root level. + * The day of the week. Possible values include: 'Sunday', 'Monday', 'Tuesday', 'Wednesday', + * 'Thursday', 'Friday', 'Saturday' */ - folder?: DatasetFolder; + day?: DayOfWeek; /** - * The location of the avro storage. + * The occurrence. */ - location: DatasetLocationUnion; + occurrence?: number; /** - * Possible values include: 'none', 'deflate', 'snappy', 'xz', 'bzip2' + * Describes unknown properties. The value of an unknown property can be of "any" type. */ - avroCompressionCodec?: AvroCompressionCodec; - avroCompressionLevel?: number; + [property: string]: any; } /** - * A single Amazon Simple Storage Service (S3) object or a set of S3 objects. + * The recurrence schedule. */ -export interface AmazonS3Dataset { +export interface RecurrenceSchedule { /** - * Polymorphic Discriminator + * The minutes. */ - type: "AmazonS3Object"; + minutes?: number[]; /** - * Dataset description. + * The hours. */ - description?: string; + hours?: number[]; /** - * Columns that define the structure of the dataset. Type: array (or Expression with resultType - * array), itemType: DatasetDataElement. + * The days of the week. */ - structure?: any; + weekDays?: DaysOfWeek[]; /** - * Columns that define the physical type schema of the dataset. Type: array (or Expression with - * resultType array), itemType: DatasetSchemaDataElement. + * The month days. */ - schema?: any; + monthDays?: number[]; /** - * Linked service reference. + * The monthly occurrences. */ - linkedServiceName: LinkedServiceReference; + monthlyOccurrences?: RecurrenceScheduleOccurrence[]; /** - * Parameters for dataset. + * Describes unknown properties. The value of an unknown property can be of "any" type. */ - parameters?: { [propertyName: string]: ParameterSpecification }; + [property: string]: any; +} + +/** + * The workflow trigger recurrence. + */ +export interface ScheduleTriggerRecurrence { /** - * List of tags that can be used for describing the Dataset. + * The frequency. Possible values include: 'NotSpecified', 'Minute', 'Hour', 'Day', 'Week', + * 'Month', 'Year' */ - annotations?: any[]; + frequency?: RecurrenceFrequency; /** - * The folder that this Dataset is in. If not specified, Dataset will appear at the root level. + * The interval. */ - folder?: DatasetFolder; + interval?: number; /** - * The name of the Amazon S3 bucket. Type: string (or Expression with resultType string). + * The start time. */ - bucketName: any; + startTime?: Date; /** - * The key of the Amazon S3 object. Type: string (or Expression with resultType string). + * The end time. */ - key?: any; + endTime?: Date; /** - * The prefix filter for the S3 object name. Type: string (or Expression with resultType string). + * The time zone. */ - prefix?: any; + timeZone?: string; /** - * The version for the S3 object. Type: string (or Expression with resultType string). + * The recurrence schedule. */ - version?: any; + schedule?: RecurrenceSchedule; /** - * The start of S3 object's modified datetime. Type: string (or Expression with resultType - * string). + * Describes unknown properties. The value of an unknown property can be of "any" type. */ - modifiedDatetimeStart?: any; + [property: string]: any; +} + +/** + * Trigger that creates pipeline runs periodically, on schedule. + */ +export interface ScheduleTrigger { /** - * The end of S3 object's modified datetime. Type: string (or Expression with resultType string). + * Polymorphic Discriminator */ - modifiedDatetimeEnd?: any; + type: "ScheduleTrigger"; /** - * The format of files. + * Trigger description. */ - format?: DatasetStorageFormatUnion; + description?: string; /** - * The data compression method used for the Amazon S3 object. + * Indicates if trigger is running or not. Updated when Start/Stop APIs are called on the + * Trigger. Possible values include: 'Started', 'Stopped', 'Disabled' + * **NOTE: This property will not be serialized. It can only be populated by the server.** */ - compression?: DatasetCompressionUnion; + readonly runtimeState?: TriggerRuntimeState; + /** + * List of tags that can be used for describing the trigger. + */ + annotations?: any[]; + /** + * Pipelines that need to be started. + */ + pipelines?: TriggerPipelineReference[]; + /** + * Recurrence schedule configuration. + */ + recurrence: ScheduleTriggerRecurrence; } /** @@ -13804,6 +13822,20 @@ export interface Office365Source { outputColumns?: any; } +/** + * Specify the column name and value of additional columns. + */ +export interface AdditionalColumns { + /** + * Additional column name. Type: string (or Expression with resultType string). + */ + name?: any; + /** + * Additional column value. Type: string (or Expression with resultType string). + */ + value?: any; +} + /** * Cursor methods for Mongodb query */ @@ -13880,6 +13912,11 @@ export interface CosmosDbMongoDbApiSource { * ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). */ queryTimeout?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; } /** @@ -13926,6 +13963,11 @@ export interface MongoDbV2Source { * ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). */ queryTimeout?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; } /** @@ -13955,6 +13997,11 @@ export interface MongoDbSource { * resultType string). */ query?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; } /** @@ -13979,6 +14026,11 @@ export interface WebSource { * Expression with resultType integer). */ maxConcurrentConnections?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; } /** @@ -14046,6 +14098,11 @@ export interface OracleSource { * The settings that will be leveraged for Oracle source partitioning. */ partitionSettings?: OraclePartitionSettings; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; } /** @@ -14085,6 +14142,11 @@ export interface AzureDataExplorerSource { * ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])).. */ queryTimeout?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; } /** @@ -14168,6 +14230,11 @@ export interface FileSystemSource { * (or Expression with resultType boolean). */ recursive?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; } /** @@ -14222,6 +14289,11 @@ export interface RestSource { * The time to await before sending next page request. */ requestInterval?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; } /** @@ -14255,6 +14327,11 @@ export interface SalesforceServiceCloudSource { * 'QueryAll' */ readBehavior?: SalesforceSourceReadBehavior; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; } /** @@ -14283,6 +14360,11 @@ export interface ODataSource { * OData query. For example, "$top=1". Type: string (or Expression with resultType string). */ query?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; } /** @@ -14311,6 +14393,11 @@ export interface MicrosoftAccessSource { * Database query. Type: string (or Expression with resultType string). */ query?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; } /** @@ -14339,6 +14426,11 @@ export interface RelationalSource { * Database query. Type: string (or Expression with resultType string). */ query?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; } /** @@ -14368,6 +14460,11 @@ export interface CommonDataServiceForAppsSource { * Apps (online & on-premises). Type: string (or Expression with resultType string). */ query?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; } /** @@ -14397,6 +14494,11 @@ export interface DynamicsCrmSource { * on-premises). Type: string (or Expression with resultType string). */ query?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; } /** @@ -14426,6 +14528,11 @@ export interface DynamicsSource { * on-premises). Type: string (or Expression with resultType string). */ query?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; } /** @@ -14462,6 +14569,11 @@ export interface CosmosDbSqlApiSource { * Preferred regions. Type: array of strings (or Expression with resultType array of strings). */ preferredRegions?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; } /** @@ -14499,6 +14611,11 @@ export interface DocumentDbCollectionSource { * ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). */ queryTimeout?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; } /** @@ -14590,6 +14707,11 @@ export interface TabularSource { * ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). */ queryTimeout?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; } /** @@ -14619,6 +14741,11 @@ export interface AmazonRedshiftSource { * ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). */ queryTimeout?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; /** * Database query. Type: string (or Expression with resultType string). */ @@ -14658,6 +14785,11 @@ export interface GoogleAdWordsSource { * ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). */ queryTimeout?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; /** * A query to retrieve data from source. Type: string (or Expression with resultType string). */ @@ -14691,6 +14823,11 @@ export interface OracleServiceCloudSource { * ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). */ queryTimeout?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; /** * A query to retrieve data from source. Type: string (or Expression with resultType string). */ @@ -14724,6 +14861,11 @@ export interface DynamicsAXSource { * ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). */ queryTimeout?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; /** * A query to retrieve data from source. Type: string (or Expression with resultType string). */ @@ -14757,6 +14899,11 @@ export interface ResponsysSource { * ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). */ queryTimeout?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; /** * A query to retrieve data from source. Type: string (or Expression with resultType string). */ @@ -14790,6 +14937,11 @@ export interface SalesforceMarketingCloudSource { * ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). */ queryTimeout?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; /** * A query to retrieve data from source. Type: string (or Expression with resultType string). */ @@ -14823,6 +14975,11 @@ export interface VerticaSource { * ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). */ queryTimeout?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; /** * A query to retrieve data from source. Type: string (or Expression with resultType string). */ @@ -14877,6 +15034,11 @@ export interface NetezzaSource { * ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). */ queryTimeout?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; /** * A query to retrieve data from source. Type: string (or Expression with resultType string). */ @@ -14919,6 +15081,11 @@ export interface ZohoSource { * ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). */ queryTimeout?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; /** * A query to retrieve data from source. Type: string (or Expression with resultType string). */ @@ -14952,6 +15119,11 @@ export interface XeroSource { * ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). */ queryTimeout?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; /** * A query to retrieve data from source. Type: string (or Expression with resultType string). */ @@ -14985,6 +15157,11 @@ export interface SquareSource { * ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). */ queryTimeout?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; /** * A query to retrieve data from source. Type: string (or Expression with resultType string). */ @@ -15018,6 +15195,11 @@ export interface SparkSource { * ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). */ queryTimeout?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; /** * A query to retrieve data from source. Type: string (or Expression with resultType string). */ @@ -15051,6 +15233,11 @@ export interface ShopifySource { * ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). */ queryTimeout?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; /** * A query to retrieve data from source. Type: string (or Expression with resultType string). */ @@ -15084,6 +15271,11 @@ export interface ServiceNowSource { * ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). */ queryTimeout?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; /** * A query to retrieve data from source. Type: string (or Expression with resultType string). */ @@ -15117,6 +15309,11 @@ export interface QuickBooksSource { * ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). */ queryTimeout?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; /** * A query to retrieve data from source. Type: string (or Expression with resultType string). */ @@ -15150,6 +15347,11 @@ export interface PrestoSource { * ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). */ queryTimeout?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; /** * A query to retrieve data from source. Type: string (or Expression with resultType string). */ @@ -15183,6 +15385,11 @@ export interface PhoenixSource { * ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). */ queryTimeout?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; /** * A query to retrieve data from source. Type: string (or Expression with resultType string). */ @@ -15216,6 +15423,11 @@ export interface PaypalSource { * ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). */ queryTimeout?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; /** * A query to retrieve data from source. Type: string (or Expression with resultType string). */ @@ -15249,6 +15461,11 @@ export interface MarketoSource { * ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). */ queryTimeout?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; /** * A query to retrieve data from source. Type: string (or Expression with resultType string). */ @@ -15282,6 +15499,11 @@ export interface AzureMariaDBSource { * ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). */ queryTimeout?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; /** * A query to retrieve data from source. Type: string (or Expression with resultType string). */ @@ -15315,6 +15537,11 @@ export interface MariaDBSource { * ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). */ queryTimeout?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; /** * A query to retrieve data from source. Type: string (or Expression with resultType string). */ @@ -15348,6 +15575,11 @@ export interface MagentoSource { * ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). */ queryTimeout?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; /** * A query to retrieve data from source. Type: string (or Expression with resultType string). */ @@ -15381,6 +15613,11 @@ export interface JiraSource { * ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). */ queryTimeout?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; /** * A query to retrieve data from source. Type: string (or Expression with resultType string). */ @@ -15414,6 +15651,11 @@ export interface ImpalaSource { * ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). */ queryTimeout?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; /** * A query to retrieve data from source. Type: string (or Expression with resultType string). */ @@ -15447,6 +15689,11 @@ export interface HubspotSource { * ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). */ queryTimeout?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; /** * A query to retrieve data from source. Type: string (or Expression with resultType string). */ @@ -15480,6 +15727,11 @@ export interface HiveSource { * ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). */ queryTimeout?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; /** * A query to retrieve data from source. Type: string (or Expression with resultType string). */ @@ -15513,6 +15765,11 @@ export interface HBaseSource { * ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). */ queryTimeout?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; /** * A query to retrieve data from source. Type: string (or Expression with resultType string). */ @@ -15546,6 +15803,11 @@ export interface GreenplumSource { * ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). */ queryTimeout?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; /** * A query to retrieve data from source. Type: string (or Expression with resultType string). */ @@ -15579,6 +15841,11 @@ export interface GoogleBigQuerySource { * ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). */ queryTimeout?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; /** * A query to retrieve data from source. Type: string (or Expression with resultType string). */ @@ -15612,6 +15879,11 @@ export interface EloquaSource { * ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). */ queryTimeout?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; /** * A query to retrieve data from source. Type: string (or Expression with resultType string). */ @@ -15645,6 +15917,11 @@ export interface DrillSource { * ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). */ queryTimeout?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; /** * A query to retrieve data from source. Type: string (or Expression with resultType string). */ @@ -15678,6 +15955,11 @@ export interface CouchbaseSource { * ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). */ queryTimeout?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; /** * A query to retrieve data from source. Type: string (or Expression with resultType string). */ @@ -15711,6 +15993,11 @@ export interface ConcurSource { * ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). */ queryTimeout?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; /** * A query to retrieve data from source. Type: string (or Expression with resultType string). */ @@ -15744,6 +16031,11 @@ export interface AzurePostgreSqlSource { * ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). */ queryTimeout?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; /** * A query to retrieve data from source. Type: string (or Expression with resultType string). */ @@ -15777,6 +16069,11 @@ export interface AmazonMWSSource { * ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). */ queryTimeout?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; /** * A query to retrieve data from source. Type: string (or Expression with resultType string). */ @@ -15810,6 +16107,11 @@ export interface CassandraSource { * ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). */ queryTimeout?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; /** * Database query. Should be a SQL-92 query expression or Cassandra Query Language (CQL) command. * Type: string (or Expression with resultType string). @@ -15874,6 +16176,11 @@ export interface TeradataSource { * ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). */ queryTimeout?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; /** * Teradata query. Type: string (or Expression with resultType string). */ @@ -15916,6 +16223,11 @@ export interface AzureMySqlSource { * ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). */ queryTimeout?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; /** * Database query. Type: string (or Expression with resultType string). */ @@ -15949,6 +16261,11 @@ export interface SqlDWSource { * ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). */ queryTimeout?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; /** * SQL Data Warehouse reader query. Type: string (or Expression with resultType string). */ @@ -16008,6 +16325,11 @@ export interface SqlMISource { * ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). */ queryTimeout?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; /** * SQL reader query. Type: string (or Expression with resultType string). */ @@ -16055,6 +16377,11 @@ export interface AzureSqlSource { * ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). */ queryTimeout?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; /** * SQL reader query. Type: string (or Expression with resultType string). */ @@ -16102,6 +16429,11 @@ export interface SqlServerSource { * ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). */ queryTimeout?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; /** * SQL reader query. Type: string (or Expression with resultType string). */ @@ -16149,6 +16481,11 @@ export interface SqlSource { * ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). */ queryTimeout?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; /** * SQL reader query. Type: string (or Expression with resultType string). */ @@ -16163,6 +16500,12 @@ export interface SqlSource { * type: "int"}}". */ storedProcedureParameters?: { [propertyName: string]: StoredProcedureParameter }; + /** + * Specifies the transaction locking behavior for the SQL source. Allowed values: + * ReadCommitted/ReadUncommitted/RepeatableRead/Serializable/Snapshot. The default value is + * ReadCommitted. Type: string (or Expression with resultType string). + */ + isolationLevel?: any; } /** @@ -16218,6 +16561,11 @@ export interface SapTableSource { * ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). */ queryTimeout?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; /** * The number of rows to be retrieved. Type: integer(or Expression with resultType integer). */ @@ -16286,6 +16634,11 @@ export interface SapOpenHubSource { * ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). */ queryTimeout?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; /** * Whether to exclude the records of the last request. The default value is true. Type: boolean * (or Expression with resultType boolean). @@ -16337,6 +16690,11 @@ export interface SapHanaSource { * ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). */ queryTimeout?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; /** * SAP HANA Sql query. Type: string (or Expression with resultType string). */ @@ -16384,6 +16742,11 @@ export interface SapEccSource { * ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). */ queryTimeout?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; /** * SAP ECC OData query. For example, "$top=1". Type: string (or Expression with resultType * string). @@ -16418,6 +16781,11 @@ export interface SapCloudForCustomerSource { * ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). */ queryTimeout?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; /** * SAP Cloud for Customer OData query. For example, "$top=1". Type: string (or Expression with * resultType string). @@ -16452,6 +16820,11 @@ export interface SalesforceSource { * ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). */ queryTimeout?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; /** * Database query. Type: string (or Expression with resultType string). */ @@ -16490,6 +16863,11 @@ export interface SapBwSource { * ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). */ queryTimeout?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; /** * MDX query. Type: string (or Expression with resultType string). */ @@ -16523,6 +16901,11 @@ export interface SybaseSource { * ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). */ queryTimeout?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; /** * Database query. Type: string (or Expression with resultType string). */ @@ -16556,6 +16939,11 @@ export interface PostgreSqlSource { * ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). */ queryTimeout?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; /** * Database query. Type: string (or Expression with resultType string). */ @@ -16589,6 +16977,11 @@ export interface MySqlSource { * ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). */ queryTimeout?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; /** * Database query. Type: string (or Expression with resultType string). */ @@ -16622,6 +17015,11 @@ export interface OdbcSource { * ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). */ queryTimeout?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; /** * Database query. Type: string (or Expression with resultType string). */ @@ -16655,6 +17053,11 @@ export interface Db2Source { * ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). */ queryTimeout?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; /** * Database query. Type: string (or Expression with resultType string). */ @@ -16688,6 +17091,11 @@ export interface InformixSource { * ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). */ queryTimeout?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; /** * Database query. Type: string (or Expression with resultType string). */ @@ -16721,6 +17129,11 @@ export interface AzureTableSource { * ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). */ queryTimeout?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; /** * Azure Table source query. Type: string (or Expression with resultType string). */ @@ -16782,6 +17195,11 @@ export interface HdfsReadSettings { * HDFS wildcardFileName. Type: string (or Expression with resultType string). */ wildcardFileName?: any; + /** + * Point to a text file that lists each file (relative path to the path configured in the + * dataset) that you want to copy. Type: string (or Expression with resultType string). + */ + fileListPath?: any; /** * Indicates whether to enable partition discovery. */ @@ -16860,6 +17278,11 @@ export interface SftpReadSettings { * Sftp wildcardFileName. Type: string (or Expression with resultType string). */ wildcardFileName?: any; + /** + * Point to a text file that lists each file (relative path to the path configured in the + * dataset) that you want to copy. Type: string (or Expression with resultType string). + */ + fileListPath?: any; /** * The start of file's modified datetime. Type: string (or Expression with resultType string). */ @@ -16896,6 +17319,11 @@ export interface FtpReadSettings { * Ftp wildcardFileName. Type: string (or Expression with resultType string). */ wildcardFileName?: any; + /** + * Point to a text file that lists each file (relative path to the path configured in the + * dataset) that you want to copy. Type: string (or Expression with resultType string). + */ + fileListPath?: any; /** * Specify whether to use binary transfer mode for FTP stores. */ @@ -16933,6 +17361,11 @@ export interface GoogleCloudStorageReadSettings { * resultType string). */ prefix?: any; + /** + * Point to a text file that lists each file (relative path to the path configured in the + * dataset) that you want to copy. Type: string (or Expression with resultType string). + */ + fileListPath?: any; /** * Indicates whether to enable partition discovery. */ @@ -16973,6 +17406,11 @@ export interface AzureFileStorageReadSettings { * Azure File Storage wildcardFileName. Type: string (or Expression with resultType string). */ wildcardFileName?: any; + /** + * Point to a text file that lists each file (relative path to the path configured in the + * dataset) that you want to copy. Type: string (or Expression with resultType string). + */ + fileListPath?: any; /** * Indicates whether to enable partition discovery. */ @@ -17013,6 +17451,11 @@ export interface FileServerReadSettings { * FileServer wildcardFileName. Type: string (or Expression with resultType string). */ wildcardFileName?: any; + /** + * Point to a text file that lists each file (relative path to the path configured in the + * dataset) that you want to copy. Type: string (or Expression with resultType string). + */ + fileListPath?: any; /** * Indicates whether to enable partition discovery. */ @@ -17057,6 +17500,11 @@ export interface AmazonS3ReadSettings { * The prefix filter for the S3 object name. Type: string (or Expression with resultType string). */ prefix?: any; + /** + * Point to a text file that lists each file (relative path to the path configured in the + * dataset) that you want to copy. Type: string (or Expression with resultType string). + */ + fileListPath?: any; /** * Indicates whether to enable partition discovery. */ @@ -17097,6 +17545,11 @@ export interface AzureDataLakeStoreReadSettings { * ADLS wildcardFileName. Type: string (or Expression with resultType string). */ wildcardFileName?: any; + /** + * Point to a text file that lists each file (relative path to the path configured in the + * dataset) that you want to copy. Type: string (or Expression with resultType string). + */ + fileListPath?: any; /** * Indicates whether to enable partition discovery. */ @@ -17137,6 +17590,11 @@ export interface AzureBlobFSReadSettings { * Azure blobFS wildcardFileName. Type: string (or Expression with resultType string). */ wildcardFileName?: any; + /** + * Point to a text file that lists each file (relative path to the path configured in the + * dataset) that you want to copy. Type: string (or Expression with resultType string). + */ + fileListPath?: any; /** * Indicates whether to enable partition discovery. */ @@ -17182,6 +17640,11 @@ export interface AzureBlobStorageReadSettings { * string). */ prefix?: any; + /** + * Point to a text file that lists each file (relative path to the path configured in the + * dataset) that you want to copy. Type: string (or Expression with resultType string). + */ + fileListPath?: any; /** * Indicates whether to enable partition discovery. */ @@ -17250,6 +17713,11 @@ export interface OrcSource { * ORC store settings. */ storeSettings?: StoreReadSettingsUnion; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; } /** @@ -17278,6 +17746,11 @@ export interface JsonSource { * Json store settings. */ storeSettings?: StoreReadSettingsUnion; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; } /** @@ -17344,6 +17817,11 @@ export interface DelimitedTextSource { * DelimitedText format settings. */ formatSettings?: DelimitedTextReadSettings; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; } /** @@ -17372,6 +17850,11 @@ export interface ParquetSource { * Parquet store settings. */ storeSettings?: StoreReadSettingsUnion; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; } /** @@ -17400,6 +17883,11 @@ export interface AvroSource { * Avro store settings. */ storeSettings?: StoreReadSettingsUnion; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; } /** @@ -18261,6 +18749,22 @@ export interface HDInsightHiveActivity { queryTimeout?: number; } +/** + * Skip error file. + */ +export interface SkipErrorFile { + /** + * Skip if file is deleted by other client during copy. Default is true. Type: boolean (or + * Expression with resultType boolean). + */ + fileMissing?: any; + /** + * Skip if source/sink file changed by other concurrent write. Default is false. Type: boolean + * (or Expression with resultType boolean). + */ + dataInconsistency?: any; +} + /** * Redirect incompatible row settings */ @@ -19523,6 +20027,12 @@ export interface AzureDataLakeStoreWriteSettings { * The type of copy behavior for copy sink. */ copyBehavior?: any; + /** + * Specifies the expiry time of the written files. The time is applied to the UTC time zone in + * the format of "2018-12-01T05:00:00Z". Default value is NULL. Type: integer (or Expression with + * resultType integer). + */ + expiryDateTime?: any; } /** @@ -19595,6 +20105,11 @@ export interface SftpWriteSettings { * hour). Type: string (or Expression with resultType string). */ operationTimeout?: any; + /** + * Upload to temporary file(s) and rename. Disable this option if your SFTP server doesn't + * support rename operation. Type: boolean (or Expression with resultType boolean). + */ + useTempFileRename?: any; } /** @@ -20172,6 +20687,10 @@ export interface CopyActivity { * Redirect incompatible row settings when EnableSkipIncompatibleRow is true. */ redirectIncompatibleRowSettings?: RedirectIncompatibleRowSettings; + /** + * Log storage settings customer need to provide when enabling session log. + */ + logStorageSettings?: LogStorageSettings; /** * Preserve Rules. */ @@ -20180,6 +20699,15 @@ export interface CopyActivity { * Preserve rules. */ preserve?: any[]; + /** + * Whether to enable Data Consistency validation. Type: boolean (or Expression with resultType + * boolean). + */ + validateDataConsistency?: any; + /** + * Specify the fault tolerance for data consistency. + */ + skipErrorFile?: SkipErrorFile; /** * List of inputs for the activity. */ @@ -21975,6 +22503,11 @@ export interface PipelinesCreateRunOptionalParams extends msRest.RequestOptionsB * will run. */ startActivityName?: string; + /** + * In recovery mode, if set to true, the rerun will start from failed activities. The property + * will be used only if startActivityName is not specified. + */ + startFromFailure?: boolean; /** * Parameters of the pipeline run. These parameters will be used only if the runId is not * specified. @@ -22147,20 +22680,6 @@ export interface TriggerListResponse extends Array { nextLink?: string; } -/** - * @interface - * A list of rerun triggers. - * @extends Array - */ -export interface RerunTriggerListResponse extends Array { - /** - * The continuation token for getting the next page of results, if any remaining results exist, - * null otherwise. - * **NOTE: This property will not be serialized. It can only be populated by the server.** - */ - readonly nextLink?: string; -} - /** * @interface * A list of data flow resources. @@ -22294,48 +22813,6 @@ export type TriggerRunStatus = 'Succeeded' | 'Failed' | 'Inprogress'; */ export type DataFlowDebugCommandType = 'executePreviewQuery' | 'executeStatisticsQuery' | 'executeExpressionQuery'; -/** - * Defines values for TumblingWindowFrequency. - * Possible values include: 'Minute', 'Hour' - * @readonly - * @enum {string} - */ -export type TumblingWindowFrequency = 'Minute' | 'Hour'; - -/** - * Defines values for BlobEventTypes. - * Possible values include: 'Microsoft.Storage.BlobCreated', 'Microsoft.Storage.BlobDeleted' - * @readonly - * @enum {string} - */ -export type BlobEventTypes = 'Microsoft.Storage.BlobCreated' | 'Microsoft.Storage.BlobDeleted'; - -/** - * Defines values for DayOfWeek. - * Possible values include: 'Sunday', 'Monday', 'Tuesday', 'Wednesday', 'Thursday', 'Friday', - * 'Saturday' - * @readonly - * @enum {string} - */ -export type DayOfWeek = 'Sunday' | 'Monday' | 'Tuesday' | 'Wednesday' | 'Thursday' | 'Friday' | 'Saturday'; - -/** - * Defines values for DaysOfWeek. - * Possible values include: 'Sunday', 'Monday', 'Tuesday', 'Wednesday', 'Thursday', 'Friday', - * 'Saturday' - * @readonly - * @enum {string} - */ -export type DaysOfWeek = 'Sunday' | 'Monday' | 'Tuesday' | 'Wednesday' | 'Thursday' | 'Friday' | 'Saturday'; - -/** - * Defines values for RecurrenceFrequency. - * Possible values include: 'NotSpecified', 'Minute', 'Hour', 'Day', 'Week', 'Month', 'Year' - * @readonly - * @enum {string} - */ -export type RecurrenceFrequency = 'NotSpecified' | 'Minute' | 'Hour' | 'Day' | 'Week' | 'Month' | 'Year'; - /** * Defines values for GoogleAdWordsAuthenticationType. * Possible values include: 'ServiceAuthentication', 'UserAuthentication' @@ -22563,6 +23040,48 @@ export type OrcCompressionCodec = 'none' | 'zlib' | 'snappy'; */ export type AvroCompressionCodec = 'none' | 'deflate' | 'snappy' | 'xz' | 'bzip2'; +/** + * Defines values for TumblingWindowFrequency. + * Possible values include: 'Minute', 'Hour' + * @readonly + * @enum {string} + */ +export type TumblingWindowFrequency = 'Minute' | 'Hour'; + +/** + * Defines values for BlobEventTypes. + * Possible values include: 'Microsoft.Storage.BlobCreated', 'Microsoft.Storage.BlobDeleted' + * @readonly + * @enum {string} + */ +export type BlobEventTypes = 'Microsoft.Storage.BlobCreated' | 'Microsoft.Storage.BlobDeleted'; + +/** + * Defines values for DayOfWeek. + * Possible values include: 'Sunday', 'Monday', 'Tuesday', 'Wednesday', 'Thursday', 'Friday', + * 'Saturday' + * @readonly + * @enum {string} + */ +export type DayOfWeek = 'Sunday' | 'Monday' | 'Tuesday' | 'Wednesday' | 'Thursday' | 'Friday' | 'Saturday'; + +/** + * Defines values for DaysOfWeek. + * Possible values include: 'Sunday', 'Monday', 'Tuesday', 'Wednesday', 'Thursday', 'Friday', + * 'Saturday' + * @readonly + * @enum {string} + */ +export type DaysOfWeek = 'Sunday' | 'Monday' | 'Tuesday' | 'Wednesday' | 'Thursday' | 'Friday' | 'Saturday'; + +/** + * Defines values for RecurrenceFrequency. + * Possible values include: 'NotSpecified', 'Minute', 'Hour', 'Day', 'Week', 'Month', 'Year' + * @readonly + * @enum {string} + */ +export type RecurrenceFrequency = 'NotSpecified' | 'Minute' | 'Hour' | 'Day' | 'Week' | 'Month' | 'Year'; + /** * Defines values for DataFlowComputeType. * Possible values include: 'General', 'MemoryOptimized', 'ComputeOptimized' @@ -23806,6 +24325,26 @@ export type TriggersListByFactoryResponse = TriggerListResponse & { }; }; +/** + * Contains response data for the queryByFactory operation. + */ +export type TriggersQueryByFactoryResponse = TriggerQueryResponse & { + /** + * The underlying HTTP response. + */ + _response: msRest.HttpResponse & { + /** + * The response body as text (string format) + */ + bodyAsText: string; + + /** + * The response body as parsed JSON or XML + */ + parsedBody: TriggerQueryResponse; + }; +}; + /** * Contains response data for the createOrUpdate operation. */ @@ -23986,66 +24525,6 @@ export type TriggerRunsQueryByFactoryResponse = TriggerRunsQueryResponse & { }; }; -/** - * Contains response data for the create operation. - */ -export type RerunTriggersCreateResponse = TriggerResource & { - /** - * The underlying HTTP response. - */ - _response: msRest.HttpResponse & { - /** - * The response body as text (string format) - */ - bodyAsText: string; - - /** - * The response body as parsed JSON or XML - */ - parsedBody: TriggerResource; - }; -}; - -/** - * Contains response data for the listByTrigger operation. - */ -export type RerunTriggersListByTriggerResponse = RerunTriggerListResponse & { - /** - * The underlying HTTP response. - */ - _response: msRest.HttpResponse & { - /** - * The response body as text (string format) - */ - bodyAsText: string; - - /** - * The response body as parsed JSON or XML - */ - parsedBody: RerunTriggerListResponse; - }; -}; - -/** - * Contains response data for the listByTriggerNext operation. - */ -export type RerunTriggersListByTriggerNextResponse = RerunTriggerListResponse & { - /** - * The underlying HTTP response. - */ - _response: msRest.HttpResponse & { - /** - * The response body as text (string format) - */ - bodyAsText: string; - - /** - * The response body as parsed JSON or XML - */ - parsedBody: RerunTriggerListResponse; - }; -}; - /** * Contains response data for the createOrUpdate operation. */ diff --git a/sdk/datafactory/arm-datafactory/src/models/integrationRuntimesMappers.ts b/sdk/datafactory/arm-datafactory/src/models/integrationRuntimesMappers.ts index 467b87ecd36d..316683c70dbc 100644 --- a/sdk/datafactory/arm-datafactory/src/models/integrationRuntimesMappers.ts +++ b/sdk/datafactory/arm-datafactory/src/models/integrationRuntimesMappers.ts @@ -11,6 +11,7 @@ export { Activity, ActivityDependency, ActivityPolicy, + AdditionalColumns, AmazonMWSLinkedService, AmazonMWSObjectDataset, AmazonMWSSource, @@ -388,7 +389,6 @@ export { RedshiftUnloadSettings, RelationalSource, RelationalTableDataset, - RerunTriggerResource, RerunTumblingWindowTrigger, Resource, ResponsysLinkedService, @@ -450,6 +450,7 @@ export { ShopifyLinkedService, ShopifyObjectDataset, ShopifySource, + SkipErrorFile, SparkLinkedService, SparkObjectDataset, SparkSource, diff --git a/sdk/datafactory/arm-datafactory/src/models/linkedServicesMappers.ts b/sdk/datafactory/arm-datafactory/src/models/linkedServicesMappers.ts index f076eff667ce..a429fa7af963 100644 --- a/sdk/datafactory/arm-datafactory/src/models/linkedServicesMappers.ts +++ b/sdk/datafactory/arm-datafactory/src/models/linkedServicesMappers.ts @@ -11,6 +11,7 @@ export { Activity, ActivityDependency, ActivityPolicy, + AdditionalColumns, AmazonMWSLinkedService, AmazonMWSObjectDataset, AmazonMWSSource, @@ -374,7 +375,6 @@ export { RedshiftUnloadSettings, RelationalSource, RelationalTableDataset, - RerunTriggerResource, RerunTumblingWindowTrigger, Resource, ResponsysLinkedService, @@ -434,6 +434,7 @@ export { ShopifyLinkedService, ShopifyObjectDataset, ShopifySource, + SkipErrorFile, SparkLinkedService, SparkObjectDataset, SparkSource, diff --git a/sdk/datafactory/arm-datafactory/src/models/mappers.ts b/sdk/datafactory/arm-datafactory/src/models/mappers.ts index e59bd06d97f9..5ceddc7199c9 100644 --- a/sdk/datafactory/arm-datafactory/src/models/mappers.ts +++ b/sdk/datafactory/arm-datafactory/src/models/mappers.ts @@ -1289,6 +1289,35 @@ export const TriggerResource: msRest.CompositeMapper = { } }; +export const TriggerQueryResponse: msRest.CompositeMapper = { + serializedName: "TriggerQueryResponse", + type: { + name: "Composite", + className: "TriggerQueryResponse", + modelProperties: { + value: { + required: true, + serializedName: "value", + type: { + name: "Sequence", + element: { + type: { + name: "Composite", + className: "TriggerResource" + } + } + } + }, + continuationToken: { + serializedName: "continuationToken", + type: { + name: "String" + } + } + } + } +}; + export const CreateRunResponse: msRest.CompositeMapper = { serializedName: "CreateRunResponse", type: { @@ -1688,6 +1717,28 @@ export const LinkedServiceDebugResource: msRest.CompositeMapper = { } }; +export const TriggerFilterParameters: msRest.CompositeMapper = { + serializedName: "TriggerFilterParameters", + type: { + name: "Composite", + className: "TriggerFilterParameters", + modelProperties: { + continuationToken: { + serializedName: "continuationToken", + type: { + name: "String" + } + }, + parentTriggerName: { + serializedName: "parentTriggerName", + type: { + name: "String" + } + } + } + } +}; + export const RunQueryFilter: msRest.CompositeMapper = { serializedName: "RunQueryFilter", type: { @@ -2263,106 +2314,6 @@ export const TriggerRunsQueryResponse: msRest.CompositeMapper = { } }; -export const RerunTumblingWindowTriggerActionParameters: msRest.CompositeMapper = { - serializedName: "RerunTumblingWindowTriggerActionParameters", - type: { - name: "Composite", - className: "RerunTumblingWindowTriggerActionParameters", - modelProperties: { - startTime: { - required: true, - serializedName: "startTime", - type: { - name: "DateTime" - } - }, - endTime: { - required: true, - serializedName: "endTime", - type: { - name: "DateTime" - } - }, - maxConcurrency: { - required: true, - serializedName: "maxConcurrency", - constraints: { - InclusiveMaximum: 50, - InclusiveMinimum: 1 - }, - type: { - name: "Number" - } - } - } - } -}; - -export const RerunTumblingWindowTrigger: msRest.CompositeMapper = { - serializedName: "RerunTumblingWindowTrigger", - type: { - name: "Composite", - polymorphicDiscriminator: Trigger.type.polymorphicDiscriminator, - uberParent: "Trigger", - className: "RerunTumblingWindowTrigger", - modelProperties: { - ...Trigger.type.modelProperties, - parentTrigger: { - serializedName: "typeProperties.parentTrigger", - type: { - name: "Object" - } - }, - requestedStartTime: { - required: true, - serializedName: "typeProperties.requestedStartTime", - type: { - name: "DateTime" - } - }, - requestedEndTime: { - required: true, - serializedName: "typeProperties.requestedEndTime", - type: { - name: "DateTime" - } - }, - maxConcurrency: { - required: true, - serializedName: "typeProperties.maxConcurrency", - constraints: { - InclusiveMaximum: 50, - InclusiveMinimum: 1 - }, - type: { - name: "Number" - } - } - }, - additionalProperties: Trigger.type.additionalProperties - } -}; - -export const RerunTriggerResource: msRest.CompositeMapper = { - serializedName: "RerunTriggerResource", - type: { - name: "Composite", - className: "RerunTriggerResource", - modelProperties: { - ...SubResource.type.modelProperties, - properties: { - required: true, - serializedName: "properties", - type: { - name: "Composite", - className: "RerunTumblingWindowTrigger", - additionalProperties: Trigger.type.additionalProperties - } - } - } - } -}; - export const OperationDisplay: msRest.CompositeMapper = { serializedName: "Operation_display", type: { @@ -3404,648 +3355,396 @@ export const MappingDataFlow: msRest.CompositeMapper = { } }; -export const ChainingTrigger: msRest.CompositeMapper = { - serializedName: "ChainingTrigger", +export const AzureFunctionLinkedService: msRest.CompositeMapper = { + serializedName: "AzureFunction", type: { name: "Composite", - polymorphicDiscriminator: Trigger.type.polymorphicDiscriminator, - uberParent: "Trigger", - className: "ChainingTrigger", + polymorphicDiscriminator: LinkedService.type.polymorphicDiscriminator, + uberParent: "LinkedService", + className: "AzureFunctionLinkedService", modelProperties: { - ...Trigger.type.modelProperties, - pipelineProperty: { + ...LinkedService.type.modelProperties, + functionAppUrl: { required: true, - serializedName: "pipeline", - defaultValue: {}, + serializedName: "typeProperties.functionAppUrl", type: { - name: "Composite", - className: "TriggerPipelineReference" + name: "Object" } }, - dependsOn: { - required: true, - serializedName: "typeProperties.dependsOn", + functionKey: { + serializedName: "typeProperties.functionKey", type: { - name: "Sequence", - element: { - type: { - name: "Composite", - className: "PipelineReference" - } - } + name: "Composite", + className: "SecretBase" } }, - runDimension: { - required: true, - serializedName: "typeProperties.runDimension", + encryptedCredential: { + serializedName: "typeProperties.encryptedCredential", type: { - name: "String" + name: "Object" } } }, - additionalProperties: Trigger.type.additionalProperties + additionalProperties: LinkedService.type.additionalProperties } }; -export const DependencyReference: msRest.CompositeMapper = { - serializedName: "DependencyReference", +export const AzureDataExplorerLinkedService: msRest.CompositeMapper = { + serializedName: "AzureDataExplorer", type: { name: "Composite", - polymorphicDiscriminator: { - serializedName: "type", - clientName: "type" - }, - uberParent: "DependencyReference", - className: "DependencyReference", + polymorphicDiscriminator: LinkedService.type.polymorphicDiscriminator, + uberParent: "LinkedService", + className: "AzureDataExplorerLinkedService", modelProperties: { - type: { + ...LinkedService.type.modelProperties, + endpoint: { required: true, - serializedName: "type", + serializedName: "typeProperties.endpoint", type: { - name: "String" + name: "Object" } - } - } - } -}; - -export const SelfDependencyTumblingWindowTriggerReference: msRest.CompositeMapper = { - serializedName: "SelfDependencyTumblingWindowTriggerReference", - type: { - name: "Composite", - polymorphicDiscriminator: DependencyReference.type.polymorphicDiscriminator, - uberParent: "DependencyReference", - className: "SelfDependencyTumblingWindowTriggerReference", - modelProperties: { - ...DependencyReference.type.modelProperties, - offset: { + }, + servicePrincipalId: { required: true, - serializedName: "offset", - constraints: { - MaxLength: 15, - MinLength: 8, - Pattern: /-((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9]))/ - }, + serializedName: "typeProperties.servicePrincipalId", type: { - name: "String" + name: "Object" } }, - size: { - serializedName: "size", - constraints: { - MaxLength: 15, - MinLength: 8, - Pattern: /((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9]))/ - }, + servicePrincipalKey: { + required: true, + serializedName: "typeProperties.servicePrincipalKey", type: { - name: "String" + name: "Composite", + className: "SecretBase" } - } - } - } -}; - -export const TriggerReference: msRest.CompositeMapper = { - serializedName: "TriggerReference", - type: { - name: "Composite", - className: "TriggerReference", - modelProperties: { - type: { + }, + database: { required: true, - isConstant: true, - serializedName: "type", - defaultValue: 'TriggerReference', + serializedName: "typeProperties.database", type: { - name: "String" + name: "Object" } }, - referenceName: { + tenant: { required: true, - serializedName: "referenceName", + serializedName: "typeProperties.tenant", type: { - name: "String" + name: "Object" } } - } + }, + additionalProperties: LinkedService.type.additionalProperties } }; -export const TriggerDependencyReference: msRest.CompositeMapper = { - serializedName: "TriggerDependencyReference", +export const SapTableLinkedService: msRest.CompositeMapper = { + serializedName: "SapTable", type: { name: "Composite", - polymorphicDiscriminator: DependencyReference.type.polymorphicDiscriminator, - uberParent: "DependencyReference", - className: "TriggerDependencyReference", + polymorphicDiscriminator: LinkedService.type.polymorphicDiscriminator, + uberParent: "LinkedService", + className: "SapTableLinkedService", modelProperties: { - ...DependencyReference.type.modelProperties, - referenceTrigger: { - required: true, - serializedName: "referenceTrigger", - defaultValue: {}, + ...LinkedService.type.modelProperties, + server: { + serializedName: "typeProperties.server", type: { - name: "Composite", - className: "TriggerReference" + name: "Object" } - } - } - } -}; - -export const TumblingWindowTriggerDependencyReference: msRest.CompositeMapper = { - serializedName: "TumblingWindowTriggerDependencyReference", - type: { - name: "Composite", - polymorphicDiscriminator: DependencyReference.type.polymorphicDiscriminator, - uberParent: "DependencyReference", - className: "TumblingWindowTriggerDependencyReference", - modelProperties: { - ...TriggerDependencyReference.type.modelProperties, - offset: { - serializedName: "offset", - constraints: { - MaxLength: 15, - MinLength: 8, - Pattern: /-?((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9]))/ - }, + }, + systemNumber: { + serializedName: "typeProperties.systemNumber", type: { - name: "String" + name: "Object" } }, - size: { - serializedName: "size", - constraints: { - MaxLength: 15, - MinLength: 8, - Pattern: /((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9]))/ - }, + clientId: { + serializedName: "typeProperties.clientId", type: { - name: "String" + name: "Object" } - } - } - } -}; - -export const RetryPolicy: msRest.CompositeMapper = { - serializedName: "RetryPolicy", - type: { - name: "Composite", - className: "RetryPolicy", - modelProperties: { - count: { - serializedName: "count", + }, + language: { + serializedName: "typeProperties.language", type: { name: "Object" } }, - intervalInSeconds: { - serializedName: "intervalInSeconds", - constraints: { - InclusiveMaximum: 86400, - InclusiveMinimum: 30 - }, + systemId: { + serializedName: "typeProperties.systemId", type: { - name: "Number" + name: "Object" } - } - } - } -}; - -export const TumblingWindowTrigger: msRest.CompositeMapper = { - serializedName: "TumblingWindowTrigger", - type: { - name: "Composite", - polymorphicDiscriminator: Trigger.type.polymorphicDiscriminator, - uberParent: "Trigger", - className: "TumblingWindowTrigger", - modelProperties: { - ...Trigger.type.modelProperties, - pipelineProperty: { - required: true, - serializedName: "pipeline", - defaultValue: {}, + }, + userName: { + serializedName: "typeProperties.userName", + type: { + name: "Object" + } + }, + password: { + serializedName: "typeProperties.password", type: { name: "Composite", - className: "TriggerPipelineReference" + className: "SecretBase" } }, - frequency: { - required: true, - serializedName: "typeProperties.frequency", + messageServer: { + serializedName: "typeProperties.messageServer", type: { - name: "String" + name: "Object" } }, - interval: { - required: true, - serializedName: "typeProperties.interval", + messageServerService: { + serializedName: "typeProperties.messageServerService", type: { - name: "Number" + name: "Object" } }, - startTime: { - required: true, - serializedName: "typeProperties.startTime", + sncMode: { + serializedName: "typeProperties.sncMode", type: { - name: "DateTime" + name: "Object" } }, - endTime: { - serializedName: "typeProperties.endTime", + sncMyName: { + serializedName: "typeProperties.sncMyName", type: { - name: "DateTime" + name: "Object" } }, - delay: { - serializedName: "typeProperties.delay", + sncPartnerName: { + serializedName: "typeProperties.sncPartnerName", type: { name: "Object" } }, - maxConcurrency: { - required: true, - serializedName: "typeProperties.maxConcurrency", - constraints: { - InclusiveMaximum: 50, - InclusiveMinimum: 1 - }, + sncLibraryPath: { + serializedName: "typeProperties.sncLibraryPath", type: { - name: "Number" + name: "Object" } }, - retryPolicy: { - serializedName: "typeProperties.retryPolicy", + sncQop: { + serializedName: "typeProperties.sncQop", type: { - name: "Composite", - className: "RetryPolicy" + name: "Object" } }, - dependsOn: { - serializedName: "typeProperties.dependsOn", + logonGroup: { + serializedName: "typeProperties.logonGroup", type: { - name: "Sequence", - element: { - type: { - name: "Composite", - className: "DependencyReference" - } - } + name: "Object" } - } - }, - additionalProperties: Trigger.type.additionalProperties - } -}; - -export const MultiplePipelineTrigger: msRest.CompositeMapper = { - serializedName: "MultiplePipelineTrigger", - type: { - name: "Composite", - polymorphicDiscriminator: Trigger.type.polymorphicDiscriminator, - uberParent: "Trigger", - className: "MultiplePipelineTrigger", - modelProperties: { - ...Trigger.type.modelProperties, - pipelines: { - serializedName: "pipelines", + }, + encryptedCredential: { + serializedName: "typeProperties.encryptedCredential", type: { - name: "Sequence", - element: { - type: { - name: "Composite", - className: "TriggerPipelineReference" - } - } + name: "Object" } } }, - additionalProperties: Trigger.type.additionalProperties + additionalProperties: LinkedService.type.additionalProperties } }; -export const BlobEventsTrigger: msRest.CompositeMapper = { - serializedName: "BlobEventsTrigger", +export const GoogleAdWordsLinkedService: msRest.CompositeMapper = { + serializedName: "GoogleAdWords", type: { name: "Composite", - polymorphicDiscriminator: Trigger.type.polymorphicDiscriminator, - uberParent: "Trigger", - className: "BlobEventsTrigger", + polymorphicDiscriminator: LinkedService.type.polymorphicDiscriminator, + uberParent: "LinkedService", + className: "GoogleAdWordsLinkedService", modelProperties: { - ...MultiplePipelineTrigger.type.modelProperties, - blobPathBeginsWith: { - serializedName: "typeProperties.blobPathBeginsWith", + ...LinkedService.type.modelProperties, + clientCustomerID: { + required: true, + serializedName: "typeProperties.clientCustomerID", type: { - name: "String" + name: "Object" } }, - blobPathEndsWith: { - serializedName: "typeProperties.blobPathEndsWith", + developerToken: { + required: true, + serializedName: "typeProperties.developerToken", type: { - name: "String" + name: "Composite", + className: "SecretBase" } }, - ignoreEmptyBlobs: { - serializedName: "typeProperties.ignoreEmptyBlobs", + authenticationType: { + required: true, + serializedName: "typeProperties.authenticationType", type: { - name: "Boolean" + name: "String" } }, - events: { - required: true, - serializedName: "typeProperties.events", + refreshToken: { + serializedName: "typeProperties.refreshToken", type: { - name: "Sequence", - element: { - type: { - name: "String" - } - } + name: "Composite", + className: "SecretBase" } }, - scope: { - required: true, - serializedName: "typeProperties.scope", + clientId: { + serializedName: "typeProperties.clientId", type: { - name: "String" + name: "Object" } - } - }, - additionalProperties: Trigger.type.additionalProperties - } -}; - -export const BlobTrigger: msRest.CompositeMapper = { - serializedName: "BlobTrigger", - type: { - name: "Composite", - polymorphicDiscriminator: Trigger.type.polymorphicDiscriminator, - uberParent: "Trigger", - className: "BlobTrigger", - modelProperties: { - ...MultiplePipelineTrigger.type.modelProperties, - folderPath: { - required: true, - serializedName: "typeProperties.folderPath", + }, + clientSecret: { + serializedName: "typeProperties.clientSecret", type: { - name: "String" + name: "Composite", + className: "SecretBase" } }, - maxConcurrency: { - required: true, - serializedName: "typeProperties.maxConcurrency", + email: { + serializedName: "typeProperties.email", type: { - name: "Number" + name: "Object" } }, - linkedService: { - required: true, - serializedName: "typeProperties.linkedService", - defaultValue: {}, + keyFilePath: { + serializedName: "typeProperties.keyFilePath", type: { - name: "Composite", - className: "LinkedServiceReference" + name: "Object" } - } - }, - additionalProperties: Trigger.type.additionalProperties - } -}; - -export const RecurrenceScheduleOccurrence: msRest.CompositeMapper = { - serializedName: "RecurrenceScheduleOccurrence", - type: { - name: "Composite", - className: "RecurrenceScheduleOccurrence", - modelProperties: { - day: { - serializedName: "day", + }, + trustedCertPath: { + serializedName: "typeProperties.trustedCertPath", type: { - name: "Enum", - allowedValues: [ - "Sunday", - "Monday", - "Tuesday", - "Wednesday", - "Thursday", - "Friday", - "Saturday" - ] + name: "Object" } }, - occurrence: { - serializedName: "occurrence", + useSystemTrustStore: { + serializedName: "typeProperties.useSystemTrustStore", type: { - name: "Number" + name: "Object" + } + }, + encryptedCredential: { + serializedName: "typeProperties.encryptedCredential", + type: { + name: "Object" } } }, - additionalProperties: { - type: { - name: "Object" - } - } + additionalProperties: LinkedService.type.additionalProperties } }; -export const RecurrenceSchedule: msRest.CompositeMapper = { - serializedName: "RecurrenceSchedule", +export const OracleServiceCloudLinkedService: msRest.CompositeMapper = { + serializedName: "OracleServiceCloud", type: { name: "Composite", - className: "RecurrenceSchedule", + polymorphicDiscriminator: LinkedService.type.polymorphicDiscriminator, + uberParent: "LinkedService", + className: "OracleServiceCloudLinkedService", modelProperties: { - minutes: { - serializedName: "minutes", + ...LinkedService.type.modelProperties, + host: { + required: true, + serializedName: "typeProperties.host", type: { - name: "Sequence", - element: { - type: { - name: "Number" - } - } + name: "Object" } }, - hours: { - serializedName: "hours", + username: { + required: true, + serializedName: "typeProperties.username", type: { - name: "Sequence", - element: { - type: { - name: "Number" - } - } + name: "Object" } }, - weekDays: { - serializedName: "weekDays", + password: { + required: true, + serializedName: "typeProperties.password", type: { - name: "Sequence", - element: { - type: { - name: "Enum", - allowedValues: [ - "Sunday", - "Monday", - "Tuesday", - "Wednesday", - "Thursday", - "Friday", - "Saturday" - ] - } - } + name: "Composite", + className: "SecretBase" } }, - monthDays: { - serializedName: "monthDays", + useEncryptedEndpoints: { + serializedName: "typeProperties.useEncryptedEndpoints", type: { - name: "Sequence", - element: { - type: { - name: "Number" - } - } + name: "Object" } }, - monthlyOccurrences: { - serializedName: "monthlyOccurrences", + useHostVerification: { + serializedName: "typeProperties.useHostVerification", type: { - name: "Sequence", - element: { - type: { - name: "Composite", - className: "RecurrenceScheduleOccurrence", - additionalProperties: { - type: { - name: "Object" - } - } - } - } + name: "Object" + } + }, + usePeerVerification: { + serializedName: "typeProperties.usePeerVerification", + type: { + name: "Object" + } + }, + encryptedCredential: { + serializedName: "typeProperties.encryptedCredential", + type: { + name: "Object" } } }, - additionalProperties: { - type: { - name: "Object" - } - } + additionalProperties: LinkedService.type.additionalProperties } }; -export const ScheduleTriggerRecurrence: msRest.CompositeMapper = { - serializedName: "ScheduleTriggerRecurrence", +export const DynamicsAXLinkedService: msRest.CompositeMapper = { + serializedName: "DynamicsAX", type: { name: "Composite", - className: "ScheduleTriggerRecurrence", + polymorphicDiscriminator: LinkedService.type.polymorphicDiscriminator, + uberParent: "LinkedService", + className: "DynamicsAXLinkedService", modelProperties: { - frequency: { - serializedName: "frequency", + ...LinkedService.type.modelProperties, + url: { + required: true, + serializedName: "typeProperties.url", type: { - name: "String" + name: "Object" } }, - interval: { - serializedName: "interval", + servicePrincipalId: { + required: true, + serializedName: "typeProperties.servicePrincipalId", type: { - name: "Number" + name: "Object" } }, - startTime: { - serializedName: "startTime", + servicePrincipalKey: { + required: true, + serializedName: "typeProperties.servicePrincipalKey", type: { - name: "DateTime" + name: "Composite", + className: "SecretBase" } }, - endTime: { - serializedName: "endTime", + tenant: { + required: true, + serializedName: "typeProperties.tenant", type: { - name: "DateTime" + name: "Object" } }, - timeZone: { - serializedName: "timeZone", + aadResourceId: { + required: true, + serializedName: "typeProperties.aadResourceId", type: { - name: "String" + name: "Object" } }, - schedule: { - serializedName: "schedule", + encryptedCredential: { + serializedName: "typeProperties.encryptedCredential", type: { - name: "Composite", - className: "RecurrenceSchedule", - additionalProperties: { - type: { - name: "Object" - } - } - } - } - }, - additionalProperties: { - type: { - name: "Object" - } - } - } -}; - -export const ScheduleTrigger: msRest.CompositeMapper = { - serializedName: "ScheduleTrigger", - type: { - name: "Composite", - polymorphicDiscriminator: Trigger.type.polymorphicDiscriminator, - uberParent: "Trigger", - className: "ScheduleTrigger", - modelProperties: { - ...MultiplePipelineTrigger.type.modelProperties, - recurrence: { - required: true, - serializedName: "typeProperties.recurrence", - type: { - name: "Composite", - className: "ScheduleTriggerRecurrence", - additionalProperties: { - type: { - name: "Object" - } - } - } - } - }, - additionalProperties: Trigger.type.additionalProperties - } -}; - -export const AzureFunctionLinkedService: msRest.CompositeMapper = { - serializedName: "AzureFunction", - type: { - name: "Composite", - polymorphicDiscriminator: LinkedService.type.polymorphicDiscriminator, - uberParent: "LinkedService", - className: "AzureFunctionLinkedService", - modelProperties: { - ...LinkedService.type.modelProperties, - functionAppUrl: { - required: true, - serializedName: "typeProperties.functionAppUrl", - type: { - name: "Object" - } - }, - functionKey: { - serializedName: "typeProperties.functionKey", - type: { - name: "Composite", - className: "SecretBase" - } - }, - encryptedCredential: { - serializedName: "typeProperties.encryptedCredential", - type: { - name: "Object" + name: "Object" } } }, @@ -4053,13 +3752,13 @@ export const AzureFunctionLinkedService: msRest.CompositeMapper = { } }; -export const AzureDataExplorerLinkedService: msRest.CompositeMapper = { - serializedName: "AzureDataExplorer", +export const ResponsysLinkedService: msRest.CompositeMapper = { + serializedName: "Responsys", type: { name: "Composite", polymorphicDiscriminator: LinkedService.type.polymorphicDiscriminator, uberParent: "LinkedService", - className: "AzureDataExplorerLinkedService", + className: "ResponsysLinkedService", modelProperties: { ...LinkedService.type.modelProperties, endpoint: { @@ -4069,31 +3768,40 @@ export const AzureDataExplorerLinkedService: msRest.CompositeMapper = { name: "Object" } }, - servicePrincipalId: { + clientId: { required: true, - serializedName: "typeProperties.servicePrincipalId", + serializedName: "typeProperties.clientId", type: { name: "Object" } }, - servicePrincipalKey: { - required: true, - serializedName: "typeProperties.servicePrincipalKey", + clientSecret: { + serializedName: "typeProperties.clientSecret", type: { name: "Composite", className: "SecretBase" } }, - database: { - required: true, - serializedName: "typeProperties.database", + useEncryptedEndpoints: { + serializedName: "typeProperties.useEncryptedEndpoints", type: { name: "Object" } }, - tenant: { - required: true, - serializedName: "typeProperties.tenant", + useHostVerification: { + serializedName: "typeProperties.useHostVerification", + type: { + name: "Object" + } + }, + usePeerVerification: { + serializedName: "typeProperties.usePeerVerification", + type: { + name: "Object" + } + }, + encryptedCredential: { + serializedName: "typeProperties.encryptedCredential", type: { name: "Object" } @@ -4103,102 +3811,107 @@ export const AzureDataExplorerLinkedService: msRest.CompositeMapper = { } }; -export const SapTableLinkedService: msRest.CompositeMapper = { - serializedName: "SapTable", +export const AzureDatabricksLinkedService: msRest.CompositeMapper = { + serializedName: "AzureDatabricks", type: { name: "Composite", polymorphicDiscriminator: LinkedService.type.polymorphicDiscriminator, uberParent: "LinkedService", - className: "SapTableLinkedService", + className: "AzureDatabricksLinkedService", modelProperties: { ...LinkedService.type.modelProperties, - server: { - serializedName: "typeProperties.server", - type: { - name: "Object" - } - }, - systemNumber: { - serializedName: "typeProperties.systemNumber", + domain: { + required: true, + serializedName: "typeProperties.domain", type: { name: "Object" } }, - clientId: { - serializedName: "typeProperties.clientId", + accessToken: { + required: true, + serializedName: "typeProperties.accessToken", type: { - name: "Object" + name: "Composite", + className: "SecretBase" } }, - language: { - serializedName: "typeProperties.language", + existingClusterId: { + serializedName: "typeProperties.existingClusterId", type: { name: "Object" } }, - systemId: { - serializedName: "typeProperties.systemId", + instancePoolId: { + serializedName: "typeProperties.instancePoolId", type: { name: "Object" } }, - userName: { - serializedName: "typeProperties.userName", + newClusterVersion: { + serializedName: "typeProperties.newClusterVersion", type: { name: "Object" } }, - password: { - serializedName: "typeProperties.password", - type: { - name: "Composite", - className: "SecretBase" - } - }, - messageServer: { - serializedName: "typeProperties.messageServer", + newClusterNumOfWorker: { + serializedName: "typeProperties.newClusterNumOfWorker", type: { name: "Object" } }, - messageServerService: { - serializedName: "typeProperties.messageServerService", + newClusterNodeType: { + serializedName: "typeProperties.newClusterNodeType", type: { name: "Object" } }, - sncMode: { - serializedName: "typeProperties.sncMode", + newClusterSparkConf: { + serializedName: "typeProperties.newClusterSparkConf", type: { - name: "Object" + name: "Dictionary", + value: { + type: { + name: "Object" + } + } } }, - sncMyName: { - serializedName: "typeProperties.sncMyName", + newClusterSparkEnvVars: { + serializedName: "typeProperties.newClusterSparkEnvVars", type: { - name: "Object" + name: "Dictionary", + value: { + type: { + name: "Object" + } + } } }, - sncPartnerName: { - serializedName: "typeProperties.sncPartnerName", + newClusterCustomTags: { + serializedName: "typeProperties.newClusterCustomTags", type: { - name: "Object" + name: "Dictionary", + value: { + type: { + name: "Object" + } + } } }, - sncLibraryPath: { - serializedName: "typeProperties.sncLibraryPath", + newClusterDriverNodeType: { + serializedName: "typeProperties.newClusterDriverNodeType", type: { name: "Object" } }, - sncQop: { - serializedName: "typeProperties.sncQop", + newClusterInitScripts: { + serializedName: "typeProperties.newClusterInitScripts", type: { name: "Object" } }, - logonGroup: { - serializedName: "typeProperties.logonGroup", + newClusterEnableElasticDisk: { + serializedName: "typeProperties.newClusterEnableElasticDisk", type: { name: "Object" } @@ -4214,77 +3927,56 @@ export const SapTableLinkedService: msRest.CompositeMapper = { } }; -export const GoogleAdWordsLinkedService: msRest.CompositeMapper = { - serializedName: "GoogleAdWords", +export const AzureDataLakeAnalyticsLinkedService: msRest.CompositeMapper = { + serializedName: "AzureDataLakeAnalytics", type: { name: "Composite", polymorphicDiscriminator: LinkedService.type.polymorphicDiscriminator, uberParent: "LinkedService", - className: "GoogleAdWordsLinkedService", + className: "AzureDataLakeAnalyticsLinkedService", modelProperties: { ...LinkedService.type.modelProperties, - clientCustomerID: { + accountName: { required: true, - serializedName: "typeProperties.clientCustomerID", + serializedName: "typeProperties.accountName", type: { name: "Object" } }, - developerToken: { - required: true, - serializedName: "typeProperties.developerToken", - type: { - name: "Composite", - className: "SecretBase" - } - }, - authenticationType: { - required: true, - serializedName: "typeProperties.authenticationType", - type: { - name: "String" - } - }, - refreshToken: { - serializedName: "typeProperties.refreshToken", - type: { - name: "Composite", - className: "SecretBase" - } - }, - clientId: { - serializedName: "typeProperties.clientId", + servicePrincipalId: { + serializedName: "typeProperties.servicePrincipalId", type: { name: "Object" } }, - clientSecret: { - serializedName: "typeProperties.clientSecret", + servicePrincipalKey: { + serializedName: "typeProperties.servicePrincipalKey", type: { name: "Composite", className: "SecretBase" } }, - email: { - serializedName: "typeProperties.email", + tenant: { + required: true, + serializedName: "typeProperties.tenant", type: { name: "Object" } }, - keyFilePath: { - serializedName: "typeProperties.keyFilePath", + subscriptionId: { + serializedName: "typeProperties.subscriptionId", type: { name: "Object" } }, - trustedCertPath: { - serializedName: "typeProperties.trustedCertPath", + resourceGroupName: { + serializedName: "typeProperties.resourceGroupName", type: { name: "Object" } }, - useSystemTrustStore: { - serializedName: "typeProperties.useSystemTrustStore", + dataLakeAnalyticsUri: { + serializedName: "typeProperties.dataLakeAnalyticsUri", type: { name: "Object" } @@ -4300,91 +3992,96 @@ export const GoogleAdWordsLinkedService: msRest.CompositeMapper = { } }; -export const OracleServiceCloudLinkedService: msRest.CompositeMapper = { - serializedName: "OracleServiceCloud", +export const ScriptAction: msRest.CompositeMapper = { + serializedName: "ScriptAction", type: { name: "Composite", - polymorphicDiscriminator: LinkedService.type.polymorphicDiscriminator, - uberParent: "LinkedService", - className: "OracleServiceCloudLinkedService", + className: "ScriptAction", modelProperties: { - ...LinkedService.type.modelProperties, - host: { + name: { required: true, - serializedName: "typeProperties.host", + serializedName: "name", type: { - name: "Object" + name: "String" } }, - username: { + uri: { required: true, - serializedName: "typeProperties.username", + serializedName: "uri", type: { - name: "Object" + name: "String" } }, - password: { + roles: { required: true, - serializedName: "typeProperties.password", + serializedName: "roles", type: { - name: "Composite", - className: "SecretBase" + name: "Object" } }, - useEncryptedEndpoints: { - serializedName: "typeProperties.useEncryptedEndpoints", + parameters: { + serializedName: "parameters", + type: { + name: "String" + } + } + } + } +}; + +export const HDInsightOnDemandLinkedService: msRest.CompositeMapper = { + serializedName: "HDInsightOnDemand", + type: { + name: "Composite", + polymorphicDiscriminator: LinkedService.type.polymorphicDiscriminator, + uberParent: "LinkedService", + className: "HDInsightOnDemandLinkedService", + modelProperties: { + ...LinkedService.type.modelProperties, + clusterSize: { + required: true, + serializedName: "typeProperties.clusterSize", type: { name: "Object" } }, - useHostVerification: { - serializedName: "typeProperties.useHostVerification", + timeToLive: { + required: true, + serializedName: "typeProperties.timeToLive", type: { name: "Object" } }, - usePeerVerification: { - serializedName: "typeProperties.usePeerVerification", + version: { + required: true, + serializedName: "typeProperties.version", type: { name: "Object" } }, - encryptedCredential: { - serializedName: "typeProperties.encryptedCredential", + linkedServiceName: { + required: true, + serializedName: "typeProperties.linkedServiceName", + defaultValue: {}, type: { - name: "Object" + name: "Composite", + className: "LinkedServiceReference" } - } - }, - additionalProperties: LinkedService.type.additionalProperties - } -}; - -export const DynamicsAXLinkedService: msRest.CompositeMapper = { - serializedName: "DynamicsAX", - type: { - name: "Composite", - polymorphicDiscriminator: LinkedService.type.polymorphicDiscriminator, - uberParent: "LinkedService", - className: "DynamicsAXLinkedService", - modelProperties: { - ...LinkedService.type.modelProperties, - url: { + }, + hostSubscriptionId: { required: true, - serializedName: "typeProperties.url", + serializedName: "typeProperties.hostSubscriptionId", type: { name: "Object" } }, servicePrincipalId: { - required: true, serializedName: "typeProperties.servicePrincipalId", type: { name: "Object" } }, servicePrincipalKey: { - required: true, serializedName: "typeProperties.servicePrincipalKey", type: { name: "Composite", @@ -4398,190 +4095,168 @@ export const DynamicsAXLinkedService: msRest.CompositeMapper = { name: "Object" } }, - aadResourceId: { + clusterResourceGroup: { required: true, - serializedName: "typeProperties.aadResourceId", + serializedName: "typeProperties.clusterResourceGroup", type: { name: "Object" } }, - encryptedCredential: { - serializedName: "typeProperties.encryptedCredential", - type: { - name: "Object" - } - } - }, - additionalProperties: LinkedService.type.additionalProperties - } -}; - -export const ResponsysLinkedService: msRest.CompositeMapper = { - serializedName: "Responsys", - type: { - name: "Composite", - polymorphicDiscriminator: LinkedService.type.polymorphicDiscriminator, - uberParent: "LinkedService", - className: "ResponsysLinkedService", - modelProperties: { - ...LinkedService.type.modelProperties, - endpoint: { - required: true, - serializedName: "typeProperties.endpoint", + clusterNamePrefix: { + serializedName: "typeProperties.clusterNamePrefix", type: { name: "Object" } }, - clientId: { - required: true, - serializedName: "typeProperties.clientId", + clusterUserName: { + serializedName: "typeProperties.clusterUserName", type: { name: "Object" } }, - clientSecret: { - serializedName: "typeProperties.clientSecret", + clusterPassword: { + serializedName: "typeProperties.clusterPassword", type: { name: "Composite", className: "SecretBase" } }, - useEncryptedEndpoints: { - serializedName: "typeProperties.useEncryptedEndpoints", + clusterSshUserName: { + serializedName: "typeProperties.clusterSshUserName", type: { name: "Object" } }, - useHostVerification: { - serializedName: "typeProperties.useHostVerification", + clusterSshPassword: { + serializedName: "typeProperties.clusterSshPassword", type: { - name: "Object" + name: "Composite", + className: "SecretBase" } }, - usePeerVerification: { - serializedName: "typeProperties.usePeerVerification", + additionalLinkedServiceNames: { + serializedName: "typeProperties.additionalLinkedServiceNames", type: { - name: "Object" + name: "Sequence", + element: { + type: { + name: "Composite", + className: "LinkedServiceReference" + } + } } }, - encryptedCredential: { - serializedName: "typeProperties.encryptedCredential", + hcatalogLinkedServiceName: { + serializedName: "typeProperties.hcatalogLinkedServiceName", type: { - name: "Object" + name: "Composite", + className: "LinkedServiceReference" } - } - }, - additionalProperties: LinkedService.type.additionalProperties - } -}; - -export const AzureDatabricksLinkedService: msRest.CompositeMapper = { - serializedName: "AzureDatabricks", - type: { - name: "Composite", - polymorphicDiscriminator: LinkedService.type.polymorphicDiscriminator, - uberParent: "LinkedService", - className: "AzureDatabricksLinkedService", - modelProperties: { - ...LinkedService.type.modelProperties, - domain: { - required: true, - serializedName: "typeProperties.domain", + }, + clusterType: { + serializedName: "typeProperties.clusterType", type: { name: "Object" } }, - accessToken: { - required: true, - serializedName: "typeProperties.accessToken", + sparkVersion: { + serializedName: "typeProperties.sparkVersion", type: { - name: "Composite", - className: "SecretBase" + name: "Object" } }, - existingClusterId: { - serializedName: "typeProperties.existingClusterId", + coreConfiguration: { + serializedName: "typeProperties.coreConfiguration", type: { name: "Object" } }, - instancePoolId: { - serializedName: "typeProperties.instancePoolId", + hBaseConfiguration: { + serializedName: "typeProperties.hBaseConfiguration", type: { name: "Object" } }, - newClusterVersion: { - serializedName: "typeProperties.newClusterVersion", + hdfsConfiguration: { + serializedName: "typeProperties.hdfsConfiguration", type: { name: "Object" } }, - newClusterNumOfWorker: { - serializedName: "typeProperties.newClusterNumOfWorker", + hiveConfiguration: { + serializedName: "typeProperties.hiveConfiguration", type: { name: "Object" } }, - newClusterNodeType: { - serializedName: "typeProperties.newClusterNodeType", + mapReduceConfiguration: { + serializedName: "typeProperties.mapReduceConfiguration", type: { name: "Object" } }, - newClusterSparkConf: { - serializedName: "typeProperties.newClusterSparkConf", + oozieConfiguration: { + serializedName: "typeProperties.oozieConfiguration", type: { - name: "Dictionary", - value: { - type: { - name: "Object" - } - } + name: "Object" } }, - newClusterSparkEnvVars: { - serializedName: "typeProperties.newClusterSparkEnvVars", + stormConfiguration: { + serializedName: "typeProperties.stormConfiguration", type: { - name: "Dictionary", - value: { - type: { - name: "Object" - } - } + name: "Object" } }, - newClusterCustomTags: { - serializedName: "typeProperties.newClusterCustomTags", + yarnConfiguration: { + serializedName: "typeProperties.yarnConfiguration", type: { - name: "Dictionary", - value: { - type: { - name: "Object" - } - } + name: "Object" } }, - newClusterDriverNodeType: { - serializedName: "typeProperties.newClusterDriverNodeType", + encryptedCredential: { + serializedName: "typeProperties.encryptedCredential", type: { name: "Object" } }, - newClusterInitScripts: { - serializedName: "typeProperties.newClusterInitScripts", + headNodeSize: { + serializedName: "typeProperties.headNodeSize", type: { name: "Object" } }, - newClusterEnableElasticDisk: { - serializedName: "typeProperties.newClusterEnableElasticDisk", + dataNodeSize: { + serializedName: "typeProperties.dataNodeSize", type: { name: "Object" } }, - encryptedCredential: { - serializedName: "typeProperties.encryptedCredential", + zookeeperNodeSize: { + serializedName: "typeProperties.zookeeperNodeSize", + type: { + name: "Object" + } + }, + scriptActions: { + serializedName: "typeProperties.scriptActions", + type: { + name: "Sequence", + element: { + type: { + name: "Composite", + className: "ScriptAction" + } + } + } + }, + virtualNetworkId: { + serializedName: "typeProperties.virtualNetworkId", + type: { + name: "Object" + } + }, + subnetName: { + serializedName: "typeProperties.subnetName", type: { name: "Object" } @@ -4591,56 +4266,43 @@ export const AzureDatabricksLinkedService: msRest.CompositeMapper = { } }; -export const AzureDataLakeAnalyticsLinkedService: msRest.CompositeMapper = { - serializedName: "AzureDataLakeAnalytics", +export const SalesforceMarketingCloudLinkedService: msRest.CompositeMapper = { + serializedName: "SalesforceMarketingCloud", type: { name: "Composite", polymorphicDiscriminator: LinkedService.type.polymorphicDiscriminator, uberParent: "LinkedService", - className: "AzureDataLakeAnalyticsLinkedService", + className: "SalesforceMarketingCloudLinkedService", modelProperties: { ...LinkedService.type.modelProperties, - accountName: { + clientId: { required: true, - serializedName: "typeProperties.accountName", - type: { - name: "Object" - } - }, - servicePrincipalId: { - serializedName: "typeProperties.servicePrincipalId", + serializedName: "typeProperties.clientId", type: { name: "Object" } }, - servicePrincipalKey: { - serializedName: "typeProperties.servicePrincipalKey", + clientSecret: { + serializedName: "typeProperties.clientSecret", type: { name: "Composite", className: "SecretBase" } }, - tenant: { - required: true, - serializedName: "typeProperties.tenant", - type: { - name: "Object" - } - }, - subscriptionId: { - serializedName: "typeProperties.subscriptionId", + useEncryptedEndpoints: { + serializedName: "typeProperties.useEncryptedEndpoints", type: { name: "Object" } }, - resourceGroupName: { - serializedName: "typeProperties.resourceGroupName", + useHostVerification: { + serializedName: "typeProperties.useHostVerification", type: { name: "Object" } }, - dataLakeAnalyticsUri: { - serializedName: "typeProperties.dataLakeAnalyticsUri", + usePeerVerification: { + serializedName: "typeProperties.usePeerVerification", type: { name: "Object" } @@ -4656,271 +4318,240 @@ export const AzureDataLakeAnalyticsLinkedService: msRest.CompositeMapper = { } }; -export const ScriptAction: msRest.CompositeMapper = { - serializedName: "ScriptAction", +export const NetezzaLinkedService: msRest.CompositeMapper = { + serializedName: "Netezza", type: { name: "Composite", - className: "ScriptAction", + polymorphicDiscriminator: LinkedService.type.polymorphicDiscriminator, + uberParent: "LinkedService", + className: "NetezzaLinkedService", modelProperties: { - name: { - required: true, - serializedName: "name", + ...LinkedService.type.modelProperties, + connectionString: { + serializedName: "typeProperties.connectionString", type: { - name: "String" + name: "Object" } }, - uri: { - required: true, - serializedName: "uri", + pwd: { + serializedName: "typeProperties.pwd", type: { - name: "String" + name: "Composite", + className: "AzureKeyVaultSecretReference" } }, - roles: { - required: true, - serializedName: "roles", + encryptedCredential: { + serializedName: "typeProperties.encryptedCredential", type: { name: "Object" } - }, - parameters: { - serializedName: "parameters", - type: { - name: "String" - } } - } + }, + additionalProperties: LinkedService.type.additionalProperties } }; -export const HDInsightOnDemandLinkedService: msRest.CompositeMapper = { - serializedName: "HDInsightOnDemand", +export const VerticaLinkedService: msRest.CompositeMapper = { + serializedName: "Vertica", type: { name: "Composite", polymorphicDiscriminator: LinkedService.type.polymorphicDiscriminator, uberParent: "LinkedService", - className: "HDInsightOnDemandLinkedService", + className: "VerticaLinkedService", modelProperties: { ...LinkedService.type.modelProperties, - clusterSize: { - required: true, - serializedName: "typeProperties.clusterSize", - type: { - name: "Object" - } - }, - timeToLive: { - required: true, - serializedName: "typeProperties.timeToLive", - type: { - name: "Object" - } - }, - version: { - required: true, - serializedName: "typeProperties.version", + connectionString: { + serializedName: "typeProperties.connectionString", type: { name: "Object" } }, - linkedServiceName: { - required: true, - serializedName: "typeProperties.linkedServiceName", - defaultValue: {}, + pwd: { + serializedName: "typeProperties.pwd", type: { name: "Composite", - className: "LinkedServiceReference" + className: "AzureKeyVaultSecretReference" } }, - hostSubscriptionId: { - required: true, - serializedName: "typeProperties.hostSubscriptionId", + encryptedCredential: { + serializedName: "typeProperties.encryptedCredential", type: { name: "Object" } - }, - servicePrincipalId: { - serializedName: "typeProperties.servicePrincipalId", + } + }, + additionalProperties: LinkedService.type.additionalProperties + } +}; + +export const ZohoLinkedService: msRest.CompositeMapper = { + serializedName: "Zoho", + type: { + name: "Composite", + polymorphicDiscriminator: LinkedService.type.polymorphicDiscriminator, + uberParent: "LinkedService", + className: "ZohoLinkedService", + modelProperties: { + ...LinkedService.type.modelProperties, + endpoint: { + required: true, + serializedName: "typeProperties.endpoint", type: { name: "Object" } }, - servicePrincipalKey: { - serializedName: "typeProperties.servicePrincipalKey", + accessToken: { + serializedName: "typeProperties.accessToken", type: { name: "Composite", className: "SecretBase" } }, - tenant: { - required: true, - serializedName: "typeProperties.tenant", + useEncryptedEndpoints: { + serializedName: "typeProperties.useEncryptedEndpoints", type: { name: "Object" } }, - clusterResourceGroup: { - required: true, - serializedName: "typeProperties.clusterResourceGroup", + useHostVerification: { + serializedName: "typeProperties.useHostVerification", type: { name: "Object" } }, - clusterNamePrefix: { - serializedName: "typeProperties.clusterNamePrefix", + usePeerVerification: { + serializedName: "typeProperties.usePeerVerification", type: { name: "Object" } }, - clusterUserName: { - serializedName: "typeProperties.clusterUserName", + encryptedCredential: { + serializedName: "typeProperties.encryptedCredential", type: { name: "Object" } - }, - clusterPassword: { - serializedName: "typeProperties.clusterPassword", - type: { - name: "Composite", - className: "SecretBase" - } - }, - clusterSshUserName: { - serializedName: "typeProperties.clusterSshUserName", + } + }, + additionalProperties: LinkedService.type.additionalProperties + } +}; + +export const XeroLinkedService: msRest.CompositeMapper = { + serializedName: "Xero", + type: { + name: "Composite", + polymorphicDiscriminator: LinkedService.type.polymorphicDiscriminator, + uberParent: "LinkedService", + className: "XeroLinkedService", + modelProperties: { + ...LinkedService.type.modelProperties, + host: { + required: true, + serializedName: "typeProperties.host", type: { name: "Object" } }, - clusterSshPassword: { - serializedName: "typeProperties.clusterSshPassword", + consumerKey: { + serializedName: "typeProperties.consumerKey", type: { name: "Composite", className: "SecretBase" } }, - additionalLinkedServiceNames: { - serializedName: "typeProperties.additionalLinkedServiceNames", - type: { - name: "Sequence", - element: { - type: { - name: "Composite", - className: "LinkedServiceReference" - } - } - } - }, - hcatalogLinkedServiceName: { - serializedName: "typeProperties.hcatalogLinkedServiceName", + privateKey: { + serializedName: "typeProperties.privateKey", type: { name: "Composite", - className: "LinkedServiceReference" - } - }, - clusterType: { - serializedName: "typeProperties.clusterType", - type: { - name: "Object" + className: "SecretBase" } }, - sparkVersion: { - serializedName: "typeProperties.sparkVersion", + useEncryptedEndpoints: { + serializedName: "typeProperties.useEncryptedEndpoints", type: { name: "Object" } }, - coreConfiguration: { - serializedName: "typeProperties.coreConfiguration", + useHostVerification: { + serializedName: "typeProperties.useHostVerification", type: { name: "Object" } }, - hBaseConfiguration: { - serializedName: "typeProperties.hBaseConfiguration", + usePeerVerification: { + serializedName: "typeProperties.usePeerVerification", type: { name: "Object" } }, - hdfsConfiguration: { - serializedName: "typeProperties.hdfsConfiguration", + encryptedCredential: { + serializedName: "typeProperties.encryptedCredential", type: { name: "Object" } - }, - hiveConfiguration: { - serializedName: "typeProperties.hiveConfiguration", - type: { - name: "Object" - } - }, - mapReduceConfiguration: { - serializedName: "typeProperties.mapReduceConfiguration", - type: { - name: "Object" - } - }, - oozieConfiguration: { - serializedName: "typeProperties.oozieConfiguration", - type: { - name: "Object" - } - }, - stormConfiguration: { - serializedName: "typeProperties.stormConfiguration", + } + }, + additionalProperties: LinkedService.type.additionalProperties + } +}; + +export const SquareLinkedService: msRest.CompositeMapper = { + serializedName: "Square", + type: { + name: "Composite", + polymorphicDiscriminator: LinkedService.type.polymorphicDiscriminator, + uberParent: "LinkedService", + className: "SquareLinkedService", + modelProperties: { + ...LinkedService.type.modelProperties, + host: { + required: true, + serializedName: "typeProperties.host", type: { name: "Object" } }, - yarnConfiguration: { - serializedName: "typeProperties.yarnConfiguration", + clientId: { + required: true, + serializedName: "typeProperties.clientId", type: { name: "Object" } }, - encryptedCredential: { - serializedName: "typeProperties.encryptedCredential", + clientSecret: { + serializedName: "typeProperties.clientSecret", type: { - name: "Object" + name: "Composite", + className: "SecretBase" } }, - headNodeSize: { - serializedName: "typeProperties.headNodeSize", + redirectUri: { + required: true, + serializedName: "typeProperties.redirectUri", type: { name: "Object" } }, - dataNodeSize: { - serializedName: "typeProperties.dataNodeSize", + useEncryptedEndpoints: { + serializedName: "typeProperties.useEncryptedEndpoints", type: { name: "Object" } }, - zookeeperNodeSize: { - serializedName: "typeProperties.zookeeperNodeSize", + useHostVerification: { + serializedName: "typeProperties.useHostVerification", type: { name: "Object" } }, - scriptActions: { - serializedName: "typeProperties.scriptActions", - type: { - name: "Sequence", - element: { - type: { - name: "Composite", - className: "ScriptAction" - } - } - } - }, - virtualNetworkId: { - serializedName: "typeProperties.virtualNetworkId", + usePeerVerification: { + serializedName: "typeProperties.usePeerVerification", type: { name: "Object" } }, - subnetName: { - serializedName: "typeProperties.subnetName", + encryptedCredential: { + serializedName: "typeProperties.encryptedCredential", type: { name: "Object" } @@ -4930,111 +4561,95 @@ export const HDInsightOnDemandLinkedService: msRest.CompositeMapper = { } }; -export const SalesforceMarketingCloudLinkedService: msRest.CompositeMapper = { - serializedName: "SalesforceMarketingCloud", +export const SparkLinkedService: msRest.CompositeMapper = { + serializedName: "Spark", type: { name: "Composite", polymorphicDiscriminator: LinkedService.type.polymorphicDiscriminator, uberParent: "LinkedService", - className: "SalesforceMarketingCloudLinkedService", + className: "SparkLinkedService", modelProperties: { ...LinkedService.type.modelProperties, - clientId: { + host: { required: true, - serializedName: "typeProperties.clientId", + serializedName: "typeProperties.host", type: { name: "Object" } }, - clientSecret: { - serializedName: "typeProperties.clientSecret", + port: { + required: true, + serializedName: "typeProperties.port", type: { - name: "Composite", - className: "SecretBase" + name: "Object" } }, - useEncryptedEndpoints: { - serializedName: "typeProperties.useEncryptedEndpoints", + serverType: { + serializedName: "typeProperties.serverType", type: { - name: "Object" + name: "String" } }, - useHostVerification: { - serializedName: "typeProperties.useHostVerification", + thriftTransportProtocol: { + serializedName: "typeProperties.thriftTransportProtocol", type: { - name: "Object" + name: "String" } }, - usePeerVerification: { - serializedName: "typeProperties.usePeerVerification", + authenticationType: { + required: true, + serializedName: "typeProperties.authenticationType", + type: { + name: "String" + } + }, + username: { + serializedName: "typeProperties.username", type: { name: "Object" } }, - encryptedCredential: { - serializedName: "typeProperties.encryptedCredential", + password: { + serializedName: "typeProperties.password", + type: { + name: "Composite", + className: "SecretBase" + } + }, + httpPath: { + serializedName: "typeProperties.httpPath", type: { name: "Object" } - } - }, - additionalProperties: LinkedService.type.additionalProperties - } -}; - -export const NetezzaLinkedService: msRest.CompositeMapper = { - serializedName: "Netezza", - type: { - name: "Composite", - polymorphicDiscriminator: LinkedService.type.polymorphicDiscriminator, - uberParent: "LinkedService", - className: "NetezzaLinkedService", - modelProperties: { - ...LinkedService.type.modelProperties, - connectionString: { - serializedName: "typeProperties.connectionString", + }, + enableSsl: { + serializedName: "typeProperties.enableSsl", type: { name: "Object" } }, - pwd: { - serializedName: "typeProperties.pwd", + trustedCertPath: { + serializedName: "typeProperties.trustedCertPath", type: { - name: "Composite", - className: "AzureKeyVaultSecretReference" + name: "Object" } }, - encryptedCredential: { - serializedName: "typeProperties.encryptedCredential", + useSystemTrustStore: { + serializedName: "typeProperties.useSystemTrustStore", type: { name: "Object" } - } - }, - additionalProperties: LinkedService.type.additionalProperties - } -}; - -export const VerticaLinkedService: msRest.CompositeMapper = { - serializedName: "Vertica", - type: { - name: "Composite", - polymorphicDiscriminator: LinkedService.type.polymorphicDiscriminator, - uberParent: "LinkedService", - className: "VerticaLinkedService", - modelProperties: { - ...LinkedService.type.modelProperties, - connectionString: { - serializedName: "typeProperties.connectionString", + }, + allowHostNameCNMismatch: { + serializedName: "typeProperties.allowHostNameCNMismatch", type: { name: "Object" } }, - pwd: { - serializedName: "typeProperties.pwd", + allowSelfSignedServerCert: { + serializedName: "typeProperties.allowSelfSignedServerCert", type: { - name: "Composite", - className: "AzureKeyVaultSecretReference" + name: "Object" } }, encryptedCredential: { @@ -5048,18 +4663,18 @@ export const VerticaLinkedService: msRest.CompositeMapper = { } }; -export const ZohoLinkedService: msRest.CompositeMapper = { - serializedName: "Zoho", +export const ShopifyLinkedService: msRest.CompositeMapper = { + serializedName: "Shopify", type: { name: "Composite", polymorphicDiscriminator: LinkedService.type.polymorphicDiscriminator, uberParent: "LinkedService", - className: "ZohoLinkedService", + className: "ShopifyLinkedService", modelProperties: { ...LinkedService.type.modelProperties, - endpoint: { + host: { required: true, - serializedName: "typeProperties.endpoint", + serializedName: "typeProperties.host", type: { name: "Object" } @@ -5100,31 +4715,50 @@ export const ZohoLinkedService: msRest.CompositeMapper = { } }; -export const XeroLinkedService: msRest.CompositeMapper = { - serializedName: "Xero", +export const ServiceNowLinkedService: msRest.CompositeMapper = { + serializedName: "ServiceNow", type: { name: "Composite", polymorphicDiscriminator: LinkedService.type.polymorphicDiscriminator, uberParent: "LinkedService", - className: "XeroLinkedService", + className: "ServiceNowLinkedService", modelProperties: { ...LinkedService.type.modelProperties, - host: { + endpoint: { required: true, - serializedName: "typeProperties.host", + serializedName: "typeProperties.endpoint", type: { name: "Object" } }, - consumerKey: { - serializedName: "typeProperties.consumerKey", + authenticationType: { + required: true, + serializedName: "typeProperties.authenticationType", + type: { + name: "String" + } + }, + username: { + serializedName: "typeProperties.username", + type: { + name: "Object" + } + }, + password: { + serializedName: "typeProperties.password", type: { name: "Composite", className: "SecretBase" } }, - privateKey: { - serializedName: "typeProperties.privateKey", + clientId: { + serializedName: "typeProperties.clientId", + type: { + name: "Object" + } + }, + clientSecret: { + serializedName: "typeProperties.clientSecret", type: { name: "Composite", className: "SecretBase" @@ -5159,57 +4793,62 @@ export const XeroLinkedService: msRest.CompositeMapper = { } }; -export const SquareLinkedService: msRest.CompositeMapper = { - serializedName: "Square", +export const QuickBooksLinkedService: msRest.CompositeMapper = { + serializedName: "QuickBooks", type: { name: "Composite", polymorphicDiscriminator: LinkedService.type.polymorphicDiscriminator, uberParent: "LinkedService", - className: "SquareLinkedService", + className: "QuickBooksLinkedService", modelProperties: { ...LinkedService.type.modelProperties, - host: { + endpoint: { required: true, - serializedName: "typeProperties.host", + serializedName: "typeProperties.endpoint", type: { name: "Object" } }, - clientId: { + companyId: { required: true, - serializedName: "typeProperties.clientId", + serializedName: "typeProperties.companyId", type: { name: "Object" } }, - clientSecret: { - serializedName: "typeProperties.clientSecret", + consumerKey: { + required: true, + serializedName: "typeProperties.consumerKey", type: { - name: "Composite", - className: "SecretBase" + name: "Object" } }, - redirectUri: { + consumerSecret: { required: true, - serializedName: "typeProperties.redirectUri", + serializedName: "typeProperties.consumerSecret", type: { - name: "Object" + name: "Composite", + className: "SecretBase" } }, - useEncryptedEndpoints: { - serializedName: "typeProperties.useEncryptedEndpoints", + accessToken: { + required: true, + serializedName: "typeProperties.accessToken", type: { - name: "Object" + name: "Composite", + className: "SecretBase" } }, - useHostVerification: { - serializedName: "typeProperties.useHostVerification", + accessTokenSecret: { + required: true, + serializedName: "typeProperties.accessTokenSecret", type: { - name: "Object" + name: "Composite", + className: "SecretBase" } }, - usePeerVerification: { - serializedName: "typeProperties.usePeerVerification", + useEncryptedEndpoints: { + serializedName: "typeProperties.useEncryptedEndpoints", type: { name: "Object" } @@ -5225,13 +4864,13 @@ export const SquareLinkedService: msRest.CompositeMapper = { } }; -export const SparkLinkedService: msRest.CompositeMapper = { - serializedName: "Spark", +export const PrestoLinkedService: msRest.CompositeMapper = { + serializedName: "Presto", type: { name: "Composite", polymorphicDiscriminator: LinkedService.type.polymorphicDiscriminator, uberParent: "LinkedService", - className: "SparkLinkedService", + className: "PrestoLinkedService", modelProperties: { ...LinkedService.type.modelProperties, host: { @@ -5241,23 +4880,24 @@ export const SparkLinkedService: msRest.CompositeMapper = { name: "Object" } }, - port: { + serverVersion: { required: true, - serializedName: "typeProperties.port", + serializedName: "typeProperties.serverVersion", type: { name: "Object" } }, - serverType: { - serializedName: "typeProperties.serverType", + catalog: { + required: true, + serializedName: "typeProperties.catalog", type: { - name: "String" + name: "Object" } }, - thriftTransportProtocol: { - serializedName: "typeProperties.thriftTransportProtocol", + port: { + serializedName: "typeProperties.port", type: { - name: "String" + name: "Object" } }, authenticationType: { @@ -5280,12 +4920,6 @@ export const SparkLinkedService: msRest.CompositeMapper = { className: "SecretBase" } }, - httpPath: { - serializedName: "typeProperties.httpPath", - type: { - name: "Object" - } - }, enableSsl: { serializedName: "typeProperties.enableSsl", type: { @@ -5316,6 +4950,12 @@ export const SparkLinkedService: msRest.CompositeMapper = { name: "Object" } }, + timeZoneID: { + serializedName: "typeProperties.timeZoneID", + type: { + name: "Object" + } + }, encryptedCredential: { serializedName: "typeProperties.encryptedCredential", type: { @@ -5327,13 +4967,13 @@ export const SparkLinkedService: msRest.CompositeMapper = { } }; -export const ShopifyLinkedService: msRest.CompositeMapper = { - serializedName: "Shopify", +export const PhoenixLinkedService: msRest.CompositeMapper = { + serializedName: "Phoenix", type: { name: "Composite", polymorphicDiscriminator: LinkedService.type.polymorphicDiscriminator, uberParent: "LinkedService", - className: "ShopifyLinkedService", + className: "PhoenixLinkedService", modelProperties: { ...LinkedService.type.modelProperties, host: { @@ -5343,27 +4983,64 @@ export const ShopifyLinkedService: msRest.CompositeMapper = { name: "Object" } }, - accessToken: { - serializedName: "typeProperties.accessToken", + port: { + serializedName: "typeProperties.port", + type: { + name: "Object" + } + }, + httpPath: { + serializedName: "typeProperties.httpPath", + type: { + name: "Object" + } + }, + authenticationType: { + required: true, + serializedName: "typeProperties.authenticationType", + type: { + name: "String" + } + }, + username: { + serializedName: "typeProperties.username", + type: { + name: "Object" + } + }, + password: { + serializedName: "typeProperties.password", type: { name: "Composite", className: "SecretBase" } }, - useEncryptedEndpoints: { - serializedName: "typeProperties.useEncryptedEndpoints", + enableSsl: { + serializedName: "typeProperties.enableSsl", type: { name: "Object" } }, - useHostVerification: { - serializedName: "typeProperties.useHostVerification", + trustedCertPath: { + serializedName: "typeProperties.trustedCertPath", type: { name: "Object" } }, - usePeerVerification: { - serializedName: "typeProperties.usePeerVerification", + useSystemTrustStore: { + serializedName: "typeProperties.useSystemTrustStore", + type: { + name: "Object" + } + }, + allowHostNameCNMismatch: { + serializedName: "typeProperties.allowHostNameCNMismatch", + type: { + name: "Object" + } + }, + allowSelfSignedServerCert: { + serializedName: "typeProperties.allowSelfSignedServerCert", type: { name: "Object" } @@ -5379,43 +5056,24 @@ export const ShopifyLinkedService: msRest.CompositeMapper = { } }; -export const ServiceNowLinkedService: msRest.CompositeMapper = { - serializedName: "ServiceNow", +export const PaypalLinkedService: msRest.CompositeMapper = { + serializedName: "Paypal", type: { name: "Composite", polymorphicDiscriminator: LinkedService.type.polymorphicDiscriminator, uberParent: "LinkedService", - className: "ServiceNowLinkedService", + className: "PaypalLinkedService", modelProperties: { ...LinkedService.type.modelProperties, - endpoint: { - required: true, - serializedName: "typeProperties.endpoint", - type: { - name: "Object" - } - }, - authenticationType: { + host: { required: true, - serializedName: "typeProperties.authenticationType", - type: { - name: "String" - } - }, - username: { - serializedName: "typeProperties.username", + serializedName: "typeProperties.host", type: { name: "Object" } }, - password: { - serializedName: "typeProperties.password", - type: { - name: "Composite", - className: "SecretBase" - } - }, clientId: { + required: true, serializedName: "typeProperties.clientId", type: { name: "Object" @@ -5457,13 +5115,13 @@ export const ServiceNowLinkedService: msRest.CompositeMapper = { } }; -export const QuickBooksLinkedService: msRest.CompositeMapper = { - serializedName: "QuickBooks", +export const MarketoLinkedService: msRest.CompositeMapper = { + serializedName: "Marketo", type: { name: "Composite", polymorphicDiscriminator: LinkedService.type.polymorphicDiscriminator, uberParent: "LinkedService", - className: "QuickBooksLinkedService", + className: "MarketoLinkedService", modelProperties: { ...LinkedService.type.modelProperties, endpoint: { @@ -5473,46 +5131,34 @@ export const QuickBooksLinkedService: msRest.CompositeMapper = { name: "Object" } }, - companyId: { - required: true, - serializedName: "typeProperties.companyId", - type: { - name: "Object" - } - }, - consumerKey: { + clientId: { required: true, - serializedName: "typeProperties.consumerKey", + serializedName: "typeProperties.clientId", type: { name: "Object" } }, - consumerSecret: { - required: true, - serializedName: "typeProperties.consumerSecret", + clientSecret: { + serializedName: "typeProperties.clientSecret", type: { name: "Composite", className: "SecretBase" } }, - accessToken: { - required: true, - serializedName: "typeProperties.accessToken", + useEncryptedEndpoints: { + serializedName: "typeProperties.useEncryptedEndpoints", type: { - name: "Composite", - className: "SecretBase" + name: "Object" } }, - accessTokenSecret: { - required: true, - serializedName: "typeProperties.accessTokenSecret", + useHostVerification: { + serializedName: "typeProperties.useHostVerification", type: { - name: "Composite", - className: "SecretBase" + name: "Object" } }, - useEncryptedEndpoints: { - serializedName: "typeProperties.useEncryptedEndpoints", + usePeerVerification: { + serializedName: "typeProperties.usePeerVerification", type: { name: "Object" } @@ -5528,94 +5174,174 @@ export const QuickBooksLinkedService: msRest.CompositeMapper = { } }; -export const PrestoLinkedService: msRest.CompositeMapper = { - serializedName: "Presto", +export const AzureMariaDBLinkedService: msRest.CompositeMapper = { + serializedName: "AzureMariaDB", type: { name: "Composite", polymorphicDiscriminator: LinkedService.type.polymorphicDiscriminator, uberParent: "LinkedService", - className: "PrestoLinkedService", + className: "AzureMariaDBLinkedService", modelProperties: { ...LinkedService.type.modelProperties, - host: { - required: true, - serializedName: "typeProperties.host", + connectionString: { + serializedName: "typeProperties.connectionString", type: { name: "Object" } }, - serverVersion: { - required: true, - serializedName: "typeProperties.serverVersion", + pwd: { + serializedName: "typeProperties.pwd", type: { - name: "Object" + name: "Composite", + className: "AzureKeyVaultSecretReference" } }, - catalog: { - required: true, - serializedName: "typeProperties.catalog", + encryptedCredential: { + serializedName: "typeProperties.encryptedCredential", type: { name: "Object" } - }, - port: { - serializedName: "typeProperties.port", + } + }, + additionalProperties: LinkedService.type.additionalProperties + } +}; + +export const MariaDBLinkedService: msRest.CompositeMapper = { + serializedName: "MariaDB", + type: { + name: "Composite", + polymorphicDiscriminator: LinkedService.type.polymorphicDiscriminator, + uberParent: "LinkedService", + className: "MariaDBLinkedService", + modelProperties: { + ...LinkedService.type.modelProperties, + connectionString: { + serializedName: "typeProperties.connectionString", type: { name: "Object" } }, - authenticationType: { - required: true, - serializedName: "typeProperties.authenticationType", + pwd: { + serializedName: "typeProperties.pwd", type: { - name: "String" + name: "Composite", + className: "AzureKeyVaultSecretReference" } }, - username: { - serializedName: "typeProperties.username", - type: { + encryptedCredential: { + serializedName: "typeProperties.encryptedCredential", + type: { + name: "Object" + } + } + }, + additionalProperties: LinkedService.type.additionalProperties + } +}; + +export const MagentoLinkedService: msRest.CompositeMapper = { + serializedName: "Magento", + type: { + name: "Composite", + polymorphicDiscriminator: LinkedService.type.polymorphicDiscriminator, + uberParent: "LinkedService", + className: "MagentoLinkedService", + modelProperties: { + ...LinkedService.type.modelProperties, + host: { + required: true, + serializedName: "typeProperties.host", + type: { name: "Object" } }, - password: { - serializedName: "typeProperties.password", + accessToken: { + serializedName: "typeProperties.accessToken", type: { name: "Composite", className: "SecretBase" } }, - enableSsl: { - serializedName: "typeProperties.enableSsl", + useEncryptedEndpoints: { + serializedName: "typeProperties.useEncryptedEndpoints", type: { name: "Object" } }, - trustedCertPath: { - serializedName: "typeProperties.trustedCertPath", + useHostVerification: { + serializedName: "typeProperties.useHostVerification", type: { name: "Object" } }, - useSystemTrustStore: { - serializedName: "typeProperties.useSystemTrustStore", + usePeerVerification: { + serializedName: "typeProperties.usePeerVerification", type: { name: "Object" } }, - allowHostNameCNMismatch: { - serializedName: "typeProperties.allowHostNameCNMismatch", + encryptedCredential: { + serializedName: "typeProperties.encryptedCredential", + type: { + name: "Object" + } + } + }, + additionalProperties: LinkedService.type.additionalProperties + } +}; + +export const JiraLinkedService: msRest.CompositeMapper = { + serializedName: "Jira", + type: { + name: "Composite", + polymorphicDiscriminator: LinkedService.type.polymorphicDiscriminator, + uberParent: "LinkedService", + className: "JiraLinkedService", + modelProperties: { + ...LinkedService.type.modelProperties, + host: { + required: true, + serializedName: "typeProperties.host", type: { name: "Object" } }, - allowSelfSignedServerCert: { - serializedName: "typeProperties.allowSelfSignedServerCert", + port: { + serializedName: "typeProperties.port", type: { name: "Object" } }, - timeZoneID: { - serializedName: "typeProperties.timeZoneID", + username: { + required: true, + serializedName: "typeProperties.username", + type: { + name: "Object" + } + }, + password: { + serializedName: "typeProperties.password", + type: { + name: "Composite", + className: "SecretBase" + } + }, + useEncryptedEndpoints: { + serializedName: "typeProperties.useEncryptedEndpoints", + type: { + name: "Object" + } + }, + useHostVerification: { + serializedName: "typeProperties.useHostVerification", + type: { + name: "Object" + } + }, + usePeerVerification: { + serializedName: "typeProperties.usePeerVerification", type: { name: "Object" } @@ -5631,13 +5357,13 @@ export const PrestoLinkedService: msRest.CompositeMapper = { } }; -export const PhoenixLinkedService: msRest.CompositeMapper = { - serializedName: "Phoenix", +export const ImpalaLinkedService: msRest.CompositeMapper = { + serializedName: "Impala", type: { name: "Composite", polymorphicDiscriminator: LinkedService.type.polymorphicDiscriminator, uberParent: "LinkedService", - className: "PhoenixLinkedService", + className: "ImpalaLinkedService", modelProperties: { ...LinkedService.type.modelProperties, host: { @@ -5653,12 +5379,6 @@ export const PhoenixLinkedService: msRest.CompositeMapper = { name: "Object" } }, - httpPath: { - serializedName: "typeProperties.httpPath", - type: { - name: "Object" - } - }, authenticationType: { required: true, serializedName: "typeProperties.authenticationType", @@ -5720,22 +5440,15 @@ export const PhoenixLinkedService: msRest.CompositeMapper = { } }; -export const PaypalLinkedService: msRest.CompositeMapper = { - serializedName: "Paypal", +export const HubspotLinkedService: msRest.CompositeMapper = { + serializedName: "Hubspot", type: { name: "Composite", polymorphicDiscriminator: LinkedService.type.polymorphicDiscriminator, uberParent: "LinkedService", - className: "PaypalLinkedService", + className: "HubspotLinkedService", modelProperties: { ...LinkedService.type.modelProperties, - host: { - required: true, - serializedName: "typeProperties.host", - type: { - name: "Object" - } - }, clientId: { required: true, serializedName: "typeProperties.clientId", @@ -5750,6 +5463,20 @@ export const PaypalLinkedService: msRest.CompositeMapper = { className: "SecretBase" } }, + accessToken: { + serializedName: "typeProperties.accessToken", + type: { + name: "Composite", + className: "SecretBase" + } + }, + refreshToken: { + serializedName: "typeProperties.refreshToken", + type: { + name: "Composite", + className: "SecretBase" + } + }, useEncryptedEndpoints: { serializedName: "typeProperties.useEncryptedEndpoints", type: { @@ -5779,85 +5506,112 @@ export const PaypalLinkedService: msRest.CompositeMapper = { } }; -export const MarketoLinkedService: msRest.CompositeMapper = { - serializedName: "Marketo", +export const HiveLinkedService: msRest.CompositeMapper = { + serializedName: "Hive", type: { name: "Composite", polymorphicDiscriminator: LinkedService.type.polymorphicDiscriminator, uberParent: "LinkedService", - className: "MarketoLinkedService", + className: "HiveLinkedService", modelProperties: { ...LinkedService.type.modelProperties, - endpoint: { + host: { required: true, - serializedName: "typeProperties.endpoint", + serializedName: "typeProperties.host", type: { name: "Object" } }, - clientId: { + port: { + serializedName: "typeProperties.port", + type: { + name: "Object" + } + }, + serverType: { + serializedName: "typeProperties.serverType", + type: { + name: "String" + } + }, + thriftTransportProtocol: { + serializedName: "typeProperties.thriftTransportProtocol", + type: { + name: "String" + } + }, + authenticationType: { required: true, - serializedName: "typeProperties.clientId", + serializedName: "typeProperties.authenticationType", + type: { + name: "String" + } + }, + serviceDiscoveryMode: { + serializedName: "typeProperties.serviceDiscoveryMode", type: { name: "Object" } }, - clientSecret: { - serializedName: "typeProperties.clientSecret", + zooKeeperNameSpace: { + serializedName: "typeProperties.zooKeeperNameSpace", + type: { + name: "Object" + } + }, + useNativeQuery: { + serializedName: "typeProperties.useNativeQuery", + type: { + name: "Object" + } + }, + username: { + serializedName: "typeProperties.username", + type: { + name: "Object" + } + }, + password: { + serializedName: "typeProperties.password", type: { name: "Composite", className: "SecretBase" } }, - useEncryptedEndpoints: { - serializedName: "typeProperties.useEncryptedEndpoints", + httpPath: { + serializedName: "typeProperties.httpPath", type: { name: "Object" } }, - useHostVerification: { - serializedName: "typeProperties.useHostVerification", + enableSsl: { + serializedName: "typeProperties.enableSsl", type: { name: "Object" } }, - usePeerVerification: { - serializedName: "typeProperties.usePeerVerification", + trustedCertPath: { + serializedName: "typeProperties.trustedCertPath", type: { name: "Object" } }, - encryptedCredential: { - serializedName: "typeProperties.encryptedCredential", + useSystemTrustStore: { + serializedName: "typeProperties.useSystemTrustStore", type: { name: "Object" } - } - }, - additionalProperties: LinkedService.type.additionalProperties - } -}; - -export const AzureMariaDBLinkedService: msRest.CompositeMapper = { - serializedName: "AzureMariaDB", - type: { - name: "Composite", - polymorphicDiscriminator: LinkedService.type.polymorphicDiscriminator, - uberParent: "LinkedService", - className: "AzureMariaDBLinkedService", - modelProperties: { - ...LinkedService.type.modelProperties, - connectionString: { - serializedName: "typeProperties.connectionString", + }, + allowHostNameCNMismatch: { + serializedName: "typeProperties.allowHostNameCNMismatch", type: { name: "Object" } }, - pwd: { - serializedName: "typeProperties.pwd", + allowSelfSignedServerCert: { + serializedName: "typeProperties.allowSelfSignedServerCert", type: { - name: "Composite", - className: "AzureKeyVaultSecretReference" + name: "Object" } }, encryptedCredential: { @@ -5871,76 +5625,74 @@ export const AzureMariaDBLinkedService: msRest.CompositeMapper = { } }; -export const MariaDBLinkedService: msRest.CompositeMapper = { - serializedName: "MariaDB", +export const HBaseLinkedService: msRest.CompositeMapper = { + serializedName: "HBase", type: { name: "Composite", polymorphicDiscriminator: LinkedService.type.polymorphicDiscriminator, uberParent: "LinkedService", - className: "MariaDBLinkedService", + className: "HBaseLinkedService", modelProperties: { ...LinkedService.type.modelProperties, - connectionString: { - serializedName: "typeProperties.connectionString", + host: { + required: true, + serializedName: "typeProperties.host", type: { name: "Object" } }, - pwd: { - serializedName: "typeProperties.pwd", + port: { + serializedName: "typeProperties.port", type: { - name: "Composite", - className: "AzureKeyVaultSecretReference" + name: "Object" } }, - encryptedCredential: { - serializedName: "typeProperties.encryptedCredential", + httpPath: { + serializedName: "typeProperties.httpPath", type: { name: "Object" } - } - }, - additionalProperties: LinkedService.type.additionalProperties - } -}; - -export const MagentoLinkedService: msRest.CompositeMapper = { - serializedName: "Magento", - type: { - name: "Composite", - polymorphicDiscriminator: LinkedService.type.polymorphicDiscriminator, - uberParent: "LinkedService", - className: "MagentoLinkedService", - modelProperties: { - ...LinkedService.type.modelProperties, - host: { + }, + authenticationType: { required: true, - serializedName: "typeProperties.host", + serializedName: "typeProperties.authenticationType", + type: { + name: "String" + } + }, + username: { + serializedName: "typeProperties.username", type: { name: "Object" } }, - accessToken: { - serializedName: "typeProperties.accessToken", + password: { + serializedName: "typeProperties.password", type: { name: "Composite", className: "SecretBase" } }, - useEncryptedEndpoints: { - serializedName: "typeProperties.useEncryptedEndpoints", + enableSsl: { + serializedName: "typeProperties.enableSsl", type: { name: "Object" } }, - useHostVerification: { - serializedName: "typeProperties.useHostVerification", + trustedCertPath: { + serializedName: "typeProperties.trustedCertPath", type: { name: "Object" } }, - usePeerVerification: { - serializedName: "typeProperties.usePeerVerification", + allowHostNameCNMismatch: { + serializedName: "typeProperties.allowHostNameCNMismatch", + type: { + name: "Object" + } + }, + allowSelfSignedServerCert: { + serializedName: "typeProperties.allowSelfSignedServerCert", type: { name: "Object" } @@ -5956,58 +5708,26 @@ export const MagentoLinkedService: msRest.CompositeMapper = { } }; -export const JiraLinkedService: msRest.CompositeMapper = { - serializedName: "Jira", +export const GreenplumLinkedService: msRest.CompositeMapper = { + serializedName: "Greenplum", type: { name: "Composite", polymorphicDiscriminator: LinkedService.type.polymorphicDiscriminator, uberParent: "LinkedService", - className: "JiraLinkedService", + className: "GreenplumLinkedService", modelProperties: { ...LinkedService.type.modelProperties, - host: { - required: true, - serializedName: "typeProperties.host", - type: { - name: "Object" - } - }, - port: { - serializedName: "typeProperties.port", - type: { - name: "Object" - } - }, - username: { - required: true, - serializedName: "typeProperties.username", + connectionString: { + serializedName: "typeProperties.connectionString", type: { name: "Object" } }, - password: { - serializedName: "typeProperties.password", + pwd: { + serializedName: "typeProperties.pwd", type: { name: "Composite", - className: "SecretBase" - } - }, - useEncryptedEndpoints: { - serializedName: "typeProperties.useEncryptedEndpoints", - type: { - name: "Object" - } - }, - useHostVerification: { - serializedName: "typeProperties.useHostVerification", - type: { - name: "Object" - } - }, - usePeerVerification: { - serializedName: "typeProperties.usePeerVerification", - type: { - name: "Object" + className: "AzureKeyVaultSecretReference" } }, encryptedCredential: { @@ -6021,24 +5741,30 @@ export const JiraLinkedService: msRest.CompositeMapper = { } }; -export const ImpalaLinkedService: msRest.CompositeMapper = { - serializedName: "Impala", +export const GoogleBigQueryLinkedService: msRest.CompositeMapper = { + serializedName: "GoogleBigQuery", type: { name: "Composite", polymorphicDiscriminator: LinkedService.type.polymorphicDiscriminator, uberParent: "LinkedService", - className: "ImpalaLinkedService", + className: "GoogleBigQueryLinkedService", modelProperties: { ...LinkedService.type.modelProperties, - host: { + project: { required: true, - serializedName: "typeProperties.host", + serializedName: "typeProperties.project", type: { name: "Object" } }, - port: { - serializedName: "typeProperties.port", + additionalProjects: { + serializedName: "typeProperties.additionalProjects", + type: { + name: "Object" + } + }, + requestGoogleDriveScope: { + serializedName: "typeProperties.requestGoogleDriveScope", type: { name: "Object" } @@ -6050,45 +5776,46 @@ export const ImpalaLinkedService: msRest.CompositeMapper = { name: "String" } }, - username: { - serializedName: "typeProperties.username", - type: { - name: "Object" - } - }, - password: { - serializedName: "typeProperties.password", + refreshToken: { + serializedName: "typeProperties.refreshToken", type: { name: "Composite", className: "SecretBase" } }, - enableSsl: { - serializedName: "typeProperties.enableSsl", + clientId: { + serializedName: "typeProperties.clientId", type: { name: "Object" } }, - trustedCertPath: { - serializedName: "typeProperties.trustedCertPath", + clientSecret: { + serializedName: "typeProperties.clientSecret", + type: { + name: "Composite", + className: "SecretBase" + } + }, + email: { + serializedName: "typeProperties.email", type: { name: "Object" } }, - useSystemTrustStore: { - serializedName: "typeProperties.useSystemTrustStore", + keyFilePath: { + serializedName: "typeProperties.keyFilePath", type: { name: "Object" } }, - allowHostNameCNMismatch: { - serializedName: "typeProperties.allowHostNameCNMismatch", + trustedCertPath: { + serializedName: "typeProperties.trustedCertPath", type: { name: "Object" } }, - allowSelfSignedServerCert: { - serializedName: "typeProperties.allowSelfSignedServerCert", + useSystemTrustStore: { + serializedName: "typeProperties.useSystemTrustStore", type: { name: "Object" } @@ -6104,38 +5831,31 @@ export const ImpalaLinkedService: msRest.CompositeMapper = { } }; -export const HubspotLinkedService: msRest.CompositeMapper = { - serializedName: "Hubspot", +export const EloquaLinkedService: msRest.CompositeMapper = { + serializedName: "Eloqua", type: { name: "Composite", polymorphicDiscriminator: LinkedService.type.polymorphicDiscriminator, uberParent: "LinkedService", - className: "HubspotLinkedService", + className: "EloquaLinkedService", modelProperties: { ...LinkedService.type.modelProperties, - clientId: { + endpoint: { required: true, - serializedName: "typeProperties.clientId", + serializedName: "typeProperties.endpoint", type: { name: "Object" } }, - clientSecret: { - serializedName: "typeProperties.clientSecret", - type: { - name: "Composite", - className: "SecretBase" - } - }, - accessToken: { - serializedName: "typeProperties.accessToken", + username: { + required: true, + serializedName: "typeProperties.username", type: { - name: "Composite", - className: "SecretBase" + name: "Object" } }, - refreshToken: { - serializedName: "typeProperties.refreshToken", + password: { + serializedName: "typeProperties.password", type: { name: "Composite", className: "SecretBase" @@ -6170,116 +5890,63 @@ export const HubspotLinkedService: msRest.CompositeMapper = { } }; -export const HiveLinkedService: msRest.CompositeMapper = { - serializedName: "Hive", +export const DrillLinkedService: msRest.CompositeMapper = { + serializedName: "Drill", type: { name: "Composite", polymorphicDiscriminator: LinkedService.type.polymorphicDiscriminator, uberParent: "LinkedService", - className: "HiveLinkedService", + className: "DrillLinkedService", modelProperties: { ...LinkedService.type.modelProperties, - host: { - required: true, - serializedName: "typeProperties.host", - type: { - name: "Object" - } - }, - port: { - serializedName: "typeProperties.port", + connectionString: { + serializedName: "typeProperties.connectionString", type: { name: "Object" } }, - serverType: { - serializedName: "typeProperties.serverType", + pwd: { + serializedName: "typeProperties.pwd", type: { - name: "String" + name: "Composite", + className: "AzureKeyVaultSecretReference" } }, - thriftTransportProtocol: { - serializedName: "typeProperties.thriftTransportProtocol", + encryptedCredential: { + serializedName: "typeProperties.encryptedCredential", type: { - name: "String" + name: "Object" } - }, - authenticationType: { - required: true, - serializedName: "typeProperties.authenticationType", + } + }, + additionalProperties: LinkedService.type.additionalProperties + } +}; + +export const CouchbaseLinkedService: msRest.CompositeMapper = { + serializedName: "Couchbase", + type: { + name: "Composite", + polymorphicDiscriminator: LinkedService.type.polymorphicDiscriminator, + uberParent: "LinkedService", + className: "CouchbaseLinkedService", + modelProperties: { + ...LinkedService.type.modelProperties, + connectionString: { + serializedName: "typeProperties.connectionString", type: { - name: "String" + name: "Object" } }, - serviceDiscoveryMode: { - serializedName: "typeProperties.serviceDiscoveryMode", + credString: { + serializedName: "typeProperties.credString", type: { - name: "Object" + name: "Composite", + className: "AzureKeyVaultSecretReference" } }, - zooKeeperNameSpace: { - serializedName: "typeProperties.zooKeeperNameSpace", - type: { - name: "Object" - } - }, - useNativeQuery: { - serializedName: "typeProperties.useNativeQuery", - type: { - name: "Object" - } - }, - username: { - serializedName: "typeProperties.username", - type: { - name: "Object" - } - }, - password: { - serializedName: "typeProperties.password", - type: { - name: "Composite", - className: "SecretBase" - } - }, - httpPath: { - serializedName: "typeProperties.httpPath", - type: { - name: "Object" - } - }, - enableSsl: { - serializedName: "typeProperties.enableSsl", - type: { - name: "Object" - } - }, - trustedCertPath: { - serializedName: "typeProperties.trustedCertPath", - type: { - name: "Object" - } - }, - useSystemTrustStore: { - serializedName: "typeProperties.useSystemTrustStore", - type: { - name: "Object" - } - }, - allowHostNameCNMismatch: { - serializedName: "typeProperties.allowHostNameCNMismatch", - type: { - name: "Object" - } - }, - allowSelfSignedServerCert: { - serializedName: "typeProperties.allowSelfSignedServerCert", - type: { - name: "Object" - } - }, - encryptedCredential: { - serializedName: "typeProperties.encryptedCredential", + encryptedCredential: { + serializedName: "typeProperties.encryptedCredential", type: { name: "Object" } @@ -6289,42 +5956,24 @@ export const HiveLinkedService: msRest.CompositeMapper = { } }; -export const HBaseLinkedService: msRest.CompositeMapper = { - serializedName: "HBase", +export const ConcurLinkedService: msRest.CompositeMapper = { + serializedName: "Concur", type: { name: "Composite", polymorphicDiscriminator: LinkedService.type.polymorphicDiscriminator, uberParent: "LinkedService", - className: "HBaseLinkedService", + className: "ConcurLinkedService", modelProperties: { ...LinkedService.type.modelProperties, - host: { + clientId: { required: true, - serializedName: "typeProperties.host", - type: { - name: "Object" - } - }, - port: { - serializedName: "typeProperties.port", - type: { - name: "Object" - } - }, - httpPath: { - serializedName: "typeProperties.httpPath", + serializedName: "typeProperties.clientId", type: { name: "Object" } }, - authenticationType: { - required: true, - serializedName: "typeProperties.authenticationType", - type: { - name: "String" - } - }, username: { + required: true, serializedName: "typeProperties.username", type: { name: "Object" @@ -6337,26 +5986,20 @@ export const HBaseLinkedService: msRest.CompositeMapper = { className: "SecretBase" } }, - enableSsl: { - serializedName: "typeProperties.enableSsl", - type: { - name: "Object" - } - }, - trustedCertPath: { - serializedName: "typeProperties.trustedCertPath", + useEncryptedEndpoints: { + serializedName: "typeProperties.useEncryptedEndpoints", type: { name: "Object" } }, - allowHostNameCNMismatch: { - serializedName: "typeProperties.allowHostNameCNMismatch", + useHostVerification: { + serializedName: "typeProperties.useHostVerification", type: { name: "Object" } }, - allowSelfSignedServerCert: { - serializedName: "typeProperties.allowSelfSignedServerCert", + usePeerVerification: { + serializedName: "typeProperties.usePeerVerification", type: { name: "Object" } @@ -6372,13 +6015,13 @@ export const HBaseLinkedService: msRest.CompositeMapper = { } }; -export const GreenplumLinkedService: msRest.CompositeMapper = { - serializedName: "Greenplum", +export const AzurePostgreSqlLinkedService: msRest.CompositeMapper = { + serializedName: "AzurePostgreSql", type: { name: "Composite", polymorphicDiscriminator: LinkedService.type.polymorphicDiscriminator, uberParent: "LinkedService", - className: "GreenplumLinkedService", + className: "AzurePostgreSqlLinkedService", modelProperties: { ...LinkedService.type.modelProperties, connectionString: { @@ -6387,8 +6030,8 @@ export const GreenplumLinkedService: msRest.CompositeMapper = { name: "Object" } }, - pwd: { - serializedName: "typeProperties.pwd", + password: { + serializedName: "typeProperties.password", type: { name: "Composite", className: "AzureKeyVaultSecretReference" @@ -6405,81 +6048,71 @@ export const GreenplumLinkedService: msRest.CompositeMapper = { } }; -export const GoogleBigQueryLinkedService: msRest.CompositeMapper = { - serializedName: "GoogleBigQuery", +export const AmazonMWSLinkedService: msRest.CompositeMapper = { + serializedName: "AmazonMWS", type: { name: "Composite", polymorphicDiscriminator: LinkedService.type.polymorphicDiscriminator, uberParent: "LinkedService", - className: "GoogleBigQueryLinkedService", + className: "AmazonMWSLinkedService", modelProperties: { ...LinkedService.type.modelProperties, - project: { + endpoint: { required: true, - serializedName: "typeProperties.project", - type: { - name: "Object" - } - }, - additionalProjects: { - serializedName: "typeProperties.additionalProjects", + serializedName: "typeProperties.endpoint", type: { name: "Object" } }, - requestGoogleDriveScope: { - serializedName: "typeProperties.requestGoogleDriveScope", + marketplaceID: { + required: true, + serializedName: "typeProperties.marketplaceID", type: { name: "Object" } }, - authenticationType: { + sellerID: { required: true, - serializedName: "typeProperties.authenticationType", + serializedName: "typeProperties.sellerID", type: { - name: "String" + name: "Object" } }, - refreshToken: { - serializedName: "typeProperties.refreshToken", + mwsAuthToken: { + serializedName: "typeProperties.mwsAuthToken", type: { name: "Composite", className: "SecretBase" } }, - clientId: { - serializedName: "typeProperties.clientId", + accessKeyId: { + required: true, + serializedName: "typeProperties.accessKeyId", type: { name: "Object" } }, - clientSecret: { - serializedName: "typeProperties.clientSecret", + secretKey: { + serializedName: "typeProperties.secretKey", type: { name: "Composite", className: "SecretBase" } }, - email: { - serializedName: "typeProperties.email", - type: { - name: "Object" - } - }, - keyFilePath: { - serializedName: "typeProperties.keyFilePath", + useEncryptedEndpoints: { + serializedName: "typeProperties.useEncryptedEndpoints", type: { name: "Object" } }, - trustedCertPath: { - serializedName: "typeProperties.trustedCertPath", + useHostVerification: { + serializedName: "typeProperties.useHostVerification", type: { name: "Object" } }, - useSystemTrustStore: { - serializedName: "typeProperties.useSystemTrustStore", + usePeerVerification: { + serializedName: "typeProperties.usePeerVerification", type: { name: "Object" } @@ -6495,52 +6128,45 @@ export const GoogleBigQueryLinkedService: msRest.CompositeMapper = { } }; -export const EloquaLinkedService: msRest.CompositeMapper = { - serializedName: "Eloqua", +export const SapHanaLinkedService: msRest.CompositeMapper = { + serializedName: "SapHana", type: { name: "Composite", polymorphicDiscriminator: LinkedService.type.polymorphicDiscriminator, uberParent: "LinkedService", - className: "EloquaLinkedService", + className: "SapHanaLinkedService", modelProperties: { ...LinkedService.type.modelProperties, - endpoint: { - required: true, - serializedName: "typeProperties.endpoint", + connectionString: { + serializedName: "typeProperties.connectionString", type: { name: "Object" } }, - username: { + server: { required: true, - serializedName: "typeProperties.username", + serializedName: "typeProperties.server", type: { name: "Object" } }, - password: { - serializedName: "typeProperties.password", - type: { - name: "Composite", - className: "SecretBase" - } - }, - useEncryptedEndpoints: { - serializedName: "typeProperties.useEncryptedEndpoints", + authenticationType: { + serializedName: "typeProperties.authenticationType", type: { - name: "Object" + name: "String" } }, - useHostVerification: { - serializedName: "typeProperties.useHostVerification", + userName: { + serializedName: "typeProperties.userName", type: { name: "Object" } }, - usePeerVerification: { - serializedName: "typeProperties.usePeerVerification", + password: { + serializedName: "typeProperties.password", type: { - name: "Object" + name: "Composite", + className: "SecretBase" } }, encryptedCredential: { @@ -6554,26 +6180,47 @@ export const EloquaLinkedService: msRest.CompositeMapper = { } }; -export const DrillLinkedService: msRest.CompositeMapper = { - serializedName: "Drill", +export const SapBWLinkedService: msRest.CompositeMapper = { + serializedName: "SapBW", type: { name: "Composite", polymorphicDiscriminator: LinkedService.type.polymorphicDiscriminator, uberParent: "LinkedService", - className: "DrillLinkedService", + className: "SapBWLinkedService", modelProperties: { ...LinkedService.type.modelProperties, - connectionString: { - serializedName: "typeProperties.connectionString", + server: { + required: true, + serializedName: "typeProperties.server", type: { name: "Object" } }, - pwd: { - serializedName: "typeProperties.pwd", + systemNumber: { + required: true, + serializedName: "typeProperties.systemNumber", + type: { + name: "Object" + } + }, + clientId: { + required: true, + serializedName: "typeProperties.clientId", + type: { + name: "Object" + } + }, + userName: { + serializedName: "typeProperties.userName", + type: { + name: "Object" + } + }, + password: { + serializedName: "typeProperties.password", type: { name: "Composite", - className: "AzureKeyVaultSecretReference" + className: "SecretBase" } }, encryptedCredential: { @@ -6587,58 +6234,36 @@ export const DrillLinkedService: msRest.CompositeMapper = { } }; -export const CouchbaseLinkedService: msRest.CompositeMapper = { - serializedName: "Couchbase", +export const SftpServerLinkedService: msRest.CompositeMapper = { + serializedName: "Sftp", type: { name: "Composite", polymorphicDiscriminator: LinkedService.type.polymorphicDiscriminator, uberParent: "LinkedService", - className: "CouchbaseLinkedService", + className: "SftpServerLinkedService", modelProperties: { ...LinkedService.type.modelProperties, - connectionString: { - serializedName: "typeProperties.connectionString", + host: { + required: true, + serializedName: "typeProperties.host", type: { name: "Object" } }, - credString: { - serializedName: "typeProperties.credString", - type: { - name: "Composite", - className: "AzureKeyVaultSecretReference" - } - }, - encryptedCredential: { - serializedName: "typeProperties.encryptedCredential", + port: { + serializedName: "typeProperties.port", type: { name: "Object" } - } - }, - additionalProperties: LinkedService.type.additionalProperties - } -}; - -export const ConcurLinkedService: msRest.CompositeMapper = { - serializedName: "Concur", - type: { - name: "Composite", - polymorphicDiscriminator: LinkedService.type.polymorphicDiscriminator, - uberParent: "LinkedService", - className: "ConcurLinkedService", - modelProperties: { - ...LinkedService.type.modelProperties, - clientId: { - required: true, - serializedName: "typeProperties.clientId", + }, + authenticationType: { + serializedName: "typeProperties.authenticationType", type: { - name: "Object" + name: "String" } }, - username: { - required: true, - serializedName: "typeProperties.username", + userName: { + serializedName: "typeProperties.userName", type: { name: "Object" } @@ -6650,59 +6275,40 @@ export const ConcurLinkedService: msRest.CompositeMapper = { className: "SecretBase" } }, - useEncryptedEndpoints: { - serializedName: "typeProperties.useEncryptedEndpoints", + encryptedCredential: { + serializedName: "typeProperties.encryptedCredential", type: { name: "Object" } }, - useHostVerification: { - serializedName: "typeProperties.useHostVerification", + privateKeyPath: { + serializedName: "typeProperties.privateKeyPath", type: { name: "Object" } }, - usePeerVerification: { - serializedName: "typeProperties.usePeerVerification", + privateKeyContent: { + serializedName: "typeProperties.privateKeyContent", type: { - name: "Object" + name: "Composite", + className: "SecretBase" } }, - encryptedCredential: { - serializedName: "typeProperties.encryptedCredential", - type: { - name: "Object" - } - } - }, - additionalProperties: LinkedService.type.additionalProperties - } -}; - -export const AzurePostgreSqlLinkedService: msRest.CompositeMapper = { - serializedName: "AzurePostgreSql", - type: { - name: "Composite", - polymorphicDiscriminator: LinkedService.type.polymorphicDiscriminator, - uberParent: "LinkedService", - className: "AzurePostgreSqlLinkedService", - modelProperties: { - ...LinkedService.type.modelProperties, - connectionString: { - serializedName: "typeProperties.connectionString", + passPhrase: { + serializedName: "typeProperties.passPhrase", type: { - name: "Object" + name: "Composite", + className: "SecretBase" } }, - password: { - serializedName: "typeProperties.password", + skipHostKeyValidation: { + serializedName: "typeProperties.skipHostKeyValidation", type: { - name: "Composite", - className: "AzureKeyVaultSecretReference" + name: "Object" } }, - encryptedCredential: { - serializedName: "typeProperties.encryptedCredential", + hostKeyFingerprint: { + serializedName: "typeProperties.hostKeyFingerprint", type: { name: "Object" } @@ -6712,77 +6318,61 @@ export const AzurePostgreSqlLinkedService: msRest.CompositeMapper = { } }; -export const AmazonMWSLinkedService: msRest.CompositeMapper = { - serializedName: "AmazonMWS", +export const FtpServerLinkedService: msRest.CompositeMapper = { + serializedName: "FtpServer", type: { name: "Composite", polymorphicDiscriminator: LinkedService.type.polymorphicDiscriminator, uberParent: "LinkedService", - className: "AmazonMWSLinkedService", + className: "FtpServerLinkedService", modelProperties: { ...LinkedService.type.modelProperties, - endpoint: { - required: true, - serializedName: "typeProperties.endpoint", - type: { - name: "Object" - } - }, - marketplaceID: { + host: { required: true, - serializedName: "typeProperties.marketplaceID", + serializedName: "typeProperties.host", type: { name: "Object" } }, - sellerID: { - required: true, - serializedName: "typeProperties.sellerID", + port: { + serializedName: "typeProperties.port", type: { name: "Object" } }, - mwsAuthToken: { - serializedName: "typeProperties.mwsAuthToken", + authenticationType: { + serializedName: "typeProperties.authenticationType", type: { - name: "Composite", - className: "SecretBase" + name: "String" } }, - accessKeyId: { - required: true, - serializedName: "typeProperties.accessKeyId", + userName: { + serializedName: "typeProperties.userName", type: { name: "Object" } }, - secretKey: { - serializedName: "typeProperties.secretKey", + password: { + serializedName: "typeProperties.password", type: { name: "Composite", className: "SecretBase" } }, - useEncryptedEndpoints: { - serializedName: "typeProperties.useEncryptedEndpoints", - type: { - name: "Object" - } - }, - useHostVerification: { - serializedName: "typeProperties.useHostVerification", + encryptedCredential: { + serializedName: "typeProperties.encryptedCredential", type: { name: "Object" } }, - usePeerVerification: { - serializedName: "typeProperties.usePeerVerification", + enableSsl: { + serializedName: "typeProperties.enableSsl", type: { name: "Object" } }, - encryptedCredential: { - serializedName: "typeProperties.encryptedCredential", + enableServerCertificateValidation: { + serializedName: "typeProperties.enableServerCertificateValidation", type: { name: "Object" } @@ -6792,24 +6382,18 @@ export const AmazonMWSLinkedService: msRest.CompositeMapper = { } }; -export const SapHanaLinkedService: msRest.CompositeMapper = { - serializedName: "SapHana", +export const HttpLinkedService: msRest.CompositeMapper = { + serializedName: "HttpServer", type: { name: "Composite", polymorphicDiscriminator: LinkedService.type.polymorphicDiscriminator, uberParent: "LinkedService", - className: "SapHanaLinkedService", + className: "HttpLinkedService", modelProperties: { ...LinkedService.type.modelProperties, - connectionString: { - serializedName: "typeProperties.connectionString", - type: { - name: "Object" - } - }, - server: { + url: { required: true, - serializedName: "typeProperties.server", + serializedName: "typeProperties.url", type: { name: "Object" } @@ -6833,55 +6417,53 @@ export const SapHanaLinkedService: msRest.CompositeMapper = { className: "SecretBase" } }, + embeddedCertData: { + serializedName: "typeProperties.embeddedCertData", + type: { + name: "Object" + } + }, + certThumbprint: { + serializedName: "typeProperties.certThumbprint", + type: { + name: "Object" + } + }, encryptedCredential: { serializedName: "typeProperties.encryptedCredential", type: { name: "Object" } + }, + enableServerCertificateValidation: { + serializedName: "typeProperties.enableServerCertificateValidation", + type: { + name: "Object" + } } }, additionalProperties: LinkedService.type.additionalProperties } }; -export const SapBWLinkedService: msRest.CompositeMapper = { - serializedName: "SapBW", +export const AzureSearchLinkedService: msRest.CompositeMapper = { + serializedName: "AzureSearch", type: { name: "Composite", polymorphicDiscriminator: LinkedService.type.polymorphicDiscriminator, uberParent: "LinkedService", - className: "SapBWLinkedService", + className: "AzureSearchLinkedService", modelProperties: { ...LinkedService.type.modelProperties, - server: { - required: true, - serializedName: "typeProperties.server", - type: { - name: "Object" - } - }, - systemNumber: { - required: true, - serializedName: "typeProperties.systemNumber", - type: { - name: "Object" - } - }, - clientId: { + url: { required: true, - serializedName: "typeProperties.clientId", - type: { - name: "Object" - } - }, - userName: { - serializedName: "typeProperties.userName", + serializedName: "typeProperties.url", type: { name: "Object" } }, - password: { - serializedName: "typeProperties.password", + key: { + serializedName: "typeProperties.key", type: { name: "Composite", className: "SecretBase" @@ -6898,36 +6480,45 @@ export const SapBWLinkedService: msRest.CompositeMapper = { } }; -export const SftpServerLinkedService: msRest.CompositeMapper = { - serializedName: "Sftp", +export const CustomDataSourceLinkedService: msRest.CompositeMapper = { + serializedName: "CustomDataSource", type: { name: "Composite", polymorphicDiscriminator: LinkedService.type.polymorphicDiscriminator, uberParent: "LinkedService", - className: "SftpServerLinkedService", + className: "CustomDataSourceLinkedService", modelProperties: { ...LinkedService.type.modelProperties, - host: { + typeProperties: { required: true, - serializedName: "typeProperties.host", + serializedName: "typeProperties", type: { name: "Object" } - }, - port: { - serializedName: "typeProperties.port", + } + }, + additionalProperties: LinkedService.type.additionalProperties + } +}; + +export const AmazonRedshiftLinkedService: msRest.CompositeMapper = { + serializedName: "AmazonRedshift", + type: { + name: "Composite", + polymorphicDiscriminator: LinkedService.type.polymorphicDiscriminator, + uberParent: "LinkedService", + className: "AmazonRedshiftLinkedService", + modelProperties: { + ...LinkedService.type.modelProperties, + server: { + required: true, + serializedName: "typeProperties.server", type: { name: "Object" } }, - authenticationType: { - serializedName: "typeProperties.authenticationType", - type: { - name: "String" - } - }, - userName: { - serializedName: "typeProperties.userName", + username: { + serializedName: "typeProperties.username", type: { name: "Object" } @@ -6939,40 +6530,60 @@ export const SftpServerLinkedService: msRest.CompositeMapper = { className: "SecretBase" } }, - encryptedCredential: { - serializedName: "typeProperties.encryptedCredential", + database: { + required: true, + serializedName: "typeProperties.database", type: { name: "Object" } }, - privateKeyPath: { - serializedName: "typeProperties.privateKeyPath", + port: { + serializedName: "typeProperties.port", type: { name: "Object" } }, - privateKeyContent: { - serializedName: "typeProperties.privateKeyContent", + encryptedCredential: { + serializedName: "typeProperties.encryptedCredential", type: { - name: "Composite", - className: "SecretBase" + name: "Object" + } + } + }, + additionalProperties: LinkedService.type.additionalProperties + } +}; + +export const AmazonS3LinkedService: msRest.CompositeMapper = { + serializedName: "AmazonS3", + type: { + name: "Composite", + polymorphicDiscriminator: LinkedService.type.polymorphicDiscriminator, + uberParent: "LinkedService", + className: "AmazonS3LinkedService", + modelProperties: { + ...LinkedService.type.modelProperties, + accessKeyId: { + serializedName: "typeProperties.accessKeyId", + type: { + name: "Object" } }, - passPhrase: { - serializedName: "typeProperties.passPhrase", + secretAccessKey: { + serializedName: "typeProperties.secretAccessKey", type: { name: "Composite", className: "SecretBase" } }, - skipHostKeyValidation: { - serializedName: "typeProperties.skipHostKeyValidation", + serviceUrl: { + serializedName: "typeProperties.serviceUrl", type: { name: "Object" } }, - hostKeyFingerprint: { - serializedName: "typeProperties.hostKeyFingerprint", + encryptedCredential: { + serializedName: "typeProperties.encryptedCredential", type: { name: "Object" } @@ -6982,29 +6593,30 @@ export const SftpServerLinkedService: msRest.CompositeMapper = { } }; -export const FtpServerLinkedService: msRest.CompositeMapper = { - serializedName: "FtpServer", +export const RestServiceLinkedService: msRest.CompositeMapper = { + serializedName: "RestService", type: { name: "Composite", polymorphicDiscriminator: LinkedService.type.polymorphicDiscriminator, uberParent: "LinkedService", - className: "FtpServerLinkedService", + className: "RestServiceLinkedService", modelProperties: { ...LinkedService.type.modelProperties, - host: { + url: { required: true, - serializedName: "typeProperties.host", + serializedName: "typeProperties.url", type: { name: "Object" } }, - port: { - serializedName: "typeProperties.port", + enableServerCertificateValidation: { + serializedName: "typeProperties.enableServerCertificateValidation", type: { name: "Object" } }, authenticationType: { + required: true, serializedName: "typeProperties.authenticationType", type: { name: "String" @@ -7023,20 +6635,33 @@ export const FtpServerLinkedService: msRest.CompositeMapper = { className: "SecretBase" } }, - encryptedCredential: { - serializedName: "typeProperties.encryptedCredential", + servicePrincipalId: { + serializedName: "typeProperties.servicePrincipalId", type: { name: "Object" } }, - enableSsl: { - serializedName: "typeProperties.enableSsl", + servicePrincipalKey: { + serializedName: "typeProperties.servicePrincipalKey", + type: { + name: "Composite", + className: "SecretBase" + } + }, + tenant: { + serializedName: "typeProperties.tenant", type: { name: "Object" } }, - enableServerCertificateValidation: { - serializedName: "typeProperties.enableServerCertificateValidation", + aadResourceId: { + serializedName: "typeProperties.aadResourceId", + type: { + name: "Object" + } + }, + encryptedCredential: { + serializedName: "typeProperties.encryptedCredential", type: { name: "Object" } @@ -7046,51 +6671,53 @@ export const FtpServerLinkedService: msRest.CompositeMapper = { } }; -export const HttpLinkedService: msRest.CompositeMapper = { - serializedName: "HttpServer", +export const SapOpenHubLinkedService: msRest.CompositeMapper = { + serializedName: "SapOpenHub", type: { name: "Composite", polymorphicDiscriminator: LinkedService.type.polymorphicDiscriminator, uberParent: "LinkedService", - className: "HttpLinkedService", + className: "SapOpenHubLinkedService", modelProperties: { ...LinkedService.type.modelProperties, - url: { + server: { required: true, - serializedName: "typeProperties.url", + serializedName: "typeProperties.server", type: { name: "Object" } }, - authenticationType: { - serializedName: "typeProperties.authenticationType", + systemNumber: { + required: true, + serializedName: "typeProperties.systemNumber", type: { - name: "String" + name: "Object" } }, - userName: { - serializedName: "typeProperties.userName", + clientId: { + required: true, + serializedName: "typeProperties.clientId", type: { name: "Object" } }, - password: { - serializedName: "typeProperties.password", + language: { + serializedName: "typeProperties.language", type: { - name: "Composite", - className: "SecretBase" + name: "Object" } }, - embeddedCertData: { - serializedName: "typeProperties.embeddedCertData", + userName: { + serializedName: "typeProperties.userName", type: { name: "Object" } }, - certThumbprint: { - serializedName: "typeProperties.certThumbprint", + password: { + serializedName: "typeProperties.password", type: { - name: "Object" + name: "Composite", + className: "SecretBase" } }, encryptedCredential: { @@ -7098,36 +6725,36 @@ export const HttpLinkedService: msRest.CompositeMapper = { type: { name: "Object" } - }, - enableServerCertificateValidation: { - serializedName: "typeProperties.enableServerCertificateValidation", - type: { - name: "Object" - } } }, additionalProperties: LinkedService.type.additionalProperties } }; -export const AzureSearchLinkedService: msRest.CompositeMapper = { - serializedName: "AzureSearch", +export const SapEccLinkedService: msRest.CompositeMapper = { + serializedName: "SapEcc", type: { name: "Composite", polymorphicDiscriminator: LinkedService.type.polymorphicDiscriminator, uberParent: "LinkedService", - className: "AzureSearchLinkedService", + className: "SapEccLinkedService", modelProperties: { ...LinkedService.type.modelProperties, url: { required: true, serializedName: "typeProperties.url", type: { - name: "Object" + name: "String" } }, - key: { - serializedName: "typeProperties.key", + username: { + serializedName: "typeProperties.username", + type: { + name: "String" + } + }, + password: { + serializedName: "typeProperties.password", type: { name: "Composite", className: "SecretBase" @@ -7136,7 +6763,7 @@ export const AzureSearchLinkedService: msRest.CompositeMapper = { encryptedCredential: { serializedName: "typeProperties.encryptedCredential", type: { - name: "Object" + name: "String" } } }, @@ -7144,18 +6771,37 @@ export const AzureSearchLinkedService: msRest.CompositeMapper = { } }; -export const CustomDataSourceLinkedService: msRest.CompositeMapper = { - serializedName: "CustomDataSource", +export const SapCloudForCustomerLinkedService: msRest.CompositeMapper = { + serializedName: "SapCloudForCustomer", type: { name: "Composite", polymorphicDiscriminator: LinkedService.type.polymorphicDiscriminator, uberParent: "LinkedService", - className: "CustomDataSourceLinkedService", + className: "SapCloudForCustomerLinkedService", modelProperties: { ...LinkedService.type.modelProperties, - typeProperties: { + url: { required: true, - serializedName: "typeProperties", + serializedName: "typeProperties.url", + type: { + name: "Object" + } + }, + username: { + serializedName: "typeProperties.username", + type: { + name: "Object" + } + }, + password: { + serializedName: "typeProperties.password", + type: { + name: "Composite", + className: "SecretBase" + } + }, + encryptedCredential: { + serializedName: "typeProperties.encryptedCredential", type: { name: "Object" } @@ -7165,18 +6811,17 @@ export const CustomDataSourceLinkedService: msRest.CompositeMapper = { } }; -export const AmazonRedshiftLinkedService: msRest.CompositeMapper = { - serializedName: "AmazonRedshift", +export const SalesforceServiceCloudLinkedService: msRest.CompositeMapper = { + serializedName: "SalesforceServiceCloud", type: { name: "Composite", polymorphicDiscriminator: LinkedService.type.polymorphicDiscriminator, uberParent: "LinkedService", - className: "AmazonRedshiftLinkedService", + className: "SalesforceServiceCloudLinkedService", modelProperties: { ...LinkedService.type.modelProperties, - server: { - required: true, - serializedName: "typeProperties.server", + environmentUrl: { + serializedName: "typeProperties.environmentUrl", type: { name: "Object" } @@ -7194,15 +6839,21 @@ export const AmazonRedshiftLinkedService: msRest.CompositeMapper = { className: "SecretBase" } }, - database: { - required: true, - serializedName: "typeProperties.database", + securityToken: { + serializedName: "typeProperties.securityToken", + type: { + name: "Composite", + className: "SecretBase" + } + }, + apiVersion: { + serializedName: "typeProperties.apiVersion", type: { name: "Object" } }, - port: { - serializedName: "typeProperties.port", + extendedProperties: { + serializedName: "typeProperties.extendedProperties", type: { name: "Object" } @@ -7218,30 +6869,43 @@ export const AmazonRedshiftLinkedService: msRest.CompositeMapper = { } }; -export const AmazonS3LinkedService: msRest.CompositeMapper = { - serializedName: "AmazonS3", +export const SalesforceLinkedService: msRest.CompositeMapper = { + serializedName: "Salesforce", type: { name: "Composite", polymorphicDiscriminator: LinkedService.type.polymorphicDiscriminator, uberParent: "LinkedService", - className: "AmazonS3LinkedService", + className: "SalesforceLinkedService", modelProperties: { ...LinkedService.type.modelProperties, - accessKeyId: { - serializedName: "typeProperties.accessKeyId", + environmentUrl: { + serializedName: "typeProperties.environmentUrl", type: { name: "Object" } }, - secretAccessKey: { - serializedName: "typeProperties.secretAccessKey", + username: { + serializedName: "typeProperties.username", + type: { + name: "Object" + } + }, + password: { + serializedName: "typeProperties.password", type: { name: "Composite", className: "SecretBase" } }, - serviceUrl: { - serializedName: "typeProperties.serviceUrl", + securityToken: { + serializedName: "typeProperties.securityToken", + type: { + name: "Composite", + className: "SecretBase" + } + }, + apiVersion: { + serializedName: "typeProperties.apiVersion", type: { name: "Object" } @@ -7257,46 +6921,75 @@ export const AmazonS3LinkedService: msRest.CompositeMapper = { } }; -export const RestServiceLinkedService: msRest.CompositeMapper = { - serializedName: "RestService", +export const Office365LinkedService: msRest.CompositeMapper = { + serializedName: "Office365", type: { name: "Composite", polymorphicDiscriminator: LinkedService.type.polymorphicDiscriminator, uberParent: "LinkedService", - className: "RestServiceLinkedService", + className: "Office365LinkedService", modelProperties: { ...LinkedService.type.modelProperties, - url: { + office365TenantId: { required: true, - serializedName: "typeProperties.url", + serializedName: "typeProperties.office365TenantId", type: { name: "Object" } }, - enableServerCertificateValidation: { - serializedName: "typeProperties.enableServerCertificateValidation", - type: { + servicePrincipalTenantId: { + required: true, + serializedName: "typeProperties.servicePrincipalTenantId", + type: { name: "Object" } }, - authenticationType: { + servicePrincipalId: { required: true, - serializedName: "typeProperties.authenticationType", + serializedName: "typeProperties.servicePrincipalId", type: { - name: "String" + name: "Object" } }, - userName: { - serializedName: "typeProperties.userName", + servicePrincipalKey: { + required: true, + serializedName: "typeProperties.servicePrincipalKey", + type: { + name: "Composite", + className: "SecretBase" + } + }, + encryptedCredential: { + serializedName: "typeProperties.encryptedCredential", + type: { + name: "Object" + } + } + }, + additionalProperties: LinkedService.type.additionalProperties + } +}; + +export const AzureBlobFSLinkedService: msRest.CompositeMapper = { + serializedName: "AzureBlobFS", + type: { + name: "Composite", + polymorphicDiscriminator: LinkedService.type.polymorphicDiscriminator, + uberParent: "LinkedService", + className: "AzureBlobFSLinkedService", + modelProperties: { + ...LinkedService.type.modelProperties, + url: { + required: true, + serializedName: "typeProperties.url", type: { name: "Object" } }, - password: { - serializedName: "typeProperties.password", + accountKey: { + serializedName: "typeProperties.accountKey", type: { - name: "Composite", - className: "SecretBase" + name: "Object" } }, servicePrincipalId: { @@ -7318,12 +7011,6 @@ export const RestServiceLinkedService: msRest.CompositeMapper = { name: "Object" } }, - aadResourceId: { - serializedName: "typeProperties.aadResourceId", - type: { - name: "Object" - } - }, encryptedCredential: { serializedName: "typeProperties.encryptedCredential", type: { @@ -7335,53 +7022,57 @@ export const RestServiceLinkedService: msRest.CompositeMapper = { } }; -export const SapOpenHubLinkedService: msRest.CompositeMapper = { - serializedName: "SapOpenHub", +export const AzureDataLakeStoreLinkedService: msRest.CompositeMapper = { + serializedName: "AzureDataLakeStore", type: { name: "Composite", polymorphicDiscriminator: LinkedService.type.polymorphicDiscriminator, uberParent: "LinkedService", - className: "SapOpenHubLinkedService", + className: "AzureDataLakeStoreLinkedService", modelProperties: { ...LinkedService.type.modelProperties, - server: { + dataLakeStoreUri: { required: true, - serializedName: "typeProperties.server", + serializedName: "typeProperties.dataLakeStoreUri", type: { name: "Object" } }, - systemNumber: { - required: true, - serializedName: "typeProperties.systemNumber", + servicePrincipalId: { + serializedName: "typeProperties.servicePrincipalId", type: { name: "Object" } }, - clientId: { - required: true, - serializedName: "typeProperties.clientId", + servicePrincipalKey: { + serializedName: "typeProperties.servicePrincipalKey", + type: { + name: "Composite", + className: "SecretBase" + } + }, + tenant: { + serializedName: "typeProperties.tenant", type: { name: "Object" } }, - language: { - serializedName: "typeProperties.language", + accountName: { + serializedName: "typeProperties.accountName", type: { name: "Object" } }, - userName: { - serializedName: "typeProperties.userName", + subscriptionId: { + serializedName: "typeProperties.subscriptionId", type: { name: "Object" } }, - password: { - serializedName: "typeProperties.password", + resourceGroupName: { + serializedName: "typeProperties.resourceGroupName", type: { - name: "Composite", - className: "SecretBase" + name: "Object" } }, encryptedCredential: { @@ -7395,39 +7086,27 @@ export const SapOpenHubLinkedService: msRest.CompositeMapper = { } }; -export const SapEccLinkedService: msRest.CompositeMapper = { - serializedName: "SapEcc", +export const CosmosDbMongoDbApiLinkedService: msRest.CompositeMapper = { + serializedName: "CosmosDbMongoDbApi", type: { name: "Composite", polymorphicDiscriminator: LinkedService.type.polymorphicDiscriminator, uberParent: "LinkedService", - className: "SapEccLinkedService", + className: "CosmosDbMongoDbApiLinkedService", modelProperties: { ...LinkedService.type.modelProperties, - url: { + connectionString: { required: true, - serializedName: "typeProperties.url", - type: { - name: "String" - } - }, - username: { - serializedName: "typeProperties.username", - type: { - name: "String" - } - }, - password: { - serializedName: "typeProperties.password", + serializedName: "typeProperties.connectionString", type: { - name: "Composite", - className: "SecretBase" + name: "Object" } }, - encryptedCredential: { - serializedName: "typeProperties.encryptedCredential", + database: { + required: true, + serializedName: "typeProperties.database", type: { - name: "String" + name: "Object" } } }, @@ -7435,37 +7114,25 @@ export const SapEccLinkedService: msRest.CompositeMapper = { } }; -export const SapCloudForCustomerLinkedService: msRest.CompositeMapper = { - serializedName: "SapCloudForCustomer", +export const MongoDbV2LinkedService: msRest.CompositeMapper = { + serializedName: "MongoDbV2", type: { name: "Composite", polymorphicDiscriminator: LinkedService.type.polymorphicDiscriminator, uberParent: "LinkedService", - className: "SapCloudForCustomerLinkedService", + className: "MongoDbV2LinkedService", modelProperties: { ...LinkedService.type.modelProperties, - url: { + connectionString: { required: true, - serializedName: "typeProperties.url", - type: { - name: "Object" - } - }, - username: { - serializedName: "typeProperties.username", + serializedName: "typeProperties.connectionString", type: { name: "Object" } }, - password: { - serializedName: "typeProperties.password", - type: { - name: "Composite", - className: "SecretBase" - } - }, - encryptedCredential: { - serializedName: "typeProperties.encryptedCredential", + database: { + required: true, + serializedName: "typeProperties.database", type: { name: "Object" } @@ -7475,17 +7142,31 @@ export const SapCloudForCustomerLinkedService: msRest.CompositeMapper = { } }; -export const SalesforceServiceCloudLinkedService: msRest.CompositeMapper = { - serializedName: "SalesforceServiceCloud", +export const MongoDbLinkedService: msRest.CompositeMapper = { + serializedName: "MongoDb", type: { name: "Composite", polymorphicDiscriminator: LinkedService.type.polymorphicDiscriminator, uberParent: "LinkedService", - className: "SalesforceServiceCloudLinkedService", + className: "MongoDbLinkedService", modelProperties: { ...LinkedService.type.modelProperties, - environmentUrl: { - serializedName: "typeProperties.environmentUrl", + server: { + required: true, + serializedName: "typeProperties.server", + type: { + name: "Object" + } + }, + authenticationType: { + serializedName: "typeProperties.authenticationType", + type: { + name: "String" + } + }, + databaseName: { + required: true, + serializedName: "typeProperties.databaseName", type: { name: "Object" } @@ -7503,15 +7184,26 @@ export const SalesforceServiceCloudLinkedService: msRest.CompositeMapper = { className: "SecretBase" } }, - securityToken: { - serializedName: "typeProperties.securityToken", + authSource: { + serializedName: "typeProperties.authSource", type: { - name: "Composite", - className: "SecretBase" + name: "Object" } }, - extendedProperties: { - serializedName: "typeProperties.extendedProperties", + port: { + serializedName: "typeProperties.port", + type: { + name: "Object" + } + }, + enableSsl: { + serializedName: "typeProperties.enableSsl", + type: { + name: "Object" + } + }, + allowSelfSignedServerCert: { + serializedName: "typeProperties.allowSelfSignedServerCert", type: { name: "Object" } @@ -7527,17 +7219,30 @@ export const SalesforceServiceCloudLinkedService: msRest.CompositeMapper = { } }; -export const SalesforceLinkedService: msRest.CompositeMapper = { - serializedName: "Salesforce", +export const CassandraLinkedService: msRest.CompositeMapper = { + serializedName: "Cassandra", type: { name: "Composite", polymorphicDiscriminator: LinkedService.type.polymorphicDiscriminator, uberParent: "LinkedService", - className: "SalesforceLinkedService", + className: "CassandraLinkedService", modelProperties: { ...LinkedService.type.modelProperties, - environmentUrl: { - serializedName: "typeProperties.environmentUrl", + host: { + required: true, + serializedName: "typeProperties.host", + type: { + name: "Object" + } + }, + authenticationType: { + serializedName: "typeProperties.authenticationType", + type: { + name: "Object" + } + }, + port: { + serializedName: "typeProperties.port", type: { name: "Object" } @@ -7555,13 +7260,6 @@ export const SalesforceLinkedService: msRest.CompositeMapper = { className: "SecretBase" } }, - securityToken: { - serializedName: "typeProperties.securityToken", - type: { - name: "Composite", - className: "SecretBase" - } - }, encryptedCredential: { serializedName: "typeProperties.encryptedCredential", type: { @@ -7573,48 +7271,120 @@ export const SalesforceLinkedService: msRest.CompositeMapper = { } }; -export const Office365LinkedService: msRest.CompositeMapper = { - serializedName: "Office365", +export const WebLinkedServiceTypeProperties: msRest.CompositeMapper = { + serializedName: "WebLinkedServiceTypeProperties", type: { name: "Composite", - polymorphicDiscriminator: LinkedService.type.polymorphicDiscriminator, - uberParent: "LinkedService", - className: "Office365LinkedService", + polymorphicDiscriminator: { + serializedName: "authenticationType", + clientName: "authenticationType" + }, + uberParent: "WebLinkedServiceTypeProperties", + className: "WebLinkedServiceTypeProperties", modelProperties: { - ...LinkedService.type.modelProperties, - office365TenantId: { + url: { required: true, - serializedName: "typeProperties.office365TenantId", + serializedName: "url", type: { name: "Object" } }, - servicePrincipalTenantId: { + authenticationType: { required: true, - serializedName: "typeProperties.servicePrincipalTenantId", + serializedName: "authenticationType", type: { - name: "Object" + name: "String" + } + } + } + } +}; + +export const WebClientCertificateAuthentication: msRest.CompositeMapper = { + serializedName: "ClientCertificate", + type: { + name: "Composite", + polymorphicDiscriminator: WebLinkedServiceTypeProperties.type.polymorphicDiscriminator, + uberParent: "WebLinkedServiceTypeProperties", + className: "WebClientCertificateAuthentication", + modelProperties: { + ...WebLinkedServiceTypeProperties.type.modelProperties, + pfx: { + required: true, + serializedName: "pfx", + type: { + name: "Composite", + className: "SecretBase" } }, - servicePrincipalId: { + password: { required: true, - serializedName: "typeProperties.servicePrincipalId", + serializedName: "password", + type: { + name: "Composite", + className: "SecretBase" + } + } + } + } +}; + +export const WebBasicAuthentication: msRest.CompositeMapper = { + serializedName: "Basic", + type: { + name: "Composite", + polymorphicDiscriminator: WebLinkedServiceTypeProperties.type.polymorphicDiscriminator, + uberParent: "WebLinkedServiceTypeProperties", + className: "WebBasicAuthentication", + modelProperties: { + ...WebLinkedServiceTypeProperties.type.modelProperties, + username: { + required: true, + serializedName: "username", type: { name: "Object" } }, - servicePrincipalKey: { + password: { required: true, - serializedName: "typeProperties.servicePrincipalKey", + serializedName: "password", type: { name: "Composite", className: "SecretBase" } - }, - encryptedCredential: { - serializedName: "typeProperties.encryptedCredential", + } + } + } +}; + +export const WebAnonymousAuthentication: msRest.CompositeMapper = { + serializedName: "Anonymous", + type: { + name: "Composite", + polymorphicDiscriminator: WebLinkedServiceTypeProperties.type.polymorphicDiscriminator, + uberParent: "WebLinkedServiceTypeProperties", + className: "WebAnonymousAuthentication", + modelProperties: { + ...WebLinkedServiceTypeProperties.type.modelProperties + } + } +}; + +export const WebLinkedService: msRest.CompositeMapper = { + serializedName: "Web", + type: { + name: "Composite", + polymorphicDiscriminator: LinkedService.type.polymorphicDiscriminator, + uberParent: "LinkedService", + className: "WebLinkedService", + modelProperties: { + ...LinkedService.type.modelProperties, + typeProperties: { + required: true, + serializedName: "typeProperties", type: { - name: "Object" + name: "Composite", + className: "WebLinkedServiceTypeProperties" } } }, @@ -7622,13 +7392,13 @@ export const Office365LinkedService: msRest.CompositeMapper = { } }; -export const AzureBlobFSLinkedService: msRest.CompositeMapper = { - serializedName: "AzureBlobFS", +export const ODataLinkedService: msRest.CompositeMapper = { + serializedName: "OData", type: { name: "Composite", polymorphicDiscriminator: LinkedService.type.polymorphicDiscriminator, uberParent: "LinkedService", - className: "AzureBlobFSLinkedService", + className: "ODataLinkedService", modelProperties: { ...LinkedService.type.modelProperties, url: { @@ -7638,20 +7408,20 @@ export const AzureBlobFSLinkedService: msRest.CompositeMapper = { name: "Object" } }, - accountKey: { - serializedName: "typeProperties.accountKey", + authenticationType: { + serializedName: "typeProperties.authenticationType", type: { - name: "Object" + name: "String" } }, - servicePrincipalId: { - serializedName: "typeProperties.servicePrincipalId", + userName: { + serializedName: "typeProperties.userName", type: { name: "Object" } }, - servicePrincipalKey: { - serializedName: "typeProperties.servicePrincipalKey", + password: { + serializedName: "typeProperties.password", type: { name: "Composite", className: "SecretBase" @@ -7663,37 +7433,22 @@ export const AzureBlobFSLinkedService: msRest.CompositeMapper = { name: "Object" } }, - encryptedCredential: { - serializedName: "typeProperties.encryptedCredential", + servicePrincipalId: { + serializedName: "typeProperties.servicePrincipalId", type: { name: "Object" } - } - }, - additionalProperties: LinkedService.type.additionalProperties - } -}; - -export const AzureDataLakeStoreLinkedService: msRest.CompositeMapper = { - serializedName: "AzureDataLakeStore", - type: { - name: "Composite", - polymorphicDiscriminator: LinkedService.type.polymorphicDiscriminator, - uberParent: "LinkedService", - className: "AzureDataLakeStoreLinkedService", - modelProperties: { - ...LinkedService.type.modelProperties, - dataLakeStoreUri: { - required: true, - serializedName: "typeProperties.dataLakeStoreUri", + }, + aadResourceId: { + serializedName: "typeProperties.aadResourceId", type: { name: "Object" } }, - servicePrincipalId: { - serializedName: "typeProperties.servicePrincipalId", + aadServicePrincipalCredentialType: { + serializedName: "typeProperties.aadServicePrincipalCredentialType", type: { - name: "Object" + name: "String" } }, servicePrincipalKey: { @@ -7703,28 +7458,18 @@ export const AzureDataLakeStoreLinkedService: msRest.CompositeMapper = { className: "SecretBase" } }, - tenant: { - serializedName: "typeProperties.tenant", - type: { - name: "Object" - } - }, - accountName: { - serializedName: "typeProperties.accountName", - type: { - name: "Object" - } - }, - subscriptionId: { - serializedName: "typeProperties.subscriptionId", + servicePrincipalEmbeddedCert: { + serializedName: "typeProperties.servicePrincipalEmbeddedCert", type: { - name: "Object" + name: "Composite", + className: "SecretBase" } }, - resourceGroupName: { - serializedName: "typeProperties.resourceGroupName", + servicePrincipalEmbeddedCertPassword: { + serializedName: "typeProperties.servicePrincipalEmbeddedCertPassword", type: { - name: "Object" + name: "Composite", + className: "SecretBase" } }, encryptedCredential: { @@ -7738,74 +7483,64 @@ export const AzureDataLakeStoreLinkedService: msRest.CompositeMapper = { } }; -export const CosmosDbMongoDbApiLinkedService: msRest.CompositeMapper = { - serializedName: "CosmosDbMongoDbApi", +export const HdfsLinkedService: msRest.CompositeMapper = { + serializedName: "Hdfs", type: { name: "Composite", polymorphicDiscriminator: LinkedService.type.polymorphicDiscriminator, uberParent: "LinkedService", - className: "CosmosDbMongoDbApiLinkedService", + className: "HdfsLinkedService", modelProperties: { ...LinkedService.type.modelProperties, - connectionString: { + url: { required: true, - serializedName: "typeProperties.connectionString", + serializedName: "typeProperties.url", type: { name: "Object" } }, - database: { - required: true, - serializedName: "typeProperties.database", + authenticationType: { + serializedName: "typeProperties.authenticationType", type: { name: "Object" } - } - }, - additionalProperties: LinkedService.type.additionalProperties - } -}; - -export const MongoDbV2LinkedService: msRest.CompositeMapper = { - serializedName: "MongoDbV2", - type: { - name: "Composite", - polymorphicDiscriminator: LinkedService.type.polymorphicDiscriminator, - uberParent: "LinkedService", - className: "MongoDbV2LinkedService", - modelProperties: { - ...LinkedService.type.modelProperties, - connectionString: { - required: true, - serializedName: "typeProperties.connectionString", + }, + encryptedCredential: { + serializedName: "typeProperties.encryptedCredential", type: { name: "Object" } }, - database: { - required: true, - serializedName: "typeProperties.database", + userName: { + serializedName: "typeProperties.userName", type: { name: "Object" } + }, + password: { + serializedName: "typeProperties.password", + type: { + name: "Composite", + className: "SecretBase" + } } }, additionalProperties: LinkedService.type.additionalProperties } }; -export const MongoDbLinkedService: msRest.CompositeMapper = { - serializedName: "MongoDb", +export const MicrosoftAccessLinkedService: msRest.CompositeMapper = { + serializedName: "MicrosoftAccess", type: { name: "Composite", polymorphicDiscriminator: LinkedService.type.polymorphicDiscriminator, uberParent: "LinkedService", - className: "MongoDbLinkedService", + className: "MicrosoftAccessLinkedService", modelProperties: { ...LinkedService.type.modelProperties, - server: { + connectionString: { required: true, - serializedName: "typeProperties.server", + serializedName: "typeProperties.connectionString", type: { name: "Object" } @@ -7813,18 +7548,18 @@ export const MongoDbLinkedService: msRest.CompositeMapper = { authenticationType: { serializedName: "typeProperties.authenticationType", type: { - name: "String" + name: "Object" } }, - databaseName: { - required: true, - serializedName: "typeProperties.databaseName", + credential: { + serializedName: "typeProperties.credential", type: { - name: "Object" + name: "Composite", + className: "SecretBase" } }, - username: { - serializedName: "typeProperties.username", + userName: { + serializedName: "typeProperties.userName", type: { name: "Object" } @@ -7836,32 +7571,8 @@ export const MongoDbLinkedService: msRest.CompositeMapper = { className: "SecretBase" } }, - authSource: { - serializedName: "typeProperties.authSource", - type: { - name: "Object" - } - }, - port: { - serializedName: "typeProperties.port", - type: { - name: "Object" - } - }, - enableSsl: { - serializedName: "typeProperties.enableSsl", - type: { - name: "Object" - } - }, - allowSelfSignedServerCert: { - serializedName: "typeProperties.allowSelfSignedServerCert", - type: { - name: "Object" - } - }, - encryptedCredential: { - serializedName: "typeProperties.encryptedCredential", + encryptedCredential: { + serializedName: "typeProperties.encryptedCredential", type: { name: "Object" } @@ -7871,18 +7582,18 @@ export const MongoDbLinkedService: msRest.CompositeMapper = { } }; -export const CassandraLinkedService: msRest.CompositeMapper = { - serializedName: "Cassandra", +export const InformixLinkedService: msRest.CompositeMapper = { + serializedName: "Informix", type: { name: "Composite", polymorphicDiscriminator: LinkedService.type.polymorphicDiscriminator, uberParent: "LinkedService", - className: "CassandraLinkedService", + className: "InformixLinkedService", modelProperties: { ...LinkedService.type.modelProperties, - host: { + connectionString: { required: true, - serializedName: "typeProperties.host", + serializedName: "typeProperties.connectionString", type: { name: "Object" } @@ -7893,14 +7604,15 @@ export const CassandraLinkedService: msRest.CompositeMapper = { name: "Object" } }, - port: { - serializedName: "typeProperties.port", + credential: { + serializedName: "typeProperties.credential", type: { - name: "Object" + name: "Composite", + className: "SecretBase" } }, - username: { - serializedName: "typeProperties.username", + userName: { + serializedName: "typeProperties.userName", type: { name: "Object" } @@ -7923,120 +7635,52 @@ export const CassandraLinkedService: msRest.CompositeMapper = { } }; -export const WebLinkedServiceTypeProperties: msRest.CompositeMapper = { - serializedName: "WebLinkedServiceTypeProperties", +export const OdbcLinkedService: msRest.CompositeMapper = { + serializedName: "Odbc", type: { name: "Composite", - polymorphicDiscriminator: { - serializedName: "authenticationType", - clientName: "authenticationType" - }, - uberParent: "WebLinkedServiceTypeProperties", - className: "WebLinkedServiceTypeProperties", + polymorphicDiscriminator: LinkedService.type.polymorphicDiscriminator, + uberParent: "LinkedService", + className: "OdbcLinkedService", modelProperties: { - url: { + ...LinkedService.type.modelProperties, + connectionString: { required: true, - serializedName: "url", + serializedName: "typeProperties.connectionString", type: { name: "Object" } }, authenticationType: { - required: true, - serializedName: "authenticationType", - type: { - name: "String" - } - } - } - } -}; - -export const WebClientCertificateAuthentication: msRest.CompositeMapper = { - serializedName: "ClientCertificate", - type: { - name: "Composite", - polymorphicDiscriminator: WebLinkedServiceTypeProperties.type.polymorphicDiscriminator, - uberParent: "WebLinkedServiceTypeProperties", - className: "WebClientCertificateAuthentication", - modelProperties: { - ...WebLinkedServiceTypeProperties.type.modelProperties, - pfx: { - required: true, - serializedName: "pfx", + serializedName: "typeProperties.authenticationType", type: { - name: "Composite", - className: "SecretBase" + name: "Object" } }, - password: { - required: true, - serializedName: "password", + credential: { + serializedName: "typeProperties.credential", type: { name: "Composite", className: "SecretBase" } - } - } - } -}; - -export const WebBasicAuthentication: msRest.CompositeMapper = { - serializedName: "Basic", - type: { - name: "Composite", - polymorphicDiscriminator: WebLinkedServiceTypeProperties.type.polymorphicDiscriminator, - uberParent: "WebLinkedServiceTypeProperties", - className: "WebBasicAuthentication", - modelProperties: { - ...WebLinkedServiceTypeProperties.type.modelProperties, - username: { - required: true, - serializedName: "username", + }, + userName: { + serializedName: "typeProperties.userName", type: { name: "Object" } }, password: { - required: true, - serializedName: "password", + serializedName: "typeProperties.password", type: { name: "Composite", className: "SecretBase" } - } - } - } -}; - -export const WebAnonymousAuthentication: msRest.CompositeMapper = { - serializedName: "Anonymous", - type: { - name: "Composite", - polymorphicDiscriminator: WebLinkedServiceTypeProperties.type.polymorphicDiscriminator, - uberParent: "WebLinkedServiceTypeProperties", - className: "WebAnonymousAuthentication", - modelProperties: { - ...WebLinkedServiceTypeProperties.type.modelProperties - } - } -}; - -export const WebLinkedService: msRest.CompositeMapper = { - serializedName: "Web", - type: { - name: "Composite", - polymorphicDiscriminator: LinkedService.type.polymorphicDiscriminator, - uberParent: "LinkedService", - className: "WebLinkedService", - modelProperties: { - ...LinkedService.type.modelProperties, - typeProperties: { - required: true, - serializedName: "typeProperties", + }, + encryptedCredential: { + serializedName: "typeProperties.encryptedCredential", type: { - name: "Composite", - className: "WebLinkedServiceTypeProperties" + name: "Object" } } }, @@ -8044,43 +7688,32 @@ export const WebLinkedService: msRest.CompositeMapper = { } }; -export const ODataLinkedService: msRest.CompositeMapper = { - serializedName: "OData", +export const AzureMLServiceLinkedService: msRest.CompositeMapper = { + serializedName: "AzureMLService", type: { name: "Composite", polymorphicDiscriminator: LinkedService.type.polymorphicDiscriminator, uberParent: "LinkedService", - className: "ODataLinkedService", + className: "AzureMLServiceLinkedService", modelProperties: { ...LinkedService.type.modelProperties, - url: { + subscriptionId: { required: true, - serializedName: "typeProperties.url", + serializedName: "typeProperties.subscriptionId", type: { name: "Object" } }, - authenticationType: { - serializedName: "typeProperties.authenticationType", - type: { - name: "String" - } - }, - userName: { - serializedName: "typeProperties.userName", + resourceGroupName: { + required: true, + serializedName: "typeProperties.resourceGroupName", type: { name: "Object" } }, - password: { - serializedName: "typeProperties.password", - type: { - name: "Composite", - className: "SecretBase" - } - }, - tenant: { - serializedName: "typeProperties.tenant", + mlWorkspaceName: { + required: true, + serializedName: "typeProperties.mlWorkspaceName", type: { name: "Object" } @@ -8091,18 +7724,6 @@ export const ODataLinkedService: msRest.CompositeMapper = { name: "Object" } }, - aadResourceId: { - serializedName: "typeProperties.aadResourceId", - type: { - name: "Object" - } - }, - aadServicePrincipalCredentialType: { - serializedName: "typeProperties.aadServicePrincipalCredentialType", - type: { - name: "String" - } - }, servicePrincipalKey: { serializedName: "typeProperties.servicePrincipalKey", type: { @@ -8110,18 +7731,10 @@ export const ODataLinkedService: msRest.CompositeMapper = { className: "SecretBase" } }, - servicePrincipalEmbeddedCert: { - serializedName: "typeProperties.servicePrincipalEmbeddedCert", - type: { - name: "Composite", - className: "SecretBase" - } - }, - servicePrincipalEmbeddedCertPassword: { - serializedName: "typeProperties.servicePrincipalEmbeddedCertPassword", + tenant: { + serializedName: "typeProperties.tenant", type: { - name: "Composite", - className: "SecretBase" + name: "Object" } }, encryptedCredential: { @@ -8135,83 +7748,95 @@ export const ODataLinkedService: msRest.CompositeMapper = { } }; -export const HdfsLinkedService: msRest.CompositeMapper = { - serializedName: "Hdfs", +export const AzureMLLinkedService: msRest.CompositeMapper = { + serializedName: "AzureML", type: { name: "Composite", polymorphicDiscriminator: LinkedService.type.polymorphicDiscriminator, uberParent: "LinkedService", - className: "HdfsLinkedService", + className: "AzureMLLinkedService", modelProperties: { ...LinkedService.type.modelProperties, - url: { + mlEndpoint: { required: true, - serializedName: "typeProperties.url", + serializedName: "typeProperties.mlEndpoint", type: { name: "Object" } }, - authenticationType: { - serializedName: "typeProperties.authenticationType", + apiKey: { + required: true, + serializedName: "typeProperties.apiKey", type: { - name: "Object" + name: "Composite", + className: "SecretBase" } }, - encryptedCredential: { - serializedName: "typeProperties.encryptedCredential", + updateResourceEndpoint: { + serializedName: "typeProperties.updateResourceEndpoint", type: { name: "Object" } }, - userName: { - serializedName: "typeProperties.userName", + servicePrincipalId: { + serializedName: "typeProperties.servicePrincipalId", type: { name: "Object" } }, - password: { - serializedName: "typeProperties.password", + servicePrincipalKey: { + serializedName: "typeProperties.servicePrincipalKey", type: { name: "Composite", className: "SecretBase" } + }, + tenant: { + serializedName: "typeProperties.tenant", + type: { + name: "Object" + } + }, + encryptedCredential: { + serializedName: "typeProperties.encryptedCredential", + type: { + name: "Object" + } } }, additionalProperties: LinkedService.type.additionalProperties } }; -export const MicrosoftAccessLinkedService: msRest.CompositeMapper = { - serializedName: "MicrosoftAccess", +export const TeradataLinkedService: msRest.CompositeMapper = { + serializedName: "Teradata", type: { name: "Composite", polymorphicDiscriminator: LinkedService.type.polymorphicDiscriminator, uberParent: "LinkedService", - className: "MicrosoftAccessLinkedService", + className: "TeradataLinkedService", modelProperties: { ...LinkedService.type.modelProperties, connectionString: { - required: true, serializedName: "typeProperties.connectionString", type: { name: "Object" } }, - authenticationType: { - serializedName: "typeProperties.authenticationType", + server: { + serializedName: "typeProperties.server", type: { name: "Object" } }, - credential: { - serializedName: "typeProperties.credential", + authenticationType: { + serializedName: "typeProperties.authenticationType", type: { - name: "Composite", - className: "SecretBase" + name: "String" } }, - userName: { - serializedName: "typeProperties.userName", + username: { + serializedName: "typeProperties.username", type: { name: "Object" } @@ -8234,37 +7859,41 @@ export const MicrosoftAccessLinkedService: msRest.CompositeMapper = { } }; -export const InformixLinkedService: msRest.CompositeMapper = { - serializedName: "Informix", +export const Db2LinkedService: msRest.CompositeMapper = { + serializedName: "Db2", type: { name: "Composite", polymorphicDiscriminator: LinkedService.type.polymorphicDiscriminator, uberParent: "LinkedService", - className: "InformixLinkedService", + className: "Db2LinkedService", modelProperties: { ...LinkedService.type.modelProperties, connectionString: { - required: true, serializedName: "typeProperties.connectionString", type: { name: "Object" } }, - authenticationType: { - serializedName: "typeProperties.authenticationType", + server: { + serializedName: "typeProperties.server", type: { name: "Object" } }, - credential: { - serializedName: "typeProperties.credential", + database: { + serializedName: "typeProperties.database", type: { - name: "Composite", - className: "SecretBase" + name: "Object" } }, - userName: { - serializedName: "typeProperties.userName", + authenticationType: { + serializedName: "typeProperties.authenticationType", + type: { + name: "String" + } + }, + username: { + serializedName: "typeProperties.username", type: { name: "Object" } @@ -8276,6 +7905,18 @@ export const InformixLinkedService: msRest.CompositeMapper = { className: "SecretBase" } }, + packageCollection: { + serializedName: "typeProperties.packageCollection", + type: { + name: "Object" + } + }, + certificateCommonName: { + serializedName: "typeProperties.certificateCommonName", + type: { + name: "Object" + } + }, encryptedCredential: { serializedName: "typeProperties.encryptedCredential", type: { @@ -8287,37 +7928,43 @@ export const InformixLinkedService: msRest.CompositeMapper = { } }; -export const OdbcLinkedService: msRest.CompositeMapper = { - serializedName: "Odbc", +export const SybaseLinkedService: msRest.CompositeMapper = { + serializedName: "Sybase", type: { name: "Composite", polymorphicDiscriminator: LinkedService.type.polymorphicDiscriminator, uberParent: "LinkedService", - className: "OdbcLinkedService", + className: "SybaseLinkedService", modelProperties: { ...LinkedService.type.modelProperties, - connectionString: { + server: { required: true, - serializedName: "typeProperties.connectionString", + serializedName: "typeProperties.server", type: { name: "Object" } }, - authenticationType: { - serializedName: "typeProperties.authenticationType", + database: { + required: true, + serializedName: "typeProperties.database", type: { name: "Object" } }, - credential: { - serializedName: "typeProperties.credential", + schema: { + serializedName: "typeProperties.schema", type: { - name: "Composite", - className: "SecretBase" + name: "Object" } }, - userName: { - serializedName: "typeProperties.userName", + authenticationType: { + serializedName: "typeProperties.authenticationType", + type: { + name: "String" + } + }, + username: { + serializedName: "typeProperties.username", type: { name: "Object" } @@ -8340,53 +7987,27 @@ export const OdbcLinkedService: msRest.CompositeMapper = { } }; -export const AzureMLServiceLinkedService: msRest.CompositeMapper = { - serializedName: "AzureMLService", +export const PostgreSqlLinkedService: msRest.CompositeMapper = { + serializedName: "PostgreSql", type: { name: "Composite", polymorphicDiscriminator: LinkedService.type.polymorphicDiscriminator, uberParent: "LinkedService", - className: "AzureMLServiceLinkedService", + className: "PostgreSqlLinkedService", modelProperties: { ...LinkedService.type.modelProperties, - subscriptionId: { - required: true, - serializedName: "typeProperties.subscriptionId", - type: { - name: "Object" - } - }, - resourceGroupName: { - required: true, - serializedName: "typeProperties.resourceGroupName", - type: { - name: "Object" - } - }, - mlWorkspaceName: { + connectionString: { required: true, - serializedName: "typeProperties.mlWorkspaceName", - type: { - name: "Object" - } - }, - servicePrincipalId: { - serializedName: "typeProperties.servicePrincipalId", + serializedName: "typeProperties.connectionString", type: { name: "Object" } }, - servicePrincipalKey: { - serializedName: "typeProperties.servicePrincipalKey", + password: { + serializedName: "typeProperties.password", type: { name: "Composite", - className: "SecretBase" - } - }, - tenant: { - serializedName: "typeProperties.tenant", - type: { - name: "Object" + className: "AzureKeyVaultSecretReference" } }, encryptedCredential: { @@ -8400,53 +8021,61 @@ export const AzureMLServiceLinkedService: msRest.CompositeMapper = { } }; -export const AzureMLLinkedService: msRest.CompositeMapper = { - serializedName: "AzureML", +export const MySqlLinkedService: msRest.CompositeMapper = { + serializedName: "MySql", type: { name: "Composite", polymorphicDiscriminator: LinkedService.type.polymorphicDiscriminator, uberParent: "LinkedService", - className: "AzureMLLinkedService", + className: "MySqlLinkedService", modelProperties: { ...LinkedService.type.modelProperties, - mlEndpoint: { + connectionString: { required: true, - serializedName: "typeProperties.mlEndpoint", + serializedName: "typeProperties.connectionString", type: { name: "Object" } }, - apiKey: { - required: true, - serializedName: "typeProperties.apiKey", + password: { + serializedName: "typeProperties.password", type: { name: "Composite", - className: "SecretBase" + className: "AzureKeyVaultSecretReference" } }, - updateResourceEndpoint: { - serializedName: "typeProperties.updateResourceEndpoint", + encryptedCredential: { + serializedName: "typeProperties.encryptedCredential", type: { name: "Object" } - }, - servicePrincipalId: { - serializedName: "typeProperties.servicePrincipalId", + } + }, + additionalProperties: LinkedService.type.additionalProperties + } +}; + +export const AzureMySqlLinkedService: msRest.CompositeMapper = { + serializedName: "AzureMySql", + type: { + name: "Composite", + polymorphicDiscriminator: LinkedService.type.polymorphicDiscriminator, + uberParent: "LinkedService", + className: "AzureMySqlLinkedService", + modelProperties: { + ...LinkedService.type.modelProperties, + connectionString: { + required: true, + serializedName: "typeProperties.connectionString", type: { name: "Object" } }, - servicePrincipalKey: { - serializedName: "typeProperties.servicePrincipalKey", + password: { + serializedName: "typeProperties.password", type: { name: "Composite", - className: "SecretBase" - } - }, - tenant: { - serializedName: "typeProperties.tenant", - type: { - name: "Object" + className: "AzureKeyVaultSecretReference" } }, encryptedCredential: { @@ -8460,44 +8089,27 @@ export const AzureMLLinkedService: msRest.CompositeMapper = { } }; -export const TeradataLinkedService: msRest.CompositeMapper = { - serializedName: "Teradata", +export const OracleLinkedService: msRest.CompositeMapper = { + serializedName: "Oracle", type: { name: "Composite", polymorphicDiscriminator: LinkedService.type.polymorphicDiscriminator, uberParent: "LinkedService", - className: "TeradataLinkedService", + className: "OracleLinkedService", modelProperties: { ...LinkedService.type.modelProperties, connectionString: { + required: true, serializedName: "typeProperties.connectionString", type: { name: "Object" } }, - server: { - serializedName: "typeProperties.server", - type: { - name: "Object" - } - }, - authenticationType: { - serializedName: "typeProperties.authenticationType", - type: { - name: "String" - } - }, - username: { - serializedName: "typeProperties.username", - type: { - name: "Object" - } - }, password: { serializedName: "typeProperties.password", type: { name: "Composite", - className: "SecretBase" + className: "AzureKeyVaultSecretReference" } }, encryptedCredential: { @@ -8511,56 +8123,30 @@ export const TeradataLinkedService: msRest.CompositeMapper = { } }; -export const Db2LinkedService: msRest.CompositeMapper = { - serializedName: "Db2", +export const GoogleCloudStorageLinkedService: msRest.CompositeMapper = { + serializedName: "GoogleCloudStorage", type: { name: "Composite", polymorphicDiscriminator: LinkedService.type.polymorphicDiscriminator, uberParent: "LinkedService", - className: "Db2LinkedService", + className: "GoogleCloudStorageLinkedService", modelProperties: { ...LinkedService.type.modelProperties, - server: { - required: true, - serializedName: "typeProperties.server", - type: { - name: "Object" - } - }, - database: { - required: true, - serializedName: "typeProperties.database", - type: { - name: "Object" - } - }, - authenticationType: { - serializedName: "typeProperties.authenticationType", - type: { - name: "String" - } - }, - username: { - serializedName: "typeProperties.username", + accessKeyId: { + serializedName: "typeProperties.accessKeyId", type: { name: "Object" } }, - password: { - serializedName: "typeProperties.password", + secretAccessKey: { + serializedName: "typeProperties.secretAccessKey", type: { name: "Composite", className: "SecretBase" } }, - packageCollection: { - serializedName: "typeProperties.packageCollection", - type: { - name: "Object" - } - }, - certificateCommonName: { - serializedName: "typeProperties.certificateCommonName", + serviceUrl: { + serializedName: "typeProperties.serviceUrl", type: { name: "Object" } @@ -8576,252 +8162,18 @@ export const Db2LinkedService: msRest.CompositeMapper = { } }; -export const SybaseLinkedService: msRest.CompositeMapper = { - serializedName: "Sybase", +export const AzureFileStorageLinkedService: msRest.CompositeMapper = { + serializedName: "AzureFileStorage", type: { name: "Composite", polymorphicDiscriminator: LinkedService.type.polymorphicDiscriminator, uberParent: "LinkedService", - className: "SybaseLinkedService", + className: "AzureFileStorageLinkedService", modelProperties: { ...LinkedService.type.modelProperties, - server: { + host: { required: true, - serializedName: "typeProperties.server", - type: { - name: "Object" - } - }, - database: { - required: true, - serializedName: "typeProperties.database", - type: { - name: "Object" - } - }, - schema: { - serializedName: "typeProperties.schema", - type: { - name: "Object" - } - }, - authenticationType: { - serializedName: "typeProperties.authenticationType", - type: { - name: "String" - } - }, - username: { - serializedName: "typeProperties.username", - type: { - name: "Object" - } - }, - password: { - serializedName: "typeProperties.password", - type: { - name: "Composite", - className: "SecretBase" - } - }, - encryptedCredential: { - serializedName: "typeProperties.encryptedCredential", - type: { - name: "Object" - } - } - }, - additionalProperties: LinkedService.type.additionalProperties - } -}; - -export const PostgreSqlLinkedService: msRest.CompositeMapper = { - serializedName: "PostgreSql", - type: { - name: "Composite", - polymorphicDiscriminator: LinkedService.type.polymorphicDiscriminator, - uberParent: "LinkedService", - className: "PostgreSqlLinkedService", - modelProperties: { - ...LinkedService.type.modelProperties, - connectionString: { - required: true, - serializedName: "typeProperties.connectionString", - type: { - name: "Object" - } - }, - password: { - serializedName: "typeProperties.password", - type: { - name: "Composite", - className: "AzureKeyVaultSecretReference" - } - }, - encryptedCredential: { - serializedName: "typeProperties.encryptedCredential", - type: { - name: "Object" - } - } - }, - additionalProperties: LinkedService.type.additionalProperties - } -}; - -export const MySqlLinkedService: msRest.CompositeMapper = { - serializedName: "MySql", - type: { - name: "Composite", - polymorphicDiscriminator: LinkedService.type.polymorphicDiscriminator, - uberParent: "LinkedService", - className: "MySqlLinkedService", - modelProperties: { - ...LinkedService.type.modelProperties, - connectionString: { - required: true, - serializedName: "typeProperties.connectionString", - type: { - name: "Object" - } - }, - password: { - serializedName: "typeProperties.password", - type: { - name: "Composite", - className: "AzureKeyVaultSecretReference" - } - }, - encryptedCredential: { - serializedName: "typeProperties.encryptedCredential", - type: { - name: "Object" - } - } - }, - additionalProperties: LinkedService.type.additionalProperties - } -}; - -export const AzureMySqlLinkedService: msRest.CompositeMapper = { - serializedName: "AzureMySql", - type: { - name: "Composite", - polymorphicDiscriminator: LinkedService.type.polymorphicDiscriminator, - uberParent: "LinkedService", - className: "AzureMySqlLinkedService", - modelProperties: { - ...LinkedService.type.modelProperties, - connectionString: { - required: true, - serializedName: "typeProperties.connectionString", - type: { - name: "Object" - } - }, - password: { - serializedName: "typeProperties.password", - type: { - name: "Composite", - className: "AzureKeyVaultSecretReference" - } - }, - encryptedCredential: { - serializedName: "typeProperties.encryptedCredential", - type: { - name: "Object" - } - } - }, - additionalProperties: LinkedService.type.additionalProperties - } -}; - -export const OracleLinkedService: msRest.CompositeMapper = { - serializedName: "Oracle", - type: { - name: "Composite", - polymorphicDiscriminator: LinkedService.type.polymorphicDiscriminator, - uberParent: "LinkedService", - className: "OracleLinkedService", - modelProperties: { - ...LinkedService.type.modelProperties, - connectionString: { - required: true, - serializedName: "typeProperties.connectionString", - type: { - name: "Object" - } - }, - password: { - serializedName: "typeProperties.password", - type: { - name: "Composite", - className: "AzureKeyVaultSecretReference" - } - }, - encryptedCredential: { - serializedName: "typeProperties.encryptedCredential", - type: { - name: "Object" - } - } - }, - additionalProperties: LinkedService.type.additionalProperties - } -}; - -export const GoogleCloudStorageLinkedService: msRest.CompositeMapper = { - serializedName: "GoogleCloudStorage", - type: { - name: "Composite", - polymorphicDiscriminator: LinkedService.type.polymorphicDiscriminator, - uberParent: "LinkedService", - className: "GoogleCloudStorageLinkedService", - modelProperties: { - ...LinkedService.type.modelProperties, - accessKeyId: { - serializedName: "typeProperties.accessKeyId", - type: { - name: "Object" - } - }, - secretAccessKey: { - serializedName: "typeProperties.secretAccessKey", - type: { - name: "Composite", - className: "SecretBase" - } - }, - serviceUrl: { - serializedName: "typeProperties.serviceUrl", - type: { - name: "Object" - } - }, - encryptedCredential: { - serializedName: "typeProperties.encryptedCredential", - type: { - name: "Object" - } - } - }, - additionalProperties: LinkedService.type.additionalProperties - } -}; - -export const AzureFileStorageLinkedService: msRest.CompositeMapper = { - serializedName: "AzureFileStorage", - type: { - name: "Composite", - polymorphicDiscriminator: LinkedService.type.polymorphicDiscriminator, - uberParent: "LinkedService", - className: "AzureFileStorageLinkedService", - modelProperties: { - ...LinkedService.type.modelProperties, - host: { - required: true, - serializedName: "typeProperties.host", + serializedName: "typeProperties.host", type: { name: "Object" } @@ -12375,314 +11727,975 @@ export const BinaryDataset: msRest.CompositeMapper = { className: "DatasetCompression", additionalProperties: { type: { - name: "Object" + name: "Object" + } + } + } + } + }, + additionalProperties: Dataset.type.additionalProperties + } +}; + +export const OrcDataset: msRest.CompositeMapper = { + serializedName: "Orc", + type: { + name: "Composite", + polymorphicDiscriminator: Dataset.type.polymorphicDiscriminator, + uberParent: "Dataset", + className: "OrcDataset", + modelProperties: { + ...Dataset.type.modelProperties, + location: { + required: true, + serializedName: "typeProperties.location", + type: { + name: "Composite", + className: "DatasetLocation", + additionalProperties: { + type: { + name: "Object" + } + } + } + }, + orcCompressionCodec: { + serializedName: "typeProperties.orcCompressionCodec", + type: { + name: "String" + } + } + }, + additionalProperties: Dataset.type.additionalProperties + } +}; + +export const JsonDataset: msRest.CompositeMapper = { + serializedName: "Json", + type: { + name: "Composite", + polymorphicDiscriminator: Dataset.type.polymorphicDiscriminator, + uberParent: "Dataset", + className: "JsonDataset", + modelProperties: { + ...Dataset.type.modelProperties, + location: { + required: true, + serializedName: "typeProperties.location", + type: { + name: "Composite", + className: "DatasetLocation", + additionalProperties: { + type: { + name: "Object" + } + } + } + }, + encodingName: { + serializedName: "typeProperties.encodingName", + type: { + name: "Object" + } + }, + compression: { + serializedName: "typeProperties.compression", + type: { + name: "Composite", + className: "DatasetCompression", + additionalProperties: { + type: { + name: "Object" + } + } + } + } + }, + additionalProperties: Dataset.type.additionalProperties + } +}; + +export const DelimitedTextDataset: msRest.CompositeMapper = { + serializedName: "DelimitedText", + type: { + name: "Composite", + polymorphicDiscriminator: Dataset.type.polymorphicDiscriminator, + uberParent: "Dataset", + className: "DelimitedTextDataset", + modelProperties: { + ...Dataset.type.modelProperties, + location: { + required: true, + serializedName: "typeProperties.location", + type: { + name: "Composite", + className: "DatasetLocation", + additionalProperties: { + type: { + name: "Object" + } + } + } + }, + columnDelimiter: { + serializedName: "typeProperties.columnDelimiter", + type: { + name: "Object" + } + }, + rowDelimiter: { + serializedName: "typeProperties.rowDelimiter", + type: { + name: "Object" + } + }, + encodingName: { + serializedName: "typeProperties.encodingName", + type: { + name: "Object" + } + }, + compressionCodec: { + serializedName: "typeProperties.compressionCodec", + type: { + name: "Object" + } + }, + compressionLevel: { + serializedName: "typeProperties.compressionLevel", + type: { + name: "Object" + } + }, + quoteChar: { + serializedName: "typeProperties.quoteChar", + type: { + name: "Object" + } + }, + escapeChar: { + serializedName: "typeProperties.escapeChar", + type: { + name: "Object" + } + }, + firstRowAsHeader: { + serializedName: "typeProperties.firstRowAsHeader", + type: { + name: "Object" + } + }, + nullValue: { + serializedName: "typeProperties.nullValue", + type: { + name: "Object" + } + } + }, + additionalProperties: Dataset.type.additionalProperties + } +}; + +export const ParquetDataset: msRest.CompositeMapper = { + serializedName: "Parquet", + type: { + name: "Composite", + polymorphicDiscriminator: Dataset.type.polymorphicDiscriminator, + uberParent: "Dataset", + className: "ParquetDataset", + modelProperties: { + ...Dataset.type.modelProperties, + location: { + required: true, + serializedName: "typeProperties.location", + type: { + name: "Composite", + className: "DatasetLocation", + additionalProperties: { + type: { + name: "Object" + } + } + } + }, + compressionCodec: { + serializedName: "typeProperties.compressionCodec", + type: { + name: "Object" + } + } + }, + additionalProperties: Dataset.type.additionalProperties + } +}; + +export const AvroDataset: msRest.CompositeMapper = { + serializedName: "Avro", + type: { + name: "Composite", + polymorphicDiscriminator: Dataset.type.polymorphicDiscriminator, + uberParent: "Dataset", + className: "AvroDataset", + modelProperties: { + ...Dataset.type.modelProperties, + location: { + required: true, + serializedName: "typeProperties.location", + type: { + name: "Composite", + className: "DatasetLocation", + additionalProperties: { + type: { + name: "Object" + } + } + } + }, + avroCompressionCodec: { + serializedName: "typeProperties.avroCompressionCodec", + type: { + name: "String" + } + }, + avroCompressionLevel: { + serializedName: "typeProperties.avroCompressionLevel", + constraints: { + InclusiveMaximum: 9, + InclusiveMinimum: 1 + }, + type: { + name: "Number" + } + } + }, + additionalProperties: Dataset.type.additionalProperties + } +}; + +export const AmazonS3Dataset: msRest.CompositeMapper = { + serializedName: "AmazonS3Object", + type: { + name: "Composite", + polymorphicDiscriminator: Dataset.type.polymorphicDiscriminator, + uberParent: "Dataset", + className: "AmazonS3Dataset", + modelProperties: { + ...Dataset.type.modelProperties, + bucketName: { + required: true, + serializedName: "typeProperties.bucketName", + type: { + name: "Object" + } + }, + key: { + serializedName: "typeProperties.key", + type: { + name: "Object" + } + }, + prefix: { + serializedName: "typeProperties.prefix", + type: { + name: "Object" + } + }, + version: { + serializedName: "typeProperties.version", + type: { + name: "Object" + } + }, + modifiedDatetimeStart: { + serializedName: "typeProperties.modifiedDatetimeStart", + type: { + name: "Object" + } + }, + modifiedDatetimeEnd: { + serializedName: "typeProperties.modifiedDatetimeEnd", + type: { + name: "Object" + } + }, + format: { + serializedName: "typeProperties.format", + type: { + name: "Composite", + className: "DatasetStorageFormat", + additionalProperties: { + type: { + name: "Object" + } + } + } + }, + compression: { + serializedName: "typeProperties.compression", + type: { + name: "Composite", + className: "DatasetCompression", + additionalProperties: { + type: { + name: "Object" + } + } + } + } + }, + additionalProperties: Dataset.type.additionalProperties + } +}; + +export const ChainingTrigger: msRest.CompositeMapper = { + serializedName: "ChainingTrigger", + type: { + name: "Composite", + polymorphicDiscriminator: Trigger.type.polymorphicDiscriminator, + uberParent: "Trigger", + className: "ChainingTrigger", + modelProperties: { + ...Trigger.type.modelProperties, + pipelineProperty: { + required: true, + serializedName: "pipeline", + defaultValue: {}, + type: { + name: "Composite", + className: "TriggerPipelineReference" + } + }, + dependsOn: { + required: true, + serializedName: "typeProperties.dependsOn", + type: { + name: "Sequence", + element: { + type: { + name: "Composite", + className: "PipelineReference" + } + } + } + }, + runDimension: { + required: true, + serializedName: "typeProperties.runDimension", + type: { + name: "String" + } + } + }, + additionalProperties: Trigger.type.additionalProperties + } +}; + +export const RerunTumblingWindowTrigger: msRest.CompositeMapper = { + serializedName: "RerunTumblingWindowTrigger", + type: { + name: "Composite", + polymorphicDiscriminator: Trigger.type.polymorphicDiscriminator, + uberParent: "Trigger", + className: "RerunTumblingWindowTrigger", + modelProperties: { + ...Trigger.type.modelProperties, + parentTrigger: { + required: true, + serializedName: "typeProperties.parentTrigger", + type: { + name: "Object" + } + }, + requestedStartTime: { + required: true, + serializedName: "typeProperties.requestedStartTime", + type: { + name: "DateTime" + } + }, + requestedEndTime: { + required: true, + serializedName: "typeProperties.requestedEndTime", + type: { + name: "DateTime" + } + }, + rerunConcurrency: { + required: true, + serializedName: "typeProperties.rerunConcurrency", + constraints: { + InclusiveMaximum: 50, + InclusiveMinimum: 1 + }, + type: { + name: "Number" + } + } + }, + additionalProperties: Trigger.type.additionalProperties + } +}; + +export const DependencyReference: msRest.CompositeMapper = { + serializedName: "DependencyReference", + type: { + name: "Composite", + polymorphicDiscriminator: { + serializedName: "type", + clientName: "type" + }, + uberParent: "DependencyReference", + className: "DependencyReference", + modelProperties: { + type: { + required: true, + serializedName: "type", + type: { + name: "String" + } + } + } + } +}; + +export const SelfDependencyTumblingWindowTriggerReference: msRest.CompositeMapper = { + serializedName: "SelfDependencyTumblingWindowTriggerReference", + type: { + name: "Composite", + polymorphicDiscriminator: DependencyReference.type.polymorphicDiscriminator, + uberParent: "DependencyReference", + className: "SelfDependencyTumblingWindowTriggerReference", + modelProperties: { + ...DependencyReference.type.modelProperties, + offset: { + required: true, + serializedName: "offset", + constraints: { + MaxLength: 15, + MinLength: 8, + Pattern: /-((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9]))/ + }, + type: { + name: "String" + } + }, + size: { + serializedName: "size", + constraints: { + MaxLength: 15, + MinLength: 8, + Pattern: /((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9]))/ + }, + type: { + name: "String" + } + } + } + } +}; + +export const TriggerReference: msRest.CompositeMapper = { + serializedName: "TriggerReference", + type: { + name: "Composite", + className: "TriggerReference", + modelProperties: { + type: { + required: true, + isConstant: true, + serializedName: "type", + defaultValue: 'TriggerReference', + type: { + name: "String" + } + }, + referenceName: { + required: true, + serializedName: "referenceName", + type: { + name: "String" + } + } + } + } +}; + +export const TriggerDependencyReference: msRest.CompositeMapper = { + serializedName: "TriggerDependencyReference", + type: { + name: "Composite", + polymorphicDiscriminator: DependencyReference.type.polymorphicDiscriminator, + uberParent: "DependencyReference", + className: "TriggerDependencyReference", + modelProperties: { + ...DependencyReference.type.modelProperties, + referenceTrigger: { + required: true, + serializedName: "referenceTrigger", + defaultValue: {}, + type: { + name: "Composite", + className: "TriggerReference" + } + } + } + } +}; + +export const TumblingWindowTriggerDependencyReference: msRest.CompositeMapper = { + serializedName: "TumblingWindowTriggerDependencyReference", + type: { + name: "Composite", + polymorphicDiscriminator: DependencyReference.type.polymorphicDiscriminator, + uberParent: "DependencyReference", + className: "TumblingWindowTriggerDependencyReference", + modelProperties: { + ...TriggerDependencyReference.type.modelProperties, + offset: { + serializedName: "offset", + constraints: { + MaxLength: 15, + MinLength: 8, + Pattern: /-?((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9]))/ + }, + type: { + name: "String" + } + }, + size: { + serializedName: "size", + constraints: { + MaxLength: 15, + MinLength: 8, + Pattern: /((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9]))/ + }, + type: { + name: "String" + } + } + } + } +}; + +export const RetryPolicy: msRest.CompositeMapper = { + serializedName: "RetryPolicy", + type: { + name: "Composite", + className: "RetryPolicy", + modelProperties: { + count: { + serializedName: "count", + type: { + name: "Object" + } + }, + intervalInSeconds: { + serializedName: "intervalInSeconds", + constraints: { + InclusiveMaximum: 86400, + InclusiveMinimum: 30 + }, + type: { + name: "Number" + } + } + } + } +}; + +export const TumblingWindowTrigger: msRest.CompositeMapper = { + serializedName: "TumblingWindowTrigger", + type: { + name: "Composite", + polymorphicDiscriminator: Trigger.type.polymorphicDiscriminator, + uberParent: "Trigger", + className: "TumblingWindowTrigger", + modelProperties: { + ...Trigger.type.modelProperties, + pipelineProperty: { + required: true, + serializedName: "pipeline", + defaultValue: {}, + type: { + name: "Composite", + className: "TriggerPipelineReference" + } + }, + frequency: { + required: true, + serializedName: "typeProperties.frequency", + type: { + name: "String" + } + }, + interval: { + required: true, + serializedName: "typeProperties.interval", + type: { + name: "Number" + } + }, + startTime: { + required: true, + serializedName: "typeProperties.startTime", + type: { + name: "DateTime" + } + }, + endTime: { + serializedName: "typeProperties.endTime", + type: { + name: "DateTime" + } + }, + delay: { + serializedName: "typeProperties.delay", + type: { + name: "Object" + } + }, + maxConcurrency: { + required: true, + serializedName: "typeProperties.maxConcurrency", + constraints: { + InclusiveMaximum: 50, + InclusiveMinimum: 1 + }, + type: { + name: "Number" + } + }, + retryPolicy: { + serializedName: "typeProperties.retryPolicy", + type: { + name: "Composite", + className: "RetryPolicy" + } + }, + dependsOn: { + serializedName: "typeProperties.dependsOn", + type: { + name: "Sequence", + element: { + type: { + name: "Composite", + className: "DependencyReference" } } } } }, - additionalProperties: Dataset.type.additionalProperties + additionalProperties: Trigger.type.additionalProperties } }; -export const OrcDataset: msRest.CompositeMapper = { - serializedName: "Orc", +export const MultiplePipelineTrigger: msRest.CompositeMapper = { + serializedName: "MultiplePipelineTrigger", type: { name: "Composite", - polymorphicDiscriminator: Dataset.type.polymorphicDiscriminator, - uberParent: "Dataset", - className: "OrcDataset", + polymorphicDiscriminator: Trigger.type.polymorphicDiscriminator, + uberParent: "Trigger", + className: "MultiplePipelineTrigger", modelProperties: { - ...Dataset.type.modelProperties, - location: { - required: true, - serializedName: "typeProperties.location", + ...Trigger.type.modelProperties, + pipelines: { + serializedName: "pipelines", type: { - name: "Composite", - className: "DatasetLocation", - additionalProperties: { + name: "Sequence", + element: { type: { - name: "Object" + name: "Composite", + className: "TriggerPipelineReference" } } } - }, - orcCompressionCodec: { - serializedName: "typeProperties.orcCompressionCodec", - type: { - name: "String" - } } }, - additionalProperties: Dataset.type.additionalProperties + additionalProperties: Trigger.type.additionalProperties } }; -export const JsonDataset: msRest.CompositeMapper = { - serializedName: "Json", +export const BlobEventsTrigger: msRest.CompositeMapper = { + serializedName: "BlobEventsTrigger", type: { name: "Composite", - polymorphicDiscriminator: Dataset.type.polymorphicDiscriminator, - uberParent: "Dataset", - className: "JsonDataset", + polymorphicDiscriminator: Trigger.type.polymorphicDiscriminator, + uberParent: "Trigger", + className: "BlobEventsTrigger", modelProperties: { - ...Dataset.type.modelProperties, - location: { - required: true, - serializedName: "typeProperties.location", + ...MultiplePipelineTrigger.type.modelProperties, + blobPathBeginsWith: { + serializedName: "typeProperties.blobPathBeginsWith", type: { - name: "Composite", - className: "DatasetLocation", - additionalProperties: { - type: { - name: "Object" - } - } + name: "String" } }, - encodingName: { - serializedName: "typeProperties.encodingName", + blobPathEndsWith: { + serializedName: "typeProperties.blobPathEndsWith", type: { - name: "Object" + name: "String" } }, - compression: { - serializedName: "typeProperties.compression", + ignoreEmptyBlobs: { + serializedName: "typeProperties.ignoreEmptyBlobs", type: { - name: "Composite", - className: "DatasetCompression", - additionalProperties: { + name: "Boolean" + } + }, + events: { + required: true, + serializedName: "typeProperties.events", + type: { + name: "Sequence", + element: { type: { - name: "Object" + name: "String" } } } + }, + scope: { + required: true, + serializedName: "typeProperties.scope", + type: { + name: "String" + } } }, - additionalProperties: Dataset.type.additionalProperties + additionalProperties: Trigger.type.additionalProperties } }; -export const DelimitedTextDataset: msRest.CompositeMapper = { - serializedName: "DelimitedText", +export const BlobTrigger: msRest.CompositeMapper = { + serializedName: "BlobTrigger", type: { name: "Composite", - polymorphicDiscriminator: Dataset.type.polymorphicDiscriminator, - uberParent: "Dataset", - className: "DelimitedTextDataset", + polymorphicDiscriminator: Trigger.type.polymorphicDiscriminator, + uberParent: "Trigger", + className: "BlobTrigger", modelProperties: { - ...Dataset.type.modelProperties, - location: { + ...MultiplePipelineTrigger.type.modelProperties, + folderPath: { required: true, - serializedName: "typeProperties.location", - type: { - name: "Composite", - className: "DatasetLocation", - additionalProperties: { - type: { - name: "Object" - } - } - } - }, - columnDelimiter: { - serializedName: "typeProperties.columnDelimiter", - type: { - name: "Object" - } - }, - rowDelimiter: { - serializedName: "typeProperties.rowDelimiter", - type: { - name: "Object" - } - }, - encodingName: { - serializedName: "typeProperties.encodingName", - type: { - name: "Object" - } - }, - compressionCodec: { - serializedName: "typeProperties.compressionCodec", - type: { - name: "Object" - } - }, - compressionLevel: { - serializedName: "typeProperties.compressionLevel", + serializedName: "typeProperties.folderPath", type: { - name: "Object" + name: "String" } }, - quoteChar: { - serializedName: "typeProperties.quoteChar", + maxConcurrency: { + required: true, + serializedName: "typeProperties.maxConcurrency", type: { - name: "Object" + name: "Number" } }, - escapeChar: { - serializedName: "typeProperties.escapeChar", + linkedService: { + required: true, + serializedName: "typeProperties.linkedService", + defaultValue: {}, type: { - name: "Object" + name: "Composite", + className: "LinkedServiceReference" } - }, - firstRowAsHeader: { - serializedName: "typeProperties.firstRowAsHeader", + } + }, + additionalProperties: Trigger.type.additionalProperties + } +}; + +export const RecurrenceScheduleOccurrence: msRest.CompositeMapper = { + serializedName: "RecurrenceScheduleOccurrence", + type: { + name: "Composite", + className: "RecurrenceScheduleOccurrence", + modelProperties: { + day: { + serializedName: "day", type: { - name: "Object" + name: "Enum", + allowedValues: [ + "Sunday", + "Monday", + "Tuesday", + "Wednesday", + "Thursday", + "Friday", + "Saturday" + ] } }, - nullValue: { - serializedName: "typeProperties.nullValue", + occurrence: { + serializedName: "occurrence", type: { - name: "Object" + name: "Number" } } }, - additionalProperties: Dataset.type.additionalProperties + additionalProperties: { + type: { + name: "Object" + } + } } }; -export const ParquetDataset: msRest.CompositeMapper = { - serializedName: "Parquet", +export const RecurrenceSchedule: msRest.CompositeMapper = { + serializedName: "RecurrenceSchedule", type: { name: "Composite", - polymorphicDiscriminator: Dataset.type.polymorphicDiscriminator, - uberParent: "Dataset", - className: "ParquetDataset", + className: "RecurrenceSchedule", modelProperties: { - ...Dataset.type.modelProperties, - location: { - required: true, - serializedName: "typeProperties.location", + minutes: { + serializedName: "minutes", type: { - name: "Composite", - className: "DatasetLocation", - additionalProperties: { + name: "Sequence", + element: { type: { - name: "Object" + name: "Number" + } + } + } + }, + hours: { + serializedName: "hours", + type: { + name: "Sequence", + element: { + type: { + name: "Number" + } + } + } + }, + weekDays: { + serializedName: "weekDays", + type: { + name: "Sequence", + element: { + type: { + name: "Enum", + allowedValues: [ + "Sunday", + "Monday", + "Tuesday", + "Wednesday", + "Thursday", + "Friday", + "Saturday" + ] } } } }, - compressionCodec: { - serializedName: "typeProperties.compressionCodec", - type: { - name: "Object" - } - } - }, - additionalProperties: Dataset.type.additionalProperties - } -}; - -export const AvroDataset: msRest.CompositeMapper = { - serializedName: "Avro", - type: { - name: "Composite", - polymorphicDiscriminator: Dataset.type.polymorphicDiscriminator, - uberParent: "Dataset", - className: "AvroDataset", - modelProperties: { - ...Dataset.type.modelProperties, - location: { - required: true, - serializedName: "typeProperties.location", + monthDays: { + serializedName: "monthDays", type: { - name: "Composite", - className: "DatasetLocation", - additionalProperties: { + name: "Sequence", + element: { type: { - name: "Object" + name: "Number" } } } }, - avroCompressionCodec: { - serializedName: "typeProperties.avroCompressionCodec", - type: { - name: "String" - } - }, - avroCompressionLevel: { - serializedName: "typeProperties.avroCompressionLevel", - constraints: { - InclusiveMaximum: 9, - InclusiveMinimum: 1 - }, + monthlyOccurrences: { + serializedName: "monthlyOccurrences", type: { - name: "Number" + name: "Sequence", + element: { + type: { + name: "Composite", + className: "RecurrenceScheduleOccurrence", + additionalProperties: { + type: { + name: "Object" + } + } + } + } } } }, - additionalProperties: Dataset.type.additionalProperties + additionalProperties: { + type: { + name: "Object" + } + } } }; -export const AmazonS3Dataset: msRest.CompositeMapper = { - serializedName: "AmazonS3Object", +export const ScheduleTriggerRecurrence: msRest.CompositeMapper = { + serializedName: "ScheduleTriggerRecurrence", type: { name: "Composite", - polymorphicDiscriminator: Dataset.type.polymorphicDiscriminator, - uberParent: "Dataset", - className: "AmazonS3Dataset", + className: "ScheduleTriggerRecurrence", modelProperties: { - ...Dataset.type.modelProperties, - bucketName: { - required: true, - serializedName: "typeProperties.bucketName", - type: { - name: "Object" - } - }, - key: { - serializedName: "typeProperties.key", + frequency: { + serializedName: "frequency", type: { - name: "Object" + name: "String" } }, - prefix: { - serializedName: "typeProperties.prefix", + interval: { + serializedName: "interval", type: { - name: "Object" + name: "Number" } }, - version: { - serializedName: "typeProperties.version", + startTime: { + serializedName: "startTime", type: { - name: "Object" + name: "DateTime" } }, - modifiedDatetimeStart: { - serializedName: "typeProperties.modifiedDatetimeStart", + endTime: { + serializedName: "endTime", type: { - name: "Object" + name: "DateTime" } }, - modifiedDatetimeEnd: { - serializedName: "typeProperties.modifiedDatetimeEnd", + timeZone: { + serializedName: "timeZone", type: { - name: "Object" + name: "String" } }, - format: { - serializedName: "typeProperties.format", + schedule: { + serializedName: "schedule", type: { name: "Composite", - className: "DatasetStorageFormat", + className: "RecurrenceSchedule", additionalProperties: { type: { name: "Object" } } } - }, - compression: { - serializedName: "typeProperties.compression", + } + }, + additionalProperties: { + type: { + name: "Object" + } + } + } +}; + +export const ScheduleTrigger: msRest.CompositeMapper = { + serializedName: "ScheduleTrigger", + type: { + name: "Composite", + polymorphicDiscriminator: Trigger.type.polymorphicDiscriminator, + uberParent: "Trigger", + className: "ScheduleTrigger", + modelProperties: { + ...MultiplePipelineTrigger.type.modelProperties, + recurrence: { + required: true, + serializedName: "typeProperties.recurrence", type: { name: "Composite", - className: "DatasetCompression", + className: "ScheduleTriggerRecurrence", additionalProperties: { type: { name: "Object" @@ -12691,7 +12704,7 @@ export const AmazonS3Dataset: msRest.CompositeMapper = { } } }, - additionalProperties: Dataset.type.additionalProperties + additionalProperties: Trigger.type.additionalProperties } }; @@ -13576,6 +13589,28 @@ export const Office365Source: msRest.CompositeMapper = { } }; +export const AdditionalColumns: msRest.CompositeMapper = { + serializedName: "AdditionalColumns", + type: { + name: "Composite", + className: "AdditionalColumns", + modelProperties: { + name: { + serializedName: "name", + type: { + name: "Object" + } + }, + value: { + serializedName: "value", + type: { + name: "Object" + } + } + } + } +}; + export const MongoDbCursorMethodsProperties: msRest.CompositeMapper = { serializedName: "MongoDbCursorMethodsProperties", type: { @@ -13653,6 +13688,18 @@ export const CosmosDbMongoDbApiSource: msRest.CompositeMapper = { type: { name: "Object" } + }, + additionalColumns: { + serializedName: "additionalColumns", + type: { + name: "Sequence", + element: { + type: { + name: "Composite", + className: "AdditionalColumns" + } + } + } } }, additionalProperties: CopySource.type.additionalProperties @@ -13697,6 +13744,18 @@ export const MongoDbV2Source: msRest.CompositeMapper = { type: { name: "Object" } + }, + additionalColumns: { + serializedName: "additionalColumns", + type: { + name: "Sequence", + element: { + type: { + name: "Composite", + className: "AdditionalColumns" + } + } + } } }, additionalProperties: CopySource.type.additionalProperties @@ -13717,6 +13776,18 @@ export const MongoDbSource: msRest.CompositeMapper = { type: { name: "Object" } + }, + additionalColumns: { + serializedName: "additionalColumns", + type: { + name: "Sequence", + element: { + type: { + name: "Composite", + className: "AdditionalColumns" + } + } + } } }, additionalProperties: CopySource.type.additionalProperties @@ -13731,7 +13802,19 @@ export const WebSource: msRest.CompositeMapper = { uberParent: "CopySource", className: "WebSource", modelProperties: { - ...CopySource.type.modelProperties + ...CopySource.type.modelProperties, + additionalColumns: { + serializedName: "additionalColumns", + type: { + name: "Sequence", + element: { + type: { + name: "Composite", + className: "AdditionalColumns" + } + } + } + } }, additionalProperties: CopySource.type.additionalProperties } @@ -13804,6 +13887,18 @@ export const OracleSource: msRest.CompositeMapper = { name: "Composite", className: "OraclePartitionSettings" } + }, + additionalColumns: { + serializedName: "additionalColumns", + type: { + name: "Sequence", + element: { + type: { + name: "Composite", + className: "AdditionalColumns" + } + } + } } }, additionalProperties: CopySource.type.additionalProperties @@ -13837,6 +13932,18 @@ export const AzureDataExplorerSource: msRest.CompositeMapper = { type: { name: "Object" } + }, + additionalColumns: { + serializedName: "additionalColumns", + type: { + name: "Sequence", + element: { + type: { + name: "Composite", + className: "AdditionalColumns" + } + } + } } }, additionalProperties: CopySource.type.additionalProperties @@ -13914,6 +14021,18 @@ export const FileSystemSource: msRest.CompositeMapper = { type: { name: "Object" } + }, + additionalColumns: { + serializedName: "additionalColumns", + type: { + name: "Sequence", + element: { + type: { + name: "Composite", + className: "AdditionalColumns" + } + } + } } }, additionalProperties: CopySource.type.additionalProperties @@ -13964,6 +14083,18 @@ export const RestSource: msRest.CompositeMapper = { type: { name: "Object" } + }, + additionalColumns: { + serializedName: "additionalColumns", + type: { + name: "Sequence", + element: { + type: { + name: "Composite", + className: "AdditionalColumns" + } + } + } } }, additionalProperties: CopySource.type.additionalProperties @@ -13990,6 +14121,18 @@ export const SalesforceServiceCloudSource: msRest.CompositeMapper = { type: { name: "String" } + }, + additionalColumns: { + serializedName: "additionalColumns", + type: { + name: "Sequence", + element: { + type: { + name: "Composite", + className: "AdditionalColumns" + } + } + } } }, additionalProperties: CopySource.type.additionalProperties @@ -14010,6 +14153,18 @@ export const ODataSource: msRest.CompositeMapper = { type: { name: "Object" } + }, + additionalColumns: { + serializedName: "additionalColumns", + type: { + name: "Sequence", + element: { + type: { + name: "Composite", + className: "AdditionalColumns" + } + } + } } }, additionalProperties: CopySource.type.additionalProperties @@ -14030,6 +14185,18 @@ export const MicrosoftAccessSource: msRest.CompositeMapper = { type: { name: "Object" } + }, + additionalColumns: { + serializedName: "additionalColumns", + type: { + name: "Sequence", + element: { + type: { + name: "Composite", + className: "AdditionalColumns" + } + } + } } }, additionalProperties: CopySource.type.additionalProperties @@ -14050,6 +14217,18 @@ export const RelationalSource: msRest.CompositeMapper = { type: { name: "Object" } + }, + additionalColumns: { + serializedName: "additionalColumns", + type: { + name: "Sequence", + element: { + type: { + name: "Composite", + className: "AdditionalColumns" + } + } + } } }, additionalProperties: CopySource.type.additionalProperties @@ -14070,6 +14249,18 @@ export const CommonDataServiceForAppsSource: msRest.CompositeMapper = { type: { name: "Object" } + }, + additionalColumns: { + serializedName: "additionalColumns", + type: { + name: "Sequence", + element: { + type: { + name: "Composite", + className: "AdditionalColumns" + } + } + } } }, additionalProperties: CopySource.type.additionalProperties @@ -14090,6 +14281,18 @@ export const DynamicsCrmSource: msRest.CompositeMapper = { type: { name: "Object" } + }, + additionalColumns: { + serializedName: "additionalColumns", + type: { + name: "Sequence", + element: { + type: { + name: "Composite", + className: "AdditionalColumns" + } + } + } } }, additionalProperties: CopySource.type.additionalProperties @@ -14110,6 +14313,18 @@ export const DynamicsSource: msRest.CompositeMapper = { type: { name: "Object" } + }, + additionalColumns: { + serializedName: "additionalColumns", + type: { + name: "Sequence", + element: { + type: { + name: "Composite", + className: "AdditionalColumns" + } + } + } } }, additionalProperties: CopySource.type.additionalProperties @@ -14142,6 +14357,18 @@ export const CosmosDbSqlApiSource: msRest.CompositeMapper = { type: { name: "Object" } + }, + additionalColumns: { + serializedName: "additionalColumns", + type: { + name: "Sequence", + element: { + type: { + name: "Composite", + className: "AdditionalColumns" + } + } + } } }, additionalProperties: CopySource.type.additionalProperties @@ -14174,6 +14401,18 @@ export const DocumentDbCollectionSource: msRest.CompositeMapper = { type: { name: "Object" } + }, + additionalColumns: { + serializedName: "additionalColumns", + type: { + name: "Sequence", + element: { + type: { + name: "Composite", + className: "AdditionalColumns" + } + } + } } }, additionalProperties: CopySource.type.additionalProperties @@ -14252,6 +14491,18 @@ export const TabularSource: msRest.CompositeMapper = { type: { name: "Object" } + }, + additionalColumns: { + serializedName: "additionalColumns", + type: { + name: "Sequence", + element: { + type: { + name: "Composite", + className: "AdditionalColumns" + } + } + } } }, additionalProperties: CopySource.type.additionalProperties @@ -15331,6 +15582,12 @@ export const SqlSource: msRest.CompositeMapper = { } } } + }, + isolationLevel: { + serializedName: "isolationLevel", + type: { + name: "Object" + } } }, additionalProperties: CopySource.type.additionalProperties @@ -15807,6 +16064,12 @@ export const HdfsReadSettings: msRest.CompositeMapper = { name: "Object" } }, + fileListPath: { + serializedName: "fileListPath", + type: { + name: "Object" + } + }, enablePartitionDiscovery: { serializedName: "enablePartitionDiscovery", type: { @@ -15902,6 +16165,12 @@ export const SftpReadSettings: msRest.CompositeMapper = { name: "Object" } }, + fileListPath: { + serializedName: "fileListPath", + type: { + name: "Object" + } + }, modifiedDatetimeStart: { serializedName: "modifiedDatetimeStart", type: { @@ -15946,6 +16215,12 @@ export const FtpReadSettings: msRest.CompositeMapper = { name: "Object" } }, + fileListPath: { + serializedName: "fileListPath", + type: { + name: "Object" + } + }, useBinaryTransfer: { serializedName: "useBinaryTransfer", type: { @@ -15990,6 +16265,12 @@ export const GoogleCloudStorageReadSettings: msRest.CompositeMapper = { name: "Object" } }, + fileListPath: { + serializedName: "fileListPath", + type: { + name: "Object" + } + }, enablePartitionDiscovery: { serializedName: "enablePartitionDiscovery", type: { @@ -16040,6 +16321,12 @@ export const AzureFileStorageReadSettings: msRest.CompositeMapper = { name: "Object" } }, + fileListPath: { + serializedName: "fileListPath", + type: { + name: "Object" + } + }, enablePartitionDiscovery: { serializedName: "enablePartitionDiscovery", type: { @@ -16090,6 +16377,12 @@ export const FileServerReadSettings: msRest.CompositeMapper = { name: "Object" } }, + fileListPath: { + serializedName: "fileListPath", + type: { + name: "Object" + } + }, enablePartitionDiscovery: { serializedName: "enablePartitionDiscovery", type: { @@ -16146,6 +16439,12 @@ export const AmazonS3ReadSettings: msRest.CompositeMapper = { name: "Object" } }, + fileListPath: { + serializedName: "fileListPath", + type: { + name: "Object" + } + }, enablePartitionDiscovery: { serializedName: "enablePartitionDiscovery", type: { @@ -16196,6 +16495,12 @@ export const AzureDataLakeStoreReadSettings: msRest.CompositeMapper = { name: "Object" } }, + fileListPath: { + serializedName: "fileListPath", + type: { + name: "Object" + } + }, enablePartitionDiscovery: { serializedName: "enablePartitionDiscovery", type: { @@ -16246,6 +16551,12 @@ export const AzureBlobFSReadSettings: msRest.CompositeMapper = { name: "Object" } }, + fileListPath: { + serializedName: "fileListPath", + type: { + name: "Object" + } + }, enablePartitionDiscovery: { serializedName: "enablePartitionDiscovery", type: { @@ -16302,6 +16613,12 @@ export const AzureBlobStorageReadSettings: msRest.CompositeMapper = { name: "Object" } }, + fileListPath: { + serializedName: "fileListPath", + type: { + name: "Object" + } + }, enablePartitionDiscovery: { serializedName: "enablePartitionDiscovery", type: { @@ -16371,6 +16688,18 @@ export const OrcSource: msRest.CompositeMapper = { } } } + }, + additionalColumns: { + serializedName: "additionalColumns", + type: { + name: "Sequence", + element: { + type: { + name: "Composite", + className: "AdditionalColumns" + } + } + } } }, additionalProperties: CopySource.type.additionalProperties @@ -16397,6 +16726,18 @@ export const JsonSource: msRest.CompositeMapper = { } } } + }, + additionalColumns: { + serializedName: "additionalColumns", + type: { + name: "Sequence", + element: { + type: { + name: "Composite", + className: "AdditionalColumns" + } + } + } } }, additionalProperties: CopySource.type.additionalProperties @@ -16478,6 +16819,18 @@ export const DelimitedTextSource: msRest.CompositeMapper = { className: "DelimitedTextReadSettings", additionalProperties: FormatReadSettings.type.additionalProperties } + }, + additionalColumns: { + serializedName: "additionalColumns", + type: { + name: "Sequence", + element: { + type: { + name: "Composite", + className: "AdditionalColumns" + } + } + } } }, additionalProperties: CopySource.type.additionalProperties @@ -16504,6 +16857,18 @@ export const ParquetSource: msRest.CompositeMapper = { } } } + }, + additionalColumns: { + serializedName: "additionalColumns", + type: { + name: "Sequence", + element: { + type: { + name: "Composite", + className: "AdditionalColumns" + } + } + } } }, additionalProperties: CopySource.type.additionalProperties @@ -16530,6 +16895,18 @@ export const AvroSource: msRest.CompositeMapper = { } } } + }, + additionalColumns: { + serializedName: "additionalColumns", + type: { + name: "Sequence", + element: { + type: { + name: "Composite", + className: "AdditionalColumns" + } + } + } } }, additionalProperties: CopySource.type.additionalProperties @@ -17625,6 +18002,28 @@ export const HDInsightHiveActivity: msRest.CompositeMapper = { } }; +export const SkipErrorFile: msRest.CompositeMapper = { + serializedName: "SkipErrorFile", + type: { + name: "Composite", + className: "SkipErrorFile", + modelProperties: { + fileMissing: { + serializedName: "fileMissing", + type: { + name: "Object" + } + }, + dataInconsistency: { + serializedName: "dataInconsistency", + type: { + name: "Object" + } + } + } + } +}; + export const RedirectIncompatibleRowSettings: msRest.CompositeMapper = { serializedName: "RedirectIncompatibleRowSettings", type: { @@ -18653,7 +19052,13 @@ export const AzureDataLakeStoreWriteSettings: msRest.CompositeMapper = { uberParent: "StoreWriteSettings", className: "AzureDataLakeStoreWriteSettings", modelProperties: { - ...StoreWriteSettings.type.modelProperties + ...StoreWriteSettings.type.modelProperties, + expiryDateTime: { + serializedName: "expiryDateTime", + type: { + name: "Object" + } + } }, additionalProperties: StoreWriteSettings.type.additionalProperties } @@ -18713,6 +19118,12 @@ export const SftpWriteSettings: msRest.CompositeMapper = { type: { name: "Object" } + }, + useTempFileRename: { + serializedName: "useTempFileRename", + type: { + name: "Object" + } } }, additionalProperties: StoreWriteSettings.type.additionalProperties @@ -19200,6 +19611,18 @@ export const CopyActivity: msRest.CompositeMapper = { } } }, + logStorageSettings: { + serializedName: "typeProperties.logStorageSettings", + type: { + name: "Composite", + className: "LogStorageSettings", + additionalProperties: { + type: { + name: "Object" + } + } + } + }, preserveRules: { serializedName: "typeProperties.preserveRules", type: { @@ -19222,6 +19645,19 @@ export const CopyActivity: msRest.CompositeMapper = { } } }, + validateDataConsistency: { + serializedName: "typeProperties.validateDataConsistency", + type: { + name: "Object" + } + }, + skipErrorFile: { + serializedName: "typeProperties.skipErrorFile", + type: { + name: "Composite", + className: "SkipErrorFile" + } + }, inputs: { serializedName: "inputs", type: { @@ -21730,36 +22166,6 @@ export const TriggerListResponse: msRest.CompositeMapper = { } }; -export const RerunTriggerListResponse: msRest.CompositeMapper = { - serializedName: "RerunTriggerListResponse", - type: { - name: "Composite", - className: "RerunTriggerListResponse", - modelProperties: { - value: { - required: true, - serializedName: "", - type: { - name: "Sequence", - element: { - type: { - name: "Composite", - className: "RerunTriggerResource" - } - } - } - }, - nextLink: { - readOnly: true, - serializedName: "nextLink", - type: { - name: "String" - } - } - } - } -}; - export const DataFlowListResponse: msRest.CompositeMapper = { serializedName: "DataFlowListResponse", type: { @@ -21835,19 +22241,8 @@ export const discriminators = { 'Trigger' : Trigger, 'FactoryRepoConfiguration.FactoryVSTSConfiguration' : FactoryVSTSConfiguration, 'FactoryRepoConfiguration.FactoryGitHubConfiguration' : FactoryGitHubConfiguration, - 'Trigger.RerunTumblingWindowTrigger' : RerunTumblingWindowTrigger, 'DataFlow' : DataFlow, 'DataFlow.MappingDataFlow' : MappingDataFlow, - 'Trigger.ChainingTrigger' : ChainingTrigger, - 'DependencyReference.SelfDependencyTumblingWindowTriggerReference' : SelfDependencyTumblingWindowTriggerReference, - 'DependencyReference.TumblingWindowTriggerDependencyReference' : TumblingWindowTriggerDependencyReference, - 'DependencyReference.TriggerDependencyReference' : TriggerDependencyReference, - 'DependencyReference' : DependencyReference, - 'Trigger.TumblingWindowTrigger' : TumblingWindowTrigger, - 'Trigger.BlobEventsTrigger' : BlobEventsTrigger, - 'Trigger.BlobTrigger' : BlobTrigger, - 'Trigger.ScheduleTrigger' : ScheduleTrigger, - 'Trigger.MultiplePipelineTrigger' : MultiplePipelineTrigger, 'LinkedService.AzureFunction' : AzureFunctionLinkedService, 'LinkedService.AzureDataExplorer' : AzureDataExplorerLinkedService, 'LinkedService.SapTable' : SapTableLinkedService, @@ -22056,6 +22451,17 @@ export const discriminators = { 'Dataset.Parquet' : ParquetDataset, 'Dataset.Avro' : AvroDataset, 'Dataset.AmazonS3Object' : AmazonS3Dataset, + 'Trigger.ChainingTrigger' : ChainingTrigger, + 'Trigger.RerunTumblingWindowTrigger' : RerunTumblingWindowTrigger, + 'DependencyReference.SelfDependencyTumblingWindowTriggerReference' : SelfDependencyTumblingWindowTriggerReference, + 'DependencyReference.TumblingWindowTriggerDependencyReference' : TumblingWindowTriggerDependencyReference, + 'DependencyReference.TriggerDependencyReference' : TriggerDependencyReference, + 'DependencyReference' : DependencyReference, + 'Trigger.TumblingWindowTrigger' : TumblingWindowTrigger, + 'Trigger.BlobEventsTrigger' : BlobEventsTrigger, + 'Trigger.BlobTrigger' : BlobTrigger, + 'Trigger.ScheduleTrigger' : ScheduleTrigger, + 'Trigger.MultiplePipelineTrigger' : MultiplePipelineTrigger, 'Activity.ExecuteDataFlow' : ExecuteDataFlowActivity, 'Activity.AzureFunctionActivity' : AzureFunctionActivity, 'Activity.DatabricksSparkPython' : DatabricksSparkPythonActivity, diff --git a/sdk/datafactory/arm-datafactory/src/models/parameters.ts b/sdk/datafactory/arm-datafactory/src/models/parameters.ts index aafb78063d22..f636c40d2c55 100644 --- a/sdk/datafactory/arm-datafactory/src/models/parameters.ts +++ b/sdk/datafactory/arm-datafactory/src/models/parameters.ts @@ -216,21 +216,6 @@ export const referencePipelineRunId: msRest.OperationQueryParameter = { } } }; -export const rerunTriggerName: msRest.OperationURLParameter = { - parameterPath: "rerunTriggerName", - mapper: { - required: true, - serializedName: "rerunTriggerName", - constraints: { - MaxLength: 260, - MinLength: 1, - Pattern: /^[A-Za-z0-9_][^<>*#.%&:\\+?\/]*$/ - }, - type: { - name: "String" - } - } -}; export const resourceGroupName: msRest.OperationURLParameter = { parameterPath: "resourceGroupName", mapper: { @@ -268,6 +253,18 @@ export const startActivityName: msRest.OperationQueryParameter = { } } }; +export const startFromFailure: msRest.OperationQueryParameter = { + parameterPath: [ + "options", + "startFromFailure" + ], + mapper: { + serializedName: "startFromFailure", + type: { + name: "Boolean" + } + } +}; export const subscriptionId: msRest.OperationURLParameter = { parameterPath: "subscriptionId", mapper: { diff --git a/sdk/datafactory/arm-datafactory/src/models/pipelinesMappers.ts b/sdk/datafactory/arm-datafactory/src/models/pipelinesMappers.ts index 86d02b87ff9f..1b25f3169331 100644 --- a/sdk/datafactory/arm-datafactory/src/models/pipelinesMappers.ts +++ b/sdk/datafactory/arm-datafactory/src/models/pipelinesMappers.ts @@ -11,6 +11,7 @@ export { Activity, ActivityDependency, ActivityPolicy, + AdditionalColumns, AmazonMWSLinkedService, AmazonMWSObjectDataset, AmazonMWSSource, @@ -375,7 +376,6 @@ export { RedshiftUnloadSettings, RelationalSource, RelationalTableDataset, - RerunTriggerResource, RerunTumblingWindowTrigger, Resource, ResponsysLinkedService, @@ -435,6 +435,7 @@ export { ShopifyLinkedService, ShopifyObjectDataset, ShopifySource, + SkipErrorFile, SparkLinkedService, SparkObjectDataset, SparkSource, diff --git a/sdk/datafactory/arm-datafactory/src/models/rerunTriggersMappers.ts b/sdk/datafactory/arm-datafactory/src/models/rerunTriggersMappers.ts deleted file mode 100644 index f02c88a4f465..000000000000 --- a/sdk/datafactory/arm-datafactory/src/models/rerunTriggersMappers.ts +++ /dev/null @@ -1,510 +0,0 @@ -/* - * Copyright (c) Microsoft Corporation. All rights reserved. - * Licensed under the MIT License. See License.txt in the project root for license information. - * - * Code generated by Microsoft (R) AutoRest Code Generator. - * Changes may cause incorrect behavior and will be lost if the code is regenerated. - */ - -export { - discriminators, - Activity, - ActivityDependency, - ActivityPolicy, - AmazonMWSLinkedService, - AmazonMWSObjectDataset, - AmazonMWSSource, - AmazonRedshiftLinkedService, - AmazonRedshiftSource, - AmazonRedshiftTableDataset, - AmazonS3Dataset, - AmazonS3LinkedService, - AmazonS3Location, - AmazonS3ReadSettings, - AppendVariableActivity, - AvroDataset, - AvroFormat, - AvroSink, - AvroSource, - AvroWriteSettings, - AzureBatchLinkedService, - AzureBlobDataset, - AzureBlobFSDataset, - AzureBlobFSLinkedService, - AzureBlobFSLocation, - AzureBlobFSReadSettings, - AzureBlobFSSink, - AzureBlobFSSource, - AzureBlobFSWriteSettings, - AzureBlobStorageLinkedService, - AzureBlobStorageLocation, - AzureBlobStorageReadSettings, - AzureBlobStorageWriteSettings, - AzureDatabricksLinkedService, - AzureDataExplorerCommandActivity, - AzureDataExplorerLinkedService, - AzureDataExplorerSink, - AzureDataExplorerSource, - AzureDataExplorerTableDataset, - AzureDataLakeAnalyticsLinkedService, - AzureDataLakeStoreDataset, - AzureDataLakeStoreLinkedService, - AzureDataLakeStoreLocation, - AzureDataLakeStoreReadSettings, - AzureDataLakeStoreSink, - AzureDataLakeStoreSource, - AzureDataLakeStoreWriteSettings, - AzureFileStorageLinkedService, - AzureFileStorageLocation, - AzureFileStorageReadSettings, - AzureFunctionActivity, - AzureFunctionLinkedService, - AzureKeyVaultLinkedService, - AzureKeyVaultSecretReference, - AzureMariaDBLinkedService, - AzureMariaDBSource, - AzureMariaDBTableDataset, - AzureMLBatchExecutionActivity, - AzureMLExecutePipelineActivity, - AzureMLLinkedService, - AzureMLServiceLinkedService, - AzureMLUpdateResourceActivity, - AzureMLWebServiceFile, - AzureMySqlLinkedService, - AzureMySqlSink, - AzureMySqlSource, - AzureMySqlTableDataset, - AzurePostgreSqlLinkedService, - AzurePostgreSqlSink, - AzurePostgreSqlSource, - AzurePostgreSqlTableDataset, - AzureQueueSink, - AzureSearchIndexDataset, - AzureSearchIndexSink, - AzureSearchLinkedService, - AzureSqlDatabaseLinkedService, - AzureSqlDWLinkedService, - AzureSqlDWTableDataset, - AzureSqlMILinkedService, - AzureSqlMITableDataset, - AzureSqlSink, - AzureSqlSource, - AzureSqlTableDataset, - AzureStorageLinkedService, - AzureTableDataset, - AzureTableSink, - AzureTableSource, - AzureTableStorageLinkedService, - BaseResource, - BinaryDataset, - BinarySink, - BinarySource, - BlobEventsTrigger, - BlobSink, - BlobSource, - BlobTrigger, - CassandraLinkedService, - CassandraSource, - CassandraTableDataset, - ChainingTrigger, - CloudError, - CmdkeySetup, - CommonDataServiceForAppsEntityDataset, - CommonDataServiceForAppsLinkedService, - CommonDataServiceForAppsSink, - CommonDataServiceForAppsSource, - ComponentSetup, - ConcurLinkedService, - ConcurObjectDataset, - ConcurSource, - ControlActivity, - CopyActivity, - CopySink, - CopySource, - CosmosDbLinkedService, - CosmosDbMongoDbApiCollectionDataset, - CosmosDbMongoDbApiLinkedService, - CosmosDbMongoDbApiSink, - CosmosDbMongoDbApiSource, - CosmosDbSqlApiCollectionDataset, - CosmosDbSqlApiSink, - CosmosDbSqlApiSource, - CouchbaseLinkedService, - CouchbaseSource, - CouchbaseTableDataset, - CustomActivity, - CustomActivityReferenceObject, - CustomDataset, - CustomDataSourceLinkedService, - CustomSetupBase, - DatabricksNotebookActivity, - DatabricksSparkJarActivity, - DatabricksSparkPythonActivity, - DataFlow, - DataFlowFolder, - DataFlowReference, - DataFlowResource, - DataFlowSink, - DataFlowSource, - DataFlowStagingInfo, - DataLakeAnalyticsUSQLActivity, - Dataset, - DatasetBZip2Compression, - DatasetCompression, - DatasetDeflateCompression, - DatasetFolder, - DatasetGZipCompression, - DatasetLocation, - DatasetReference, - DatasetResource, - DatasetStorageFormat, - DatasetZipDeflateCompression, - Db2LinkedService, - Db2Source, - Db2TableDataset, - DeleteActivity, - DelimitedTextDataset, - DelimitedTextReadSettings, - DelimitedTextSink, - DelimitedTextSource, - DelimitedTextWriteSettings, - DependencyReference, - DistcpSettings, - DocumentDbCollectionDataset, - DocumentDbCollectionSink, - DocumentDbCollectionSource, - DrillLinkedService, - DrillSource, - DrillTableDataset, - DWCopyCommandDefaultValue, - DWCopyCommandSettings, - DynamicsAXLinkedService, - DynamicsAXResourceDataset, - DynamicsAXSource, - DynamicsCrmEntityDataset, - DynamicsCrmLinkedService, - DynamicsCrmSink, - DynamicsCrmSource, - DynamicsEntityDataset, - DynamicsLinkedService, - DynamicsSink, - DynamicsSource, - EloquaLinkedService, - EloquaObjectDataset, - EloquaSource, - EntityReference, - EnvironmentVariableSetup, - ExecuteDataFlowActivity, - ExecuteDataFlowActivityTypePropertiesCompute, - ExecutePipelineActivity, - ExecuteSSISPackageActivity, - ExecutionActivity, - Expression, - Factory, - FactoryGitHubConfiguration, - FactoryIdentity, - FactoryRepoConfiguration, - FactoryVSTSConfiguration, - FileServerLinkedService, - FileServerLocation, - FileServerReadSettings, - FileServerWriteSettings, - FileShareDataset, - FileSystemSink, - FileSystemSource, - FilterActivity, - ForEachActivity, - FormatReadSettings, - FormatWriteSettings, - FtpReadSettings, - FtpServerLinkedService, - FtpServerLocation, - GetMetadataActivity, - GoogleAdWordsLinkedService, - GoogleAdWordsObjectDataset, - GoogleAdWordsSource, - GoogleBigQueryLinkedService, - GoogleBigQueryObjectDataset, - GoogleBigQuerySource, - GoogleCloudStorageLinkedService, - GoogleCloudStorageLocation, - GoogleCloudStorageReadSettings, - GreenplumLinkedService, - GreenplumSource, - GreenplumTableDataset, - HBaseLinkedService, - HBaseObjectDataset, - HBaseSource, - HdfsLinkedService, - HdfsLocation, - HdfsReadSettings, - HdfsSource, - HDInsightHiveActivity, - HDInsightLinkedService, - HDInsightMapReduceActivity, - HDInsightOnDemandLinkedService, - HDInsightPigActivity, - HDInsightSparkActivity, - HDInsightStreamingActivity, - HiveLinkedService, - HiveObjectDataset, - HiveSource, - HttpDataset, - HttpLinkedService, - HttpReadSettings, - HttpServerLocation, - HttpSource, - HubspotLinkedService, - HubspotObjectDataset, - HubspotSource, - IfConditionActivity, - ImpalaLinkedService, - ImpalaObjectDataset, - ImpalaSource, - InformixLinkedService, - InformixSink, - InformixSource, - InformixTableDataset, - IntegrationRuntime, - IntegrationRuntimeComputeProperties, - IntegrationRuntimeCustomSetupScriptProperties, - IntegrationRuntimeDataFlowProperties, - IntegrationRuntimeDataProxyProperties, - IntegrationRuntimeReference, - IntegrationRuntimeResource, - IntegrationRuntimeSsisCatalogInfo, - IntegrationRuntimeSsisProperties, - IntegrationRuntimeVNetProperties, - JiraLinkedService, - JiraObjectDataset, - JiraSource, - JsonDataset, - JsonFormat, - JsonSink, - JsonSource, - JsonWriteSettings, - LinkedIntegrationRuntimeKeyAuthorization, - LinkedIntegrationRuntimeRbacAuthorization, - LinkedIntegrationRuntimeType, - LinkedService, - LinkedServiceReference, - LinkedServiceResource, - LogStorageSettings, - LookupActivity, - MagentoLinkedService, - MagentoObjectDataset, - MagentoSource, - ManagedIntegrationRuntime, - MappingDataFlow, - MariaDBLinkedService, - MariaDBSource, - MariaDBTableDataset, - MarketoLinkedService, - MarketoObjectDataset, - MarketoSource, - MicrosoftAccessLinkedService, - MicrosoftAccessSink, - MicrosoftAccessSource, - MicrosoftAccessTableDataset, - MongoDbCollectionDataset, - MongoDbCursorMethodsProperties, - MongoDbLinkedService, - MongoDbSource, - MongoDbV2CollectionDataset, - MongoDbV2LinkedService, - MongoDbV2Source, - MultiplePipelineTrigger, - MySqlLinkedService, - MySqlSource, - MySqlTableDataset, - NetezzaLinkedService, - NetezzaPartitionSettings, - NetezzaSource, - NetezzaTableDataset, - ODataLinkedService, - ODataResourceDataset, - ODataSource, - OdbcLinkedService, - OdbcSink, - OdbcSource, - OdbcTableDataset, - Office365Dataset, - Office365LinkedService, - Office365Source, - OracleLinkedService, - OraclePartitionSettings, - OracleServiceCloudLinkedService, - OracleServiceCloudObjectDataset, - OracleServiceCloudSource, - OracleSink, - OracleSource, - OracleTableDataset, - OrcDataset, - OrcFormat, - OrcSink, - OrcSource, - ParameterSpecification, - ParquetDataset, - ParquetFormat, - ParquetSink, - ParquetSource, - PaypalLinkedService, - PaypalObjectDataset, - PaypalSource, - PhoenixLinkedService, - PhoenixObjectDataset, - PhoenixSource, - PipelineFolder, - PipelineReference, - PipelineResource, - PolybaseSettings, - PostgreSqlLinkedService, - PostgreSqlSource, - PostgreSqlTableDataset, - PrestoLinkedService, - PrestoObjectDataset, - PrestoSource, - QuickBooksLinkedService, - QuickBooksObjectDataset, - QuickBooksSource, - RecurrenceSchedule, - RecurrenceScheduleOccurrence, - RedirectIncompatibleRowSettings, - RedshiftUnloadSettings, - RelationalSource, - RelationalTableDataset, - RerunTriggerListResponse, - RerunTriggerResource, - RerunTumblingWindowTrigger, - RerunTumblingWindowTriggerActionParameters, - Resource, - ResponsysLinkedService, - ResponsysObjectDataset, - ResponsysSource, - RestResourceDataset, - RestServiceLinkedService, - RestSource, - RetryPolicy, - SalesforceLinkedService, - SalesforceMarketingCloudLinkedService, - SalesforceMarketingCloudObjectDataset, - SalesforceMarketingCloudSource, - SalesforceObjectDataset, - SalesforceServiceCloudLinkedService, - SalesforceServiceCloudObjectDataset, - SalesforceServiceCloudSink, - SalesforceServiceCloudSource, - SalesforceSink, - SalesforceSource, - SapBwCubeDataset, - SapBWLinkedService, - SapBwSource, - SapCloudForCustomerLinkedService, - SapCloudForCustomerResourceDataset, - SapCloudForCustomerSink, - SapCloudForCustomerSource, - SapEccLinkedService, - SapEccResourceDataset, - SapEccSource, - SapHanaLinkedService, - SapHanaPartitionSettings, - SapHanaSource, - SapHanaTableDataset, - SapOpenHubLinkedService, - SapOpenHubSource, - SapOpenHubTableDataset, - SapTableLinkedService, - SapTablePartitionSettings, - SapTableResourceDataset, - SapTableSource, - ScheduleTrigger, - ScheduleTriggerRecurrence, - ScriptAction, - SecretBase, - SecureString, - SelfDependencyTumblingWindowTriggerReference, - SelfHostedIntegrationRuntime, - ServiceNowLinkedService, - ServiceNowObjectDataset, - ServiceNowSource, - SetVariableActivity, - SftpLocation, - SftpReadSettings, - SftpServerLinkedService, - SftpWriteSettings, - ShopifyLinkedService, - ShopifyObjectDataset, - ShopifySource, - SparkLinkedService, - SparkObjectDataset, - SparkSource, - SqlDWSink, - SqlDWSource, - SqlMISink, - SqlMISource, - SqlServerLinkedService, - SqlServerSink, - SqlServerSource, - SqlServerStoredProcedureActivity, - SqlServerTableDataset, - SqlSink, - SqlSource, - SquareLinkedService, - SquareObjectDataset, - SquareSource, - SSISAccessCredential, - SSISChildPackage, - SSISExecutionCredential, - SSISExecutionParameter, - SSISLogLocation, - SSISPackageLocation, - SSISPropertyOverride, - StagingSettings, - StoredProcedureParameter, - StoreReadSettings, - StoreWriteSettings, - SubResource, - SwitchActivity, - SwitchCase, - SybaseLinkedService, - SybaseSource, - SybaseTableDataset, - TabularSource, - TeradataLinkedService, - TeradataPartitionSettings, - TeradataSource, - TeradataTableDataset, - TextFormat, - Transformation, - Trigger, - TriggerDependencyReference, - TriggerPipelineReference, - TriggerReference, - TriggerResource, - TumblingWindowTrigger, - TumblingWindowTriggerDependencyReference, - UntilActivity, - UserProperty, - ValidationActivity, - VariableSpecification, - VerticaLinkedService, - VerticaSource, - VerticaTableDataset, - WaitActivity, - WebActivity, - WebActivityAuthentication, - WebAnonymousAuthentication, - WebBasicAuthentication, - WebClientCertificateAuthentication, - WebHookActivity, - WebLinkedService, - WebLinkedServiceTypeProperties, - WebSource, - WebTableDataset, - XeroLinkedService, - XeroObjectDataset, - XeroSource, - ZohoLinkedService, - ZohoObjectDataset, - ZohoSource -} from "../models/mappers"; diff --git a/sdk/datafactory/arm-datafactory/src/models/triggersMappers.ts b/sdk/datafactory/arm-datafactory/src/models/triggersMappers.ts index ce8f2bc7097d..c6cb571e82c9 100644 --- a/sdk/datafactory/arm-datafactory/src/models/triggersMappers.ts +++ b/sdk/datafactory/arm-datafactory/src/models/triggersMappers.ts @@ -11,6 +11,7 @@ export { Activity, ActivityDependency, ActivityPolicy, + AdditionalColumns, AmazonMWSLinkedService, AmazonMWSObjectDataset, AmazonMWSSource, @@ -373,7 +374,6 @@ export { RedshiftUnloadSettings, RelationalSource, RelationalTableDataset, - RerunTriggerResource, RerunTumblingWindowTrigger, Resource, ResponsysLinkedService, @@ -433,6 +433,7 @@ export { ShopifyLinkedService, ShopifyObjectDataset, ShopifySource, + SkipErrorFile, SparkLinkedService, SparkObjectDataset, SparkSource, @@ -476,8 +477,10 @@ export { Transformation, Trigger, TriggerDependencyReference, + TriggerFilterParameters, TriggerListResponse, TriggerPipelineReference, + TriggerQueryResponse, TriggerReference, TriggerResource, TriggerSubscriptionOperationStatus, diff --git a/sdk/datafactory/arm-datafactory/src/operations/index.ts b/sdk/datafactory/arm-datafactory/src/operations/index.ts index feda4fb51a5e..a4e051342b30 100644 --- a/sdk/datafactory/arm-datafactory/src/operations/index.ts +++ b/sdk/datafactory/arm-datafactory/src/operations/index.ts @@ -21,6 +21,5 @@ export * from "./pipelineRuns"; export * from "./activityRuns"; export * from "./triggers"; export * from "./triggerRuns"; -export * from "./rerunTriggers"; export * from "./dataFlows"; export * from "./dataFlowDebugSession"; diff --git a/sdk/datafactory/arm-datafactory/src/operations/pipelines.ts b/sdk/datafactory/arm-datafactory/src/operations/pipelines.ts index e0d7c1b37157..0b0d35920ac5 100644 --- a/sdk/datafactory/arm-datafactory/src/operations/pipelines.ts +++ b/sdk/datafactory/arm-datafactory/src/operations/pipelines.ts @@ -362,7 +362,8 @@ const createRunOperationSpec: msRest.OperationSpec = { Parameters.apiVersion, Parameters.referencePipelineRunId, Parameters.isRecovery, - Parameters.startActivityName + Parameters.startActivityName, + Parameters.startFromFailure ], headerParameters: [ Parameters.acceptLanguage diff --git a/sdk/datafactory/arm-datafactory/src/operations/rerunTriggers.ts b/sdk/datafactory/arm-datafactory/src/operations/rerunTriggers.ts deleted file mode 100644 index 22693ccbabba..000000000000 --- a/sdk/datafactory/arm-datafactory/src/operations/rerunTriggers.ts +++ /dev/null @@ -1,406 +0,0 @@ -/* - * Copyright (c) Microsoft Corporation. All rights reserved. - * Licensed under the MIT License. See License.txt in the project root for - * license information. - * - * Code generated by Microsoft (R) AutoRest Code Generator. - * Changes may cause incorrect behavior and will be lost if the code is - * regenerated. - */ - -import * as msRest from "@azure/ms-rest-js"; -import * as msRestAzure from "@azure/ms-rest-azure-js"; -import * as Models from "../models"; -import * as Mappers from "../models/rerunTriggersMappers"; -import * as Parameters from "../models/parameters"; -import { DataFactoryManagementClientContext } from "../dataFactoryManagementClientContext"; - -/** Class representing a RerunTriggers. */ -export class RerunTriggers { - private readonly client: DataFactoryManagementClientContext; - - /** - * Create a RerunTriggers. - * @param {DataFactoryManagementClientContext} client Reference to the service client. - */ - constructor(client: DataFactoryManagementClientContext) { - this.client = client; - } - - /** - * Creates a rerun trigger. - * @param resourceGroupName The resource group name. - * @param factoryName The factory name. - * @param triggerName The trigger name. - * @param rerunTriggerName The rerun trigger name. - * @param rerunTumblingWindowTriggerActionParameters Rerun tumbling window trigger action - * parameters. - * @param [options] The optional parameters - * @returns Promise - */ - create(resourceGroupName: string, factoryName: string, triggerName: string, rerunTriggerName: string, rerunTumblingWindowTriggerActionParameters: Models.RerunTumblingWindowTriggerActionParameters, options?: msRest.RequestOptionsBase): Promise; - /** - * @param resourceGroupName The resource group name. - * @param factoryName The factory name. - * @param triggerName The trigger name. - * @param rerunTriggerName The rerun trigger name. - * @param rerunTumblingWindowTriggerActionParameters Rerun tumbling window trigger action - * parameters. - * @param callback The callback - */ - create(resourceGroupName: string, factoryName: string, triggerName: string, rerunTriggerName: string, rerunTumblingWindowTriggerActionParameters: Models.RerunTumblingWindowTriggerActionParameters, callback: msRest.ServiceCallback): void; - /** - * @param resourceGroupName The resource group name. - * @param factoryName The factory name. - * @param triggerName The trigger name. - * @param rerunTriggerName The rerun trigger name. - * @param rerunTumblingWindowTriggerActionParameters Rerun tumbling window trigger action - * parameters. - * @param options The optional parameters - * @param callback The callback - */ - create(resourceGroupName: string, factoryName: string, triggerName: string, rerunTriggerName: string, rerunTumblingWindowTriggerActionParameters: Models.RerunTumblingWindowTriggerActionParameters, options: msRest.RequestOptionsBase, callback: msRest.ServiceCallback): void; - create(resourceGroupName: string, factoryName: string, triggerName: string, rerunTriggerName: string, rerunTumblingWindowTriggerActionParameters: Models.RerunTumblingWindowTriggerActionParameters, options?: msRest.RequestOptionsBase | msRest.ServiceCallback, callback?: msRest.ServiceCallback): Promise { - return this.client.sendOperationRequest( - { - resourceGroupName, - factoryName, - triggerName, - rerunTriggerName, - rerunTumblingWindowTriggerActionParameters, - options - }, - createOperationSpec, - callback) as Promise; - } - - /** - * Starts a trigger. - * @param resourceGroupName The resource group name. - * @param factoryName The factory name. - * @param triggerName The trigger name. - * @param rerunTriggerName The rerun trigger name. - * @param [options] The optional parameters - * @returns Promise - */ - start(resourceGroupName: string, factoryName: string, triggerName: string, rerunTriggerName: string, options?: msRest.RequestOptionsBase): Promise { - return this.beginStart(resourceGroupName,factoryName,triggerName,rerunTriggerName,options) - .then(lroPoller => lroPoller.pollUntilFinished()); - } - - /** - * Stops a trigger. - * @param resourceGroupName The resource group name. - * @param factoryName The factory name. - * @param triggerName The trigger name. - * @param rerunTriggerName The rerun trigger name. - * @param [options] The optional parameters - * @returns Promise - */ - stop(resourceGroupName: string, factoryName: string, triggerName: string, rerunTriggerName: string, options?: msRest.RequestOptionsBase): Promise { - return this.beginStop(resourceGroupName,factoryName,triggerName,rerunTriggerName,options) - .then(lroPoller => lroPoller.pollUntilFinished()); - } - - /** - * Cancels a trigger. - * @param resourceGroupName The resource group name. - * @param factoryName The factory name. - * @param triggerName The trigger name. - * @param rerunTriggerName The rerun trigger name. - * @param [options] The optional parameters - * @returns Promise - */ - cancel(resourceGroupName: string, factoryName: string, triggerName: string, rerunTriggerName: string, options?: msRest.RequestOptionsBase): Promise { - return this.beginCancel(resourceGroupName,factoryName,triggerName,rerunTriggerName,options) - .then(lroPoller => lroPoller.pollUntilFinished()); - } - - /** - * Lists rerun triggers by an original trigger name. - * @param resourceGroupName The resource group name. - * @param factoryName The factory name. - * @param triggerName The trigger name. - * @param [options] The optional parameters - * @returns Promise - */ - listByTrigger(resourceGroupName: string, factoryName: string, triggerName: string, options?: msRest.RequestOptionsBase): Promise; - /** - * @param resourceGroupName The resource group name. - * @param factoryName The factory name. - * @param triggerName The trigger name. - * @param callback The callback - */ - listByTrigger(resourceGroupName: string, factoryName: string, triggerName: string, callback: msRest.ServiceCallback): void; - /** - * @param resourceGroupName The resource group name. - * @param factoryName The factory name. - * @param triggerName The trigger name. - * @param options The optional parameters - * @param callback The callback - */ - listByTrigger(resourceGroupName: string, factoryName: string, triggerName: string, options: msRest.RequestOptionsBase, callback: msRest.ServiceCallback): void; - listByTrigger(resourceGroupName: string, factoryName: string, triggerName: string, options?: msRest.RequestOptionsBase | msRest.ServiceCallback, callback?: msRest.ServiceCallback): Promise { - return this.client.sendOperationRequest( - { - resourceGroupName, - factoryName, - triggerName, - options - }, - listByTriggerOperationSpec, - callback) as Promise; - } - - /** - * Starts a trigger. - * @param resourceGroupName The resource group name. - * @param factoryName The factory name. - * @param triggerName The trigger name. - * @param rerunTriggerName The rerun trigger name. - * @param [options] The optional parameters - * @returns Promise - */ - beginStart(resourceGroupName: string, factoryName: string, triggerName: string, rerunTriggerName: string, options?: msRest.RequestOptionsBase): Promise { - return this.client.sendLRORequest( - { - resourceGroupName, - factoryName, - triggerName, - rerunTriggerName, - options - }, - beginStartOperationSpec, - options); - } - - /** - * Stops a trigger. - * @param resourceGroupName The resource group name. - * @param factoryName The factory name. - * @param triggerName The trigger name. - * @param rerunTriggerName The rerun trigger name. - * @param [options] The optional parameters - * @returns Promise - */ - beginStop(resourceGroupName: string, factoryName: string, triggerName: string, rerunTriggerName: string, options?: msRest.RequestOptionsBase): Promise { - return this.client.sendLRORequest( - { - resourceGroupName, - factoryName, - triggerName, - rerunTriggerName, - options - }, - beginStopOperationSpec, - options); - } - - /** - * Cancels a trigger. - * @param resourceGroupName The resource group name. - * @param factoryName The factory name. - * @param triggerName The trigger name. - * @param rerunTriggerName The rerun trigger name. - * @param [options] The optional parameters - * @returns Promise - */ - beginCancel(resourceGroupName: string, factoryName: string, triggerName: string, rerunTriggerName: string, options?: msRest.RequestOptionsBase): Promise { - return this.client.sendLRORequest( - { - resourceGroupName, - factoryName, - triggerName, - rerunTriggerName, - options - }, - beginCancelOperationSpec, - options); - } - - /** - * Lists rerun triggers by an original trigger name. - * @param nextPageLink The NextLink from the previous successful call to List operation. - * @param [options] The optional parameters - * @returns Promise - */ - listByTriggerNext(nextPageLink: string, options?: msRest.RequestOptionsBase): Promise; - /** - * @param nextPageLink The NextLink from the previous successful call to List operation. - * @param callback The callback - */ - listByTriggerNext(nextPageLink: string, callback: msRest.ServiceCallback): void; - /** - * @param nextPageLink The NextLink from the previous successful call to List operation. - * @param options The optional parameters - * @param callback The callback - */ - listByTriggerNext(nextPageLink: string, options: msRest.RequestOptionsBase, callback: msRest.ServiceCallback): void; - listByTriggerNext(nextPageLink: string, options?: msRest.RequestOptionsBase | msRest.ServiceCallback, callback?: msRest.ServiceCallback): Promise { - return this.client.sendOperationRequest( - { - nextPageLink, - options - }, - listByTriggerNextOperationSpec, - callback) as Promise; - } -} - -// Operation Specifications -const serializer = new msRest.Serializer(Mappers); -const createOperationSpec: msRest.OperationSpec = { - httpMethod: "PUT", - path: "subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.DataFactory/factories/{factoryName}/triggers/{triggerName}/rerunTriggers/{rerunTriggerName}", - urlParameters: [ - Parameters.subscriptionId, - Parameters.resourceGroupName, - Parameters.factoryName, - Parameters.triggerName, - Parameters.rerunTriggerName - ], - queryParameters: [ - Parameters.apiVersion - ], - headerParameters: [ - Parameters.acceptLanguage - ], - requestBody: { - parameterPath: "rerunTumblingWindowTriggerActionParameters", - mapper: { - ...Mappers.RerunTumblingWindowTriggerActionParameters, - required: true - } - }, - responses: { - 200: { - bodyMapper: Mappers.TriggerResource - }, - default: { - bodyMapper: Mappers.CloudError - } - }, - serializer -}; - -const listByTriggerOperationSpec: msRest.OperationSpec = { - httpMethod: "GET", - path: "subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.DataFactory/factories/{factoryName}/triggers/{triggerName}/rerunTriggers", - urlParameters: [ - Parameters.subscriptionId, - Parameters.resourceGroupName, - Parameters.factoryName, - Parameters.triggerName - ], - queryParameters: [ - Parameters.apiVersion - ], - headerParameters: [ - Parameters.acceptLanguage - ], - responses: { - 200: { - bodyMapper: Mappers.RerunTriggerListResponse - }, - default: { - bodyMapper: Mappers.CloudError - } - }, - serializer -}; - -const beginStartOperationSpec: msRest.OperationSpec = { - httpMethod: "POST", - path: "subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.DataFactory/factories/{factoryName}/triggers/{triggerName}/rerunTriggers/{rerunTriggerName}/start", - urlParameters: [ - Parameters.subscriptionId, - Parameters.resourceGroupName, - Parameters.factoryName, - Parameters.triggerName, - Parameters.rerunTriggerName - ], - queryParameters: [ - Parameters.apiVersion - ], - headerParameters: [ - Parameters.acceptLanguage - ], - responses: { - 200: {}, - default: { - bodyMapper: Mappers.CloudError - } - }, - serializer -}; - -const beginStopOperationSpec: msRest.OperationSpec = { - httpMethod: "POST", - path: "subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.DataFactory/factories/{factoryName}/triggers/{triggerName}/rerunTriggers/{rerunTriggerName}/stop", - urlParameters: [ - Parameters.subscriptionId, - Parameters.resourceGroupName, - Parameters.factoryName, - Parameters.triggerName, - Parameters.rerunTriggerName - ], - queryParameters: [ - Parameters.apiVersion - ], - headerParameters: [ - Parameters.acceptLanguage - ], - responses: { - 200: {}, - default: { - bodyMapper: Mappers.CloudError - } - }, - serializer -}; - -const beginCancelOperationSpec: msRest.OperationSpec = { - httpMethod: "POST", - path: "subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.DataFactory/factories/{factoryName}/triggers/{triggerName}/rerunTriggers/{rerunTriggerName}/cancel", - urlParameters: [ - Parameters.subscriptionId, - Parameters.resourceGroupName, - Parameters.factoryName, - Parameters.triggerName, - Parameters.rerunTriggerName - ], - queryParameters: [ - Parameters.apiVersion - ], - headerParameters: [ - Parameters.acceptLanguage - ], - responses: { - 200: {}, - default: { - bodyMapper: Mappers.CloudError - } - }, - serializer -}; - -const listByTriggerNextOperationSpec: msRest.OperationSpec = { - httpMethod: "GET", - baseUrl: "https://management.azure.com", - path: "{nextLink}", - urlParameters: [ - Parameters.nextPageLink - ], - headerParameters: [ - Parameters.acceptLanguage - ], - responses: { - 200: { - bodyMapper: Mappers.RerunTriggerListResponse - }, - default: { - bodyMapper: Mappers.CloudError - } - }, - serializer -}; diff --git a/sdk/datafactory/arm-datafactory/src/operations/triggers.ts b/sdk/datafactory/arm-datafactory/src/operations/triggers.ts index 49a41a8960d2..5218d661220e 100644 --- a/sdk/datafactory/arm-datafactory/src/operations/triggers.ts +++ b/sdk/datafactory/arm-datafactory/src/operations/triggers.ts @@ -59,6 +59,42 @@ export class Triggers { callback) as Promise; } + /** + * Query triggers. + * @param resourceGroupName The resource group name. + * @param factoryName The factory name. + * @param filterParameters Parameters to filter the triggers. + * @param [options] The optional parameters + * @returns Promise + */ + queryByFactory(resourceGroupName: string, factoryName: string, filterParameters: Models.TriggerFilterParameters, options?: msRest.RequestOptionsBase): Promise; + /** + * @param resourceGroupName The resource group name. + * @param factoryName The factory name. + * @param filterParameters Parameters to filter the triggers. + * @param callback The callback + */ + queryByFactory(resourceGroupName: string, factoryName: string, filterParameters: Models.TriggerFilterParameters, callback: msRest.ServiceCallback): void; + /** + * @param resourceGroupName The resource group name. + * @param factoryName The factory name. + * @param filterParameters Parameters to filter the triggers. + * @param options The optional parameters + * @param callback The callback + */ + queryByFactory(resourceGroupName: string, factoryName: string, filterParameters: Models.TriggerFilterParameters, options: msRest.RequestOptionsBase, callback: msRest.ServiceCallback): void; + queryByFactory(resourceGroupName: string, factoryName: string, filterParameters: Models.TriggerFilterParameters, options?: msRest.RequestOptionsBase | msRest.ServiceCallback, callback?: msRest.ServiceCallback): Promise { + return this.client.sendOperationRequest( + { + resourceGroupName, + factoryName, + filterParameters, + options + }, + queryByFactoryOperationSpec, + callback) as Promise; + } + /** * Creates or updates a trigger. * @param resourceGroupName The resource group name. @@ -395,6 +431,38 @@ const listByFactoryOperationSpec: msRest.OperationSpec = { serializer }; +const queryByFactoryOperationSpec: msRest.OperationSpec = { + httpMethod: "POST", + path: "subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.DataFactory/factories/{factoryName}/querytriggers", + urlParameters: [ + Parameters.subscriptionId, + Parameters.resourceGroupName, + Parameters.factoryName + ], + queryParameters: [ + Parameters.apiVersion + ], + headerParameters: [ + Parameters.acceptLanguage + ], + requestBody: { + parameterPath: "filterParameters", + mapper: { + ...Mappers.TriggerFilterParameters, + required: true + } + }, + responses: { + 200: { + bodyMapper: Mappers.TriggerQueryResponse + }, + default: { + bodyMapper: Mappers.CloudError + } + }, + serializer +}; + const createOrUpdateOperationSpec: msRest.OperationSpec = { httpMethod: "PUT", path: "subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.DataFactory/factories/{factoryName}/triggers/{triggerName}",