From ea145ff4992a6181f93f67a5005ed8598ee74fc3 Mon Sep 17 00:00:00 2001 From: SDK Automation Date: Mon, 9 Mar 2020 08:48:04 +0000 Subject: [PATCH] Generated from 700e3874a1554d3918adebf8c57526cf48deb625 fix prettry check --- .../src/dataFactoryManagementClient.ts | 2 - .../src/models/dataFlowDebugSessionMappers.ts | 2 + .../src/models/dataFlowsMappers.ts | 11 +- .../src/models/datasetsMappers.ts | 11 +- .../src/models/factoriesMappers.ts | 11 +- .../arm-datafactory/src/models/index.ts | 8724 +++++++++-------- .../src/models/integrationRuntimesMappers.ts | 11 +- .../src/models/linkedServicesMappers.ts | 11 +- .../arm-datafactory/src/models/mappers.ts | 3798 ++++--- .../arm-datafactory/src/models/parameters.ts | 27 +- .../src/models/pipelinesMappers.ts | 11 +- .../src/models/rerunTriggersMappers.ts | 510 - .../src/models/triggersMappers.ts | 13 +- .../arm-datafactory/src/operations/index.ts | 1 - .../src/operations/pipelines.ts | 3 +- .../src/operations/rerunTriggers.ts | 406 - .../src/operations/triggers.ts | 68 + 17 files changed, 7106 insertions(+), 6514 deletions(-) delete mode 100644 sdk/datafactory/arm-datafactory/src/models/rerunTriggersMappers.ts delete mode 100644 sdk/datafactory/arm-datafactory/src/operations/rerunTriggers.ts diff --git a/sdk/datafactory/arm-datafactory/src/dataFactoryManagementClient.ts b/sdk/datafactory/arm-datafactory/src/dataFactoryManagementClient.ts index 21184128a935..be96a827a17d 100644 --- a/sdk/datafactory/arm-datafactory/src/dataFactoryManagementClient.ts +++ b/sdk/datafactory/arm-datafactory/src/dataFactoryManagementClient.ts @@ -30,7 +30,6 @@ class DataFactoryManagementClient extends DataFactoryManagementClientContext { activityRuns: operations.ActivityRuns; triggers: operations.Triggers; triggerRuns: operations.TriggerRuns; - rerunTriggers: operations.RerunTriggers; dataFlows: operations.DataFlows; dataFlowDebugSession: operations.DataFlowDebugSession; @@ -55,7 +54,6 @@ class DataFactoryManagementClient extends DataFactoryManagementClientContext { this.activityRuns = new operations.ActivityRuns(this); this.triggers = new operations.Triggers(this); this.triggerRuns = new operations.TriggerRuns(this); - this.rerunTriggers = new operations.RerunTriggers(this); this.dataFlows = new operations.DataFlows(this); this.dataFlowDebugSession = new operations.DataFlowDebugSession(this); } diff --git a/sdk/datafactory/arm-datafactory/src/models/dataFlowDebugSessionMappers.ts b/sdk/datafactory/arm-datafactory/src/models/dataFlowDebugSessionMappers.ts index 0d470f86b98c..91593c3c53a9 100644 --- a/sdk/datafactory/arm-datafactory/src/models/dataFlowDebugSessionMappers.ts +++ b/sdk/datafactory/arm-datafactory/src/models/dataFlowDebugSessionMappers.ts @@ -247,6 +247,8 @@ export { SftpServerLinkedService, ShopifyLinkedService, ShopifyObjectDataset, + SnowflakeDataset, + SnowflakeLinkedService, SparkLinkedService, SparkObjectDataset, SqlServerLinkedService, diff --git a/sdk/datafactory/arm-datafactory/src/models/dataFlowsMappers.ts b/sdk/datafactory/arm-datafactory/src/models/dataFlowsMappers.ts index f302ede884dc..8cfcd369859f 100644 --- a/sdk/datafactory/arm-datafactory/src/models/dataFlowsMappers.ts +++ b/sdk/datafactory/arm-datafactory/src/models/dataFlowsMappers.ts @@ -11,6 +11,7 @@ export { Activity, ActivityDependency, ActivityPolicy, + AdditionalColumns, AmazonMWSLinkedService, AmazonMWSObjectDataset, AmazonMWSSource, @@ -200,6 +201,7 @@ export { ExecutePipelineActivity, ExecuteSSISPackageActivity, ExecutionActivity, + ExportSettings, Expression, Factory, FactoryGitHubConfiguration, @@ -262,6 +264,7 @@ export { ImpalaLinkedService, ImpalaObjectDataset, ImpalaSource, + ImportSettings, InformixLinkedService, InformixSink, InformixSource, @@ -374,7 +377,6 @@ export { RedshiftUnloadSettings, RelationalSource, RelationalTableDataset, - RerunTriggerResource, RerunTumblingWindowTrigger, Resource, ResponsysLinkedService, @@ -434,6 +436,13 @@ export { ShopifyLinkedService, ShopifyObjectDataset, ShopifySource, + SkipErrorFile, + SnowflakeDataset, + SnowflakeExportCopyCommand, + SnowflakeImportCopyCommand, + SnowflakeLinkedService, + SnowflakeSink, + SnowflakeSource, SparkLinkedService, SparkObjectDataset, SparkSource, diff --git a/sdk/datafactory/arm-datafactory/src/models/datasetsMappers.ts b/sdk/datafactory/arm-datafactory/src/models/datasetsMappers.ts index bb0bb8162f24..70411706127a 100644 --- a/sdk/datafactory/arm-datafactory/src/models/datasetsMappers.ts +++ b/sdk/datafactory/arm-datafactory/src/models/datasetsMappers.ts @@ -11,6 +11,7 @@ export { Activity, ActivityDependency, ActivityPolicy, + AdditionalColumns, AmazonMWSLinkedService, AmazonMWSObjectDataset, AmazonMWSSource, @@ -200,6 +201,7 @@ export { ExecutePipelineActivity, ExecuteSSISPackageActivity, ExecutionActivity, + ExportSettings, Expression, Factory, FactoryGitHubConfiguration, @@ -262,6 +264,7 @@ export { ImpalaLinkedService, ImpalaObjectDataset, ImpalaSource, + ImportSettings, InformixLinkedService, InformixSink, InformixSource, @@ -374,7 +377,6 @@ export { RedshiftUnloadSettings, RelationalSource, RelationalTableDataset, - RerunTriggerResource, RerunTumblingWindowTrigger, Resource, ResponsysLinkedService, @@ -434,6 +436,13 @@ export { ShopifyLinkedService, ShopifyObjectDataset, ShopifySource, + SkipErrorFile, + SnowflakeDataset, + SnowflakeExportCopyCommand, + SnowflakeImportCopyCommand, + SnowflakeLinkedService, + SnowflakeSink, + SnowflakeSource, SparkLinkedService, SparkObjectDataset, SparkSource, diff --git a/sdk/datafactory/arm-datafactory/src/models/factoriesMappers.ts b/sdk/datafactory/arm-datafactory/src/models/factoriesMappers.ts index a29de38c9fba..866f45f542c1 100644 --- a/sdk/datafactory/arm-datafactory/src/models/factoriesMappers.ts +++ b/sdk/datafactory/arm-datafactory/src/models/factoriesMappers.ts @@ -12,6 +12,7 @@ export { Activity, ActivityDependency, ActivityPolicy, + AdditionalColumns, AmazonMWSLinkedService, AmazonMWSObjectDataset, AmazonMWSSource, @@ -200,6 +201,7 @@ export { ExecutePipelineActivity, ExecuteSSISPackageActivity, ExecutionActivity, + ExportSettings, Expression, Factory, FactoryGitHubConfiguration, @@ -267,6 +269,7 @@ export { ImpalaLinkedService, ImpalaObjectDataset, ImpalaSource, + ImportSettings, InformixLinkedService, InformixSink, InformixSource, @@ -379,7 +382,6 @@ export { RedshiftUnloadSettings, RelationalSource, RelationalTableDataset, - RerunTriggerResource, RerunTumblingWindowTrigger, Resource, ResponsysLinkedService, @@ -439,6 +441,13 @@ export { ShopifyLinkedService, ShopifyObjectDataset, ShopifySource, + SkipErrorFile, + SnowflakeDataset, + SnowflakeExportCopyCommand, + SnowflakeImportCopyCommand, + SnowflakeLinkedService, + SnowflakeSink, + SnowflakeSource, SparkLinkedService, SparkObjectDataset, SparkSource, diff --git a/sdk/datafactory/arm-datafactory/src/models/index.ts b/sdk/datafactory/arm-datafactory/src/models/index.ts index f9c994ec0132..337937487268 100644 --- a/sdk/datafactory/arm-datafactory/src/models/index.ts +++ b/sdk/datafactory/arm-datafactory/src/models/index.ts @@ -436,7 +436,7 @@ export interface ParameterSpecification { /** * Contains the possible cases for LinkedService. */ -export type LinkedServiceUnion = LinkedService | AzureFunctionLinkedService | AzureDataExplorerLinkedService | SapTableLinkedService | GoogleAdWordsLinkedService | OracleServiceCloudLinkedService | DynamicsAXLinkedService | ResponsysLinkedService | AzureDatabricksLinkedService | AzureDataLakeAnalyticsLinkedService | HDInsightOnDemandLinkedService | SalesforceMarketingCloudLinkedService | NetezzaLinkedService | VerticaLinkedService | ZohoLinkedService | XeroLinkedService | SquareLinkedService | SparkLinkedService | ShopifyLinkedService | ServiceNowLinkedService | QuickBooksLinkedService | PrestoLinkedService | PhoenixLinkedService | PaypalLinkedService | MarketoLinkedService | AzureMariaDBLinkedService | MariaDBLinkedService | MagentoLinkedService | JiraLinkedService | ImpalaLinkedService | HubspotLinkedService | HiveLinkedService | HBaseLinkedService | GreenplumLinkedService | GoogleBigQueryLinkedService | EloquaLinkedService | DrillLinkedService | CouchbaseLinkedService | ConcurLinkedService | AzurePostgreSqlLinkedService | AmazonMWSLinkedService | SapHanaLinkedService | SapBWLinkedService | SftpServerLinkedService | FtpServerLinkedService | HttpLinkedService | AzureSearchLinkedService | CustomDataSourceLinkedService | AmazonRedshiftLinkedService | AmazonS3LinkedService | RestServiceLinkedService | SapOpenHubLinkedService | SapEccLinkedService | SapCloudForCustomerLinkedService | SalesforceServiceCloudLinkedService | SalesforceLinkedService | Office365LinkedService | AzureBlobFSLinkedService | AzureDataLakeStoreLinkedService | CosmosDbMongoDbApiLinkedService | MongoDbV2LinkedService | MongoDbLinkedService | CassandraLinkedService | WebLinkedService | ODataLinkedService | HdfsLinkedService | MicrosoftAccessLinkedService | InformixLinkedService | OdbcLinkedService | AzureMLServiceLinkedService | AzureMLLinkedService | TeradataLinkedService | Db2LinkedService | SybaseLinkedService | PostgreSqlLinkedService | MySqlLinkedService | AzureMySqlLinkedService | OracleLinkedService | GoogleCloudStorageLinkedService | AzureFileStorageLinkedService | FileServerLinkedService | HDInsightLinkedService | CommonDataServiceForAppsLinkedService | DynamicsCrmLinkedService | DynamicsLinkedService | CosmosDbLinkedService | AzureKeyVaultLinkedService | AzureBatchLinkedService | AzureSqlMILinkedService | AzureSqlDatabaseLinkedService | SqlServerLinkedService | AzureSqlDWLinkedService | AzureTableStorageLinkedService | AzureBlobStorageLinkedService | AzureStorageLinkedService; +export type LinkedServiceUnion = LinkedService | SnowflakeLinkedService | AzureFunctionLinkedService | AzureDataExplorerLinkedService | SapTableLinkedService | GoogleAdWordsLinkedService | OracleServiceCloudLinkedService | DynamicsAXLinkedService | ResponsysLinkedService | AzureDatabricksLinkedService | AzureDataLakeAnalyticsLinkedService | HDInsightOnDemandLinkedService | SalesforceMarketingCloudLinkedService | NetezzaLinkedService | VerticaLinkedService | ZohoLinkedService | XeroLinkedService | SquareLinkedService | SparkLinkedService | ShopifyLinkedService | ServiceNowLinkedService | QuickBooksLinkedService | PrestoLinkedService | PhoenixLinkedService | PaypalLinkedService | MarketoLinkedService | AzureMariaDBLinkedService | MariaDBLinkedService | MagentoLinkedService | JiraLinkedService | ImpalaLinkedService | HubspotLinkedService | HiveLinkedService | HBaseLinkedService | GreenplumLinkedService | GoogleBigQueryLinkedService | EloquaLinkedService | DrillLinkedService | CouchbaseLinkedService | ConcurLinkedService | AzurePostgreSqlLinkedService | AmazonMWSLinkedService | SapHanaLinkedService | SapBWLinkedService | SftpServerLinkedService | FtpServerLinkedService | HttpLinkedService | AzureSearchLinkedService | CustomDataSourceLinkedService | AmazonRedshiftLinkedService | AmazonS3LinkedService | RestServiceLinkedService | SapOpenHubLinkedService | SapEccLinkedService | SapCloudForCustomerLinkedService | SalesforceServiceCloudLinkedService | SalesforceLinkedService | Office365LinkedService | AzureBlobFSLinkedService | AzureDataLakeStoreLinkedService | CosmosDbMongoDbApiLinkedService | MongoDbV2LinkedService | MongoDbLinkedService | CassandraLinkedService | WebLinkedService | ODataLinkedService | HdfsLinkedService | MicrosoftAccessLinkedService | InformixLinkedService | OdbcLinkedService | AzureMLServiceLinkedService | AzureMLLinkedService | TeradataLinkedService | Db2LinkedService | SybaseLinkedService | PostgreSqlLinkedService | MySqlLinkedService | AzureMySqlLinkedService | OracleLinkedService | GoogleCloudStorageLinkedService | AzureFileStorageLinkedService | FileServerLinkedService | HDInsightLinkedService | CommonDataServiceForAppsLinkedService | DynamicsCrmLinkedService | DynamicsLinkedService | CosmosDbLinkedService | AzureKeyVaultLinkedService | AzureBatchLinkedService | AzureSqlMILinkedService | AzureSqlDatabaseLinkedService | SqlServerLinkedService | AzureSqlDWLinkedService | AzureTableStorageLinkedService | AzureBlobStorageLinkedService | AzureStorageLinkedService; /** * The Azure Data Factory nested object which contains the information and credential which can be @@ -492,7 +492,7 @@ export interface DatasetFolder { /** * Contains the possible cases for Dataset. */ -export type DatasetUnion = Dataset | GoogleAdWordsObjectDataset | AzureDataExplorerTableDataset | OracleServiceCloudObjectDataset | DynamicsAXResourceDataset | ResponsysObjectDataset | SalesforceMarketingCloudObjectDataset | VerticaTableDataset | NetezzaTableDataset | ZohoObjectDataset | XeroObjectDataset | SquareObjectDataset | SparkObjectDataset | ShopifyObjectDataset | ServiceNowObjectDataset | QuickBooksObjectDataset | PrestoObjectDataset | PhoenixObjectDataset | PaypalObjectDataset | MarketoObjectDataset | AzureMariaDBTableDataset | MariaDBTableDataset | MagentoObjectDataset | JiraObjectDataset | ImpalaObjectDataset | HubspotObjectDataset | HiveObjectDataset | HBaseObjectDataset | GreenplumTableDataset | GoogleBigQueryObjectDataset | EloquaObjectDataset | DrillTableDataset | CouchbaseTableDataset | ConcurObjectDataset | AzurePostgreSqlTableDataset | AmazonMWSObjectDataset | HttpDataset | AzureSearchIndexDataset | WebTableDataset | SapTableResourceDataset | RestResourceDataset | SqlServerTableDataset | SapOpenHubTableDataset | SapHanaTableDataset | SapEccResourceDataset | SapCloudForCustomerResourceDataset | SapBwCubeDataset | SybaseTableDataset | SalesforceServiceCloudObjectDataset | SalesforceObjectDataset | MicrosoftAccessTableDataset | PostgreSqlTableDataset | MySqlTableDataset | OdbcTableDataset | InformixTableDataset | RelationalTableDataset | Db2TableDataset | AmazonRedshiftTableDataset | AzureMySqlTableDataset | TeradataTableDataset | OracleTableDataset | ODataResourceDataset | CosmosDbMongoDbApiCollectionDataset | MongoDbV2CollectionDataset | MongoDbCollectionDataset | FileShareDataset | Office365Dataset | AzureBlobFSDataset | AzureDataLakeStoreDataset | CommonDataServiceForAppsEntityDataset | DynamicsCrmEntityDataset | DynamicsEntityDataset | DocumentDbCollectionDataset | CosmosDbSqlApiCollectionDataset | CustomDataset | CassandraTableDataset | AzureSqlDWTableDataset | AzureSqlMITableDataset | AzureSqlTableDataset | AzureTableDataset | AzureBlobDataset | BinaryDataset | OrcDataset | JsonDataset | DelimitedTextDataset | ParquetDataset | AvroDataset | AmazonS3Dataset; +export type DatasetUnion = Dataset | SnowflakeDataset | GoogleAdWordsObjectDataset | AzureDataExplorerTableDataset | OracleServiceCloudObjectDataset | DynamicsAXResourceDataset | ResponsysObjectDataset | SalesforceMarketingCloudObjectDataset | VerticaTableDataset | NetezzaTableDataset | ZohoObjectDataset | XeroObjectDataset | SquareObjectDataset | SparkObjectDataset | ShopifyObjectDataset | ServiceNowObjectDataset | QuickBooksObjectDataset | PrestoObjectDataset | PhoenixObjectDataset | PaypalObjectDataset | MarketoObjectDataset | AzureMariaDBTableDataset | MariaDBTableDataset | MagentoObjectDataset | JiraObjectDataset | ImpalaObjectDataset | HubspotObjectDataset | HiveObjectDataset | HBaseObjectDataset | GreenplumTableDataset | GoogleBigQueryObjectDataset | EloquaObjectDataset | DrillTableDataset | CouchbaseTableDataset | ConcurObjectDataset | AzurePostgreSqlTableDataset | AmazonMWSObjectDataset | HttpDataset | AzureSearchIndexDataset | WebTableDataset | SapTableResourceDataset | RestResourceDataset | SqlServerTableDataset | SapOpenHubTableDataset | SapHanaTableDataset | SapEccResourceDataset | SapCloudForCustomerResourceDataset | SapBwCubeDataset | SybaseTableDataset | SalesforceServiceCloudObjectDataset | SalesforceObjectDataset | MicrosoftAccessTableDataset | PostgreSqlTableDataset | MySqlTableDataset | OdbcTableDataset | InformixTableDataset | RelationalTableDataset | Db2TableDataset | AmazonRedshiftTableDataset | AzureMySqlTableDataset | TeradataTableDataset | OracleTableDataset | ODataResourceDataset | CosmosDbMongoDbApiCollectionDataset | MongoDbV2CollectionDataset | MongoDbCollectionDataset | FileShareDataset | Office365Dataset | AzureBlobFSDataset | AzureDataLakeStoreDataset | CommonDataServiceForAppsEntityDataset | DynamicsCrmEntityDataset | DynamicsEntityDataset | DocumentDbCollectionDataset | CosmosDbSqlApiCollectionDataset | CustomDataset | CassandraTableDataset | AzureSqlDWTableDataset | AzureSqlMITableDataset | AzureSqlTableDataset | AzureTableDataset | AzureBlobDataset | BinaryDataset | OrcDataset | JsonDataset | DelimitedTextDataset | ParquetDataset | AvroDataset | AmazonS3Dataset; /** * The Azure Data Factory nested object which identifies data within different data stores, such as @@ -685,7 +685,7 @@ export interface PipelineResource extends SubResource { /** * Contains the possible cases for Trigger. */ -export type TriggerUnion = Trigger | RerunTumblingWindowTrigger | ChainingTrigger | TumblingWindowTrigger | MultiplePipelineTriggerUnion; +export type TriggerUnion = Trigger | ChainingTrigger | RerunTumblingWindowTrigger | TumblingWindowTrigger | MultiplePipelineTriggerUnion; /** * Azure data factory nested object which contains information about creating pipeline run @@ -725,6 +725,21 @@ export interface TriggerResource extends SubResource { properties: TriggerUnion; } +/** + * A query of triggers. + */ +export interface TriggerQueryResponse { + /** + * List of triggers. + */ + value: TriggerResource[]; + /** + * The continuation token for getting the next page of results, if any remaining results exist, + * null otherwise. + */ + continuationToken?: string; +} + /** * Response body with a run identifier. */ @@ -990,6 +1005,20 @@ export interface LinkedServiceDebugResource extends SubResourceDebugResource { properties: LinkedServiceUnion; } +/** + * Query parameters for triggers. + */ +export interface TriggerFilterParameters { + /** + * The continuation token for getting the next page of results. Null for first page. + */ + continuationToken?: string; + /** + * The name of the parent TumblingWindowTrigger to get the child rerun triggers + */ + parentTriggerName?: string; +} + /** * Query filter option for listing runs. */ @@ -1335,79 +1364,6 @@ export interface TriggerRunsQueryResponse { continuationToken?: string; } -/** - * Rerun tumbling window trigger Parameters. - */ -export interface RerunTumblingWindowTriggerActionParameters { - /** - * The start time for the time period for which restatement is initiated. Only UTC time is - * currently supported. - */ - startTime: Date; - /** - * The end time for the time period for which restatement is initiated. Only UTC time is - * currently supported. - */ - endTime: Date; - /** - * The max number of parallel time windows (ready for execution) for which a rerun is triggered. - */ - maxConcurrency: number; -} - -/** - * Trigger that schedules pipeline reruns for all fixed time interval windows from a requested - * start time to requested end time. - */ -export interface RerunTumblingWindowTrigger { - /** - * Polymorphic Discriminator - */ - type: "RerunTumblingWindowTrigger"; - /** - * Trigger description. - */ - description?: string; - /** - * Indicates if trigger is running or not. Updated when Start/Stop APIs are called on the - * Trigger. Possible values include: 'Started', 'Stopped', 'Disabled' - * **NOTE: This property will not be serialized. It can only be populated by the server.** - */ - readonly runtimeState?: TriggerRuntimeState; - /** - * List of tags that can be used for describing the trigger. - */ - annotations?: any[]; - /** - * The parent trigger reference. - */ - parentTrigger?: any; - /** - * The start time for the time period for which restatement is initiated. Only UTC time is - * currently supported. - */ - requestedStartTime: Date; - /** - * The end time for the time period for which restatement is initiated. Only UTC time is - * currently supported. - */ - requestedEndTime: Date; - /** - * The max number of parallel time windows (ready for execution) for which a rerun is triggered. - */ - maxConcurrency: number; -} - -/** - * RerunTrigger resource type. - */ -export interface RerunTriggerResource extends SubResource { - /** - * Properties of the rerun trigger. - */ - properties: RerunTumblingWindowTrigger; -} - /** * Metadata associated with the operation. */ @@ -2031,461 +1987,385 @@ export interface MappingDataFlow { } /** - * Trigger that allows the referenced pipeline to depend on other pipeline runs based on - * runDimension Name/Value pairs. Upstream pipelines should declare the same runDimension Name and - * their runs should have the values for those runDimensions. The referenced pipeline run would be - * triggered if the values for the runDimension match for all upstream pipeline runs. + * Snowflake linked service. */ -export interface ChainingTrigger { +export interface SnowflakeLinkedService { /** * Polymorphic Discriminator */ - type: "ChainingTrigger"; + type: "Snowflake"; /** - * Trigger description. + * The integration runtime reference. + */ + connectVia?: IntegrationRuntimeReference; + /** + * Linked service description. */ description?: string; /** - * Indicates if trigger is running or not. Updated when Start/Stop APIs are called on the - * Trigger. Possible values include: 'Started', 'Stopped', 'Disabled' - * **NOTE: This property will not be serialized. It can only be populated by the server.** + * Parameters for linked service. */ - readonly runtimeState?: TriggerRuntimeState; + parameters?: { [propertyName: string]: ParameterSpecification }; /** - * List of tags that can be used for describing the trigger. + * List of tags that can be used for describing the linked service. */ annotations?: any[]; /** - * Pipeline for which runs are created when all upstream pipelines complete successfully. + * The connection string of snowflake. Type: string, SecureString. */ - pipelineProperty: TriggerPipelineReference; + connectionString: any; /** - * Upstream Pipelines. + * The Azure key vault secret reference of password in connection string. */ - dependsOn: PipelineReference[]; + password?: AzureKeyVaultSecretReference; /** - * Run Dimension property that needs to be emitted by upstream pipelines. + * The encrypted credential used for authentication. Credentials are encrypted using the + * integration runtime credential manager. Type: string (or Expression with resultType string). */ - runDimension: string; + encryptedCredential?: any; } /** - * Contains the possible cases for DependencyReference. - */ -export type DependencyReferenceUnion = DependencyReference | SelfDependencyTumblingWindowTriggerReference | TriggerDependencyReferenceUnion; - -/** - * Referenced dependency. + * Azure Function linked service. */ -export interface DependencyReference { +export interface AzureFunctionLinkedService { /** * Polymorphic Discriminator */ - type: "DependencyReference"; -} - -/** - * Self referenced tumbling window trigger dependency. - */ -export interface SelfDependencyTumblingWindowTriggerReference { + type: "AzureFunction"; /** - * Polymorphic Discriminator + * The integration runtime reference. */ - type: "SelfDependencyTumblingWindowTriggerReference"; + connectVia?: IntegrationRuntimeReference; /** - * Timespan applied to the start time of a tumbling window when evaluating dependency. + * Linked service description. */ - offset: string; + description?: string; /** - * The size of the window when evaluating the dependency. If undefined the frequency of the - * tumbling window will be used. + * Parameters for linked service. */ - size?: string; -} - -/** - * Trigger reference type. - */ -export interface TriggerReference { + parameters?: { [propertyName: string]: ParameterSpecification }; /** - * Reference trigger name. + * List of tags that can be used for describing the linked service. */ - referenceName: string; -} - -/** - * Contains the possible cases for TriggerDependencyReference. - */ -export type TriggerDependencyReferenceUnion = TriggerDependencyReference | TumblingWindowTriggerDependencyReference; - -/** - * Trigger referenced dependency. - */ -export interface TriggerDependencyReference { + annotations?: any[]; /** - * Polymorphic Discriminator + * The endpoint of the Azure Function App. URL will be in the format + * https://.azurewebsites.net. */ - type: "TriggerDependencyReference"; + functionAppUrl: any; /** - * Referenced trigger. + * Function or Host key for Azure Function App. */ - referenceTrigger: TriggerReference; + functionKey?: SecretBaseUnion; + /** + * The encrypted credential used for authentication. Credentials are encrypted using the + * integration runtime credential manager. Type: string (or Expression with resultType string). + */ + encryptedCredential?: any; } /** - * Referenced tumbling window trigger dependency. + * Azure Data Explorer (Kusto) linked service. */ -export interface TumblingWindowTriggerDependencyReference { +export interface AzureDataExplorerLinkedService { /** * Polymorphic Discriminator */ - type: "TumblingWindowTriggerDependencyReference"; + type: "AzureDataExplorer"; /** - * Referenced trigger. + * The integration runtime reference. */ - referenceTrigger: TriggerReference; + connectVia?: IntegrationRuntimeReference; /** - * Timespan applied to the start time of a tumbling window when evaluating dependency. + * Linked service description. */ - offset?: string; + description?: string; /** - * The size of the window when evaluating the dependency. If undefined the frequency of the - * tumbling window will be used. + * Parameters for linked service. */ - size?: string; -} - -/** - * Execution policy for an activity. - */ -export interface RetryPolicy { + parameters?: { [propertyName: string]: ParameterSpecification }; /** - * Maximum ordinary retry attempts. Default is 0. Type: integer (or Expression with resultType - * integer), minimum: 0. + * List of tags that can be used for describing the linked service. */ - count?: any; + annotations?: any[]; /** - * Interval between retries in seconds. Default is 30. + * The endpoint of Azure Data Explorer (the engine's endpoint). URL will be in the format + * https://..kusto.windows.net. Type: string (or Expression with + * resultType string) */ - intervalInSeconds?: number; + endpoint: any; + /** + * The ID of the service principal used to authenticate against Azure Data Explorer. Type: string + * (or Expression with resultType string). + */ + servicePrincipalId: any; + /** + * The key of the service principal used to authenticate against Kusto. + */ + servicePrincipalKey: SecretBaseUnion; + /** + * Database name for connection. Type: string (or Expression with resultType string). + */ + database: any; + /** + * The name or ID of the tenant to which the service principal belongs. Type: string (or + * Expression with resultType string). + */ + tenant: any; } /** - * Trigger that schedules pipeline runs for all fixed time interval windows from a start time - * without gaps and also supports backfill scenarios (when start time is in the past). + * SAP Table Linked Service. */ -export interface TumblingWindowTrigger { +export interface SapTableLinkedService { /** * Polymorphic Discriminator */ - type: "TumblingWindowTrigger"; + type: "SapTable"; /** - * Trigger description. + * The integration runtime reference. */ - description?: string; + connectVia?: IntegrationRuntimeReference; /** - * Indicates if trigger is running or not. Updated when Start/Stop APIs are called on the - * Trigger. Possible values include: 'Started', 'Stopped', 'Disabled' - * **NOTE: This property will not be serialized. It can only be populated by the server.** + * Linked service description. */ - readonly runtimeState?: TriggerRuntimeState; + description?: string; /** - * List of tags that can be used for describing the trigger. + * Parameters for linked service. */ - annotations?: any[]; - /** - * Pipeline for which runs are created when an event is fired for trigger window that is ready. - */ - pipelineProperty: TriggerPipelineReference; - /** - * The frequency of the time windows. Possible values include: 'Minute', 'Hour' - */ - frequency: TumblingWindowFrequency; - /** - * The interval of the time windows. The minimum interval allowed is 15 Minutes. - */ - interval: number; - /** - * The start time for the time period for the trigger during which events are fired for windows - * that are ready. Only UTC time is currently supported. - */ - startTime: Date; - /** - * The end time for the time period for the trigger during which events are fired for windows - * that are ready. Only UTC time is currently supported. - */ - endTime?: Date; - /** - * Specifies how long the trigger waits past due time before triggering new run. It doesn't alter - * window start and end time. The default is 0. Type: string (or Expression with resultType - * string), pattern: ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). - */ - delay?: any; - /** - * The max number of parallel time windows (ready for execution) for which a new run is - * triggered. - */ - maxConcurrency: number; + parameters?: { [propertyName: string]: ParameterSpecification }; /** - * Retry policy that will be applied for failed pipeline runs. + * List of tags that can be used for describing the linked service. */ - retryPolicy?: RetryPolicy; + annotations?: any[]; /** - * Triggers that this trigger depends on. Only tumbling window triggers are supported. + * Host name of the SAP instance where the table is located. Type: string (or Expression with + * resultType string). */ - dependsOn?: DependencyReferenceUnion[]; -} - -/** - * Contains the possible cases for MultiplePipelineTrigger. - */ -export type MultiplePipelineTriggerUnion = MultiplePipelineTrigger | BlobEventsTrigger | BlobTrigger | ScheduleTrigger; - -/** - * Base class for all triggers that support one to many model for trigger to pipeline. - */ -export interface MultiplePipelineTrigger { + server?: any; /** - * Polymorphic Discriminator + * System number of the SAP system where the table is located. (Usually a two-digit decimal + * number represented as a string.) Type: string (or Expression with resultType string). */ - type: "MultiplePipelineTrigger"; + systemNumber?: any; /** - * Trigger description. + * Client ID of the client on the SAP system where the table is located. (Usually a three-digit + * decimal number represented as a string) Type: string (or Expression with resultType string). */ - description?: string; + clientId?: any; /** - * Indicates if trigger is running or not. Updated when Start/Stop APIs are called on the - * Trigger. Possible values include: 'Started', 'Stopped', 'Disabled' - * **NOTE: This property will not be serialized. It can only be populated by the server.** + * Language of the SAP system where the table is located. The default value is EN. Type: string + * (or Expression with resultType string). */ - readonly runtimeState?: TriggerRuntimeState; + language?: any; /** - * List of tags that can be used for describing the trigger. + * SystemID of the SAP system where the table is located. Type: string (or Expression with + * resultType string). */ - annotations?: any[]; + systemId?: any; /** - * Pipelines that need to be started. + * Username to access the SAP server where the table is located. Type: string (or Expression with + * resultType string). */ - pipelines?: TriggerPipelineReference[]; -} - -/** - * Trigger that runs every time a Blob event occurs. - */ -export interface BlobEventsTrigger { + userName?: any; /** - * Polymorphic Discriminator + * Password to access the SAP server where the table is located. */ - type: "BlobEventsTrigger"; + password?: SecretBaseUnion; /** - * Trigger description. + * The hostname of the SAP Message Server. Type: string (or Expression with resultType string). */ - description?: string; + messageServer?: any; /** - * Indicates if trigger is running or not. Updated when Start/Stop APIs are called on the - * Trigger. Possible values include: 'Started', 'Stopped', 'Disabled' - * **NOTE: This property will not be serialized. It can only be populated by the server.** + * The service name or port number of the Message Server. Type: string (or Expression with + * resultType string). */ - readonly runtimeState?: TriggerRuntimeState; + messageServerService?: any; /** - * List of tags that can be used for describing the trigger. + * SNC activation indicator to access the SAP server where the table is located. Must be either 0 + * (off) or 1 (on). Type: string (or Expression with resultType string). */ - annotations?: any[]; + sncMode?: any; /** - * Pipelines that need to be started. + * Initiator's SNC name to access the SAP server where the table is located. Type: string (or + * Expression with resultType string). */ - pipelines?: TriggerPipelineReference[]; + sncMyName?: any; /** - * The blob path must begin with the pattern provided for trigger to fire. For example, - * '/records/blobs/december/' will only fire the trigger for blobs in the december folder under - * the records container. At least one of these must be provided: blobPathBeginsWith, - * blobPathEndsWith. + * Communication partner's SNC name to access the SAP server where the table is located. Type: + * string (or Expression with resultType string). */ - blobPathBeginsWith?: string; + sncPartnerName?: any; /** - * The blob path must end with the pattern provided for trigger to fire. For example, - * 'december/boxes.csv' will only fire the trigger for blobs named boxes in a december folder. At - * least one of these must be provided: blobPathBeginsWith, blobPathEndsWith. + * External security product's library to access the SAP server where the table is located. Type: + * string (or Expression with resultType string). */ - blobPathEndsWith?: string; + sncLibraryPath?: any; /** - * If set to true, blobs with zero bytes will be ignored. + * SNC Quality of Protection. Allowed value include: 1, 2, 3, 8, 9. Type: string (or Expression + * with resultType string). */ - ignoreEmptyBlobs?: boolean; + sncQop?: any; /** - * The type of events that cause this trigger to fire. + * The Logon Group for the SAP System. Type: string (or Expression with resultType string). */ - events: BlobEventTypes[]; + logonGroup?: any; /** - * The ARM resource ID of the Storage Account. + * The encrypted credential used for authentication. Credentials are encrypted using the + * integration runtime credential manager. Type: string (or Expression with resultType string). */ - scope: string; + encryptedCredential?: any; } /** - * Trigger that runs every time the selected Blob container changes. + * Google AdWords service linked service. */ -export interface BlobTrigger { +export interface GoogleAdWordsLinkedService { /** * Polymorphic Discriminator */ - type: "BlobTrigger"; - /** - * Trigger description. - */ - description?: string; + type: "GoogleAdWords"; /** - * Indicates if trigger is running or not. Updated when Start/Stop APIs are called on the - * Trigger. Possible values include: 'Started', 'Stopped', 'Disabled' - * **NOTE: This property will not be serialized. It can only be populated by the server.** + * The integration runtime reference. */ - readonly runtimeState?: TriggerRuntimeState; + connectVia?: IntegrationRuntimeReference; /** - * List of tags that can be used for describing the trigger. + * Linked service description. */ - annotations?: any[]; + description?: string; /** - * Pipelines that need to be started. + * Parameters for linked service. */ - pipelines?: TriggerPipelineReference[]; + parameters?: { [propertyName: string]: ParameterSpecification }; /** - * The path of the container/folder that will trigger the pipeline. + * List of tags that can be used for describing the linked service. */ - folderPath: string; + annotations?: any[]; /** - * The max number of parallel files to handle when it is triggered. + * The Client customer ID of the AdWords account that you want to fetch report data for. */ - maxConcurrency: number; + clientCustomerID: any; /** - * The Azure Storage linked service reference. + * The developer token associated with the manager account that you use to grant access to the + * AdWords API. */ - linkedService: LinkedServiceReference; -} - -/** - * The recurrence schedule occurrence. - */ -export interface RecurrenceScheduleOccurrence { + developerToken: SecretBaseUnion; /** - * The day of the week. Possible values include: 'Sunday', 'Monday', 'Tuesday', 'Wednesday', - * 'Thursday', 'Friday', 'Saturday' + * The OAuth 2.0 authentication mechanism used for authentication. ServiceAuthentication can only + * be used on self-hosted IR. Possible values include: 'ServiceAuthentication', + * 'UserAuthentication' */ - day?: DayOfWeek; + authenticationType: GoogleAdWordsAuthenticationType; /** - * The occurrence. + * The refresh token obtained from Google for authorizing access to AdWords for + * UserAuthentication. */ - occurrence?: number; + refreshToken?: SecretBaseUnion; /** - * Describes unknown properties. The value of an unknown property can be of "any" type. + * The client id of the google application used to acquire the refresh token. Type: string (or + * Expression with resultType string). */ - [property: string]: any; -} - -/** - * The recurrence schedule. - */ -export interface RecurrenceSchedule { + clientId?: any; /** - * The minutes. + * The client secret of the google application used to acquire the refresh token. */ - minutes?: number[]; + clientSecret?: SecretBaseUnion; /** - * The hours. + * The service account email ID that is used for ServiceAuthentication and can only be used on + * self-hosted IR. */ - hours?: number[]; + email?: any; /** - * The days of the week. + * The full path to the .p12 key file that is used to authenticate the service account email + * address and can only be used on self-hosted IR. */ - weekDays?: DaysOfWeek[]; + keyFilePath?: any; /** - * The month days. + * The full path of the .pem file containing trusted CA certificates for verifying the server + * when connecting over SSL. This property can only be set when using SSL on self-hosted IR. The + * default value is the cacerts.pem file installed with the IR. */ - monthDays?: number[]; + trustedCertPath?: any; /** - * The monthly occurrences. + * Specifies whether to use a CA certificate from the system trust store or from a specified PEM + * file. The default value is false. */ - monthlyOccurrences?: RecurrenceScheduleOccurrence[]; + useSystemTrustStore?: any; /** - * Describes unknown properties. The value of an unknown property can be of "any" type. + * The encrypted credential used for authentication. Credentials are encrypted using the + * integration runtime credential manager. Type: string (or Expression with resultType string). */ - [property: string]: any; + encryptedCredential?: any; } /** - * The workflow trigger recurrence. + * Oracle Service Cloud linked service. */ -export interface ScheduleTriggerRecurrence { - /** - * The frequency. Possible values include: 'NotSpecified', 'Minute', 'Hour', 'Day', 'Week', - * 'Month', 'Year' - */ - frequency?: RecurrenceFrequency; +export interface OracleServiceCloudLinkedService { /** - * The interval. + * Polymorphic Discriminator */ - interval?: number; + type: "OracleServiceCloud"; /** - * The start time. + * The integration runtime reference. */ - startTime?: Date; + connectVia?: IntegrationRuntimeReference; /** - * The end time. + * Linked service description. */ - endTime?: Date; + description?: string; /** - * The time zone. + * Parameters for linked service. */ - timeZone?: string; + parameters?: { [propertyName: string]: ParameterSpecification }; /** - * The recurrence schedule. + * List of tags that can be used for describing the linked service. */ - schedule?: RecurrenceSchedule; + annotations?: any[]; /** - * Describes unknown properties. The value of an unknown property can be of "any" type. + * The URL of the Oracle Service Cloud instance. */ - [property: string]: any; -} - -/** - * Trigger that creates pipeline runs periodically, on schedule. - */ -export interface ScheduleTrigger { + host: any; /** - * Polymorphic Discriminator + * The user name that you use to access Oracle Service Cloud server. */ - type: "ScheduleTrigger"; + username: any; /** - * Trigger description. + * The password corresponding to the user name that you provided in the username key. */ - description?: string; + password: SecretBaseUnion; /** - * Indicates if trigger is running or not. Updated when Start/Stop APIs are called on the - * Trigger. Possible values include: 'Started', 'Stopped', 'Disabled' - * **NOTE: This property will not be serialized. It can only be populated by the server.** + * Specifies whether the data source endpoints are encrypted using HTTPS. The default value is + * true. Type: boolean (or Expression with resultType boolean). */ - readonly runtimeState?: TriggerRuntimeState; + useEncryptedEndpoints?: any; /** - * List of tags that can be used for describing the trigger. + * Specifies whether to require the host name in the server's certificate to match the host name + * of the server when connecting over SSL. The default value is true. Type: boolean (or + * Expression with resultType boolean). */ - annotations?: any[]; + useHostVerification?: any; /** - * Pipelines that need to be started. + * Specifies whether to verify the identity of the server when connecting over SSL. The default + * value is true. Type: boolean (or Expression with resultType boolean). */ - pipelines?: TriggerPipelineReference[]; + usePeerVerification?: any; /** - * Recurrence schedule configuration. + * The encrypted credential used for authentication. Credentials are encrypted using the + * integration runtime credential manager. Type: string (or Expression with resultType string). */ - recurrence: ScheduleTriggerRecurrence; + encryptedCredential?: any; } /** - * Azure Function linked service. + * Dynamics AX linked service. */ -export interface AzureFunctionLinkedService { +export interface DynamicsAXLinkedService { /** * Polymorphic Discriminator */ - type: "AzureFunction"; + type: "DynamicsAX"; /** * The integration runtime reference. */ @@ -2503,14 +2383,30 @@ export interface AzureFunctionLinkedService { */ annotations?: any[]; /** - * The endpoint of the Azure Function App. URL will be in the format - * https://.azurewebsites.net. + * The Dynamics AX (or Dynamics 365 Finance and Operations) instance OData endpoint. */ - functionAppUrl: any; + url: any; /** - * Function or Host key for Azure Function App. + * Specify the application's client ID. Type: string (or Expression with resultType string). */ - functionKey?: SecretBaseUnion; + servicePrincipalId: any; + /** + * Specify the application's key. Mark this field as a SecureString to store it securely in Data + * Factory, or reference a secret stored in Azure Key Vault. Type: string (or Expression with + * resultType string). + */ + servicePrincipalKey: SecretBaseUnion; + /** + * Specify the tenant information (domain name or tenant ID) under which your application + * resides. Retrieve it by hovering the mouse in the top-right corner of the Azure portal. Type: + * string (or Expression with resultType string). + */ + tenant: any; + /** + * Specify the resource you are requesting authorization. Type: string (or Expression with + * resultType string). + */ + aadResourceId: any; /** * The encrypted credential used for authentication. Credentials are encrypted using the * integration runtime credential manager. Type: string (or Expression with resultType string). @@ -2519,13 +2415,13 @@ export interface AzureFunctionLinkedService { } /** - * Azure Data Explorer (Kusto) linked service. + * Responsys linked service. */ -export interface AzureDataExplorerLinkedService { +export interface ResponsysLinkedService { /** * Polymorphic Discriminator */ - type: "AzureDataExplorer"; + type: "Responsys"; /** * The integration runtime reference. */ @@ -2543,39 +2439,50 @@ export interface AzureDataExplorerLinkedService { */ annotations?: any[]; /** - * The endpoint of Azure Data Explorer (the engine's endpoint). URL will be in the format - * https://..kusto.windows.net. Type: string (or Expression with - * resultType string) + * The endpoint of the Responsys server. */ endpoint: any; /** - * The ID of the service principal used to authenticate against Azure Data Explorer. Type: string - * (or Expression with resultType string). + * The client ID associated with the Responsys application. Type: string (or Expression with + * resultType string). */ - servicePrincipalId: any; + clientId: any; /** - * The key of the service principal used to authenticate against Kusto. + * The client secret associated with the Responsys application. Type: string (or Expression with + * resultType string). */ - servicePrincipalKey: SecretBaseUnion; + clientSecret?: SecretBaseUnion; + /** + * Specifies whether the data source endpoints are encrypted using HTTPS. The default value is + * true. Type: boolean (or Expression with resultType boolean). + */ + useEncryptedEndpoints?: any; + /** + * Specifies whether to require the host name in the server's certificate to match the host name + * of the server when connecting over SSL. The default value is true. Type: boolean (or + * Expression with resultType boolean). + */ + useHostVerification?: any; /** - * Database name for connection. Type: string (or Expression with resultType string). + * Specifies whether to verify the identity of the server when connecting over SSL. The default + * value is true. Type: boolean (or Expression with resultType boolean). */ - database: any; + usePeerVerification?: any; /** - * The name or ID of the tenant to which the service principal belongs. Type: string (or - * Expression with resultType string). + * The encrypted credential used for authentication. Credentials are encrypted using the + * integration runtime credential manager. Type: string (or Expression with resultType string). */ - tenant: any; + encryptedCredential?: any; } /** - * SAP Table Linked Service. + * Azure Databricks linked service. */ -export interface SapTableLinkedService { +export interface AzureDatabricksLinkedService { /** * Polymorphic Discriminator */ - type: "SapTable"; + type: "AzureDatabricks"; /** * The integration runtime reference. */ @@ -2593,77 +2500,76 @@ export interface SapTableLinkedService { */ annotations?: any[]; /** - * Host name of the SAP instance where the table is located. Type: string (or Expression with - * resultType string). - */ - server?: any; - /** - * System number of the SAP system where the table is located. (Usually a two-digit decimal - * number represented as a string.) Type: string (or Expression with resultType string). - */ - systemNumber?: any; - /** - * Client ID of the client on the SAP system where the table is located. (Usually a three-digit - * decimal number represented as a string) Type: string (or Expression with resultType string). + * .azuredatabricks.net, domain name of your Databricks deployment. Type: string (or + * Expression with resultType string). */ - clientId?: any; + domain: any; /** - * Language of the SAP system where the table is located. The default value is EN. Type: string - * (or Expression with resultType string). + * Access token for databricks REST API. Refer to + * https://docs.azuredatabricks.net/api/latest/authentication.html. Type: string (or Expression + * with resultType string). */ - language?: any; + accessToken: SecretBaseUnion; /** - * SystemID of the SAP system where the table is located. Type: string (or Expression with - * resultType string). + * The id of an existing interactive cluster that will be used for all runs of this activity. + * Type: string (or Expression with resultType string). */ - systemId?: any; + existingClusterId?: any; /** - * Username to access the SAP server where the table is located. Type: string (or Expression with - * resultType string). + * The id of an existing instance pool that will be used for all runs of this activity. Type: + * string (or Expression with resultType string). */ - userName?: any; + instancePoolId?: any; /** - * Password to access the SAP server where the table is located. + * If not using an existing interactive cluster, this specifies the Spark version of a new job + * cluster or instance pool nodes created for each run of this activity. Required if + * instancePoolId is specified. Type: string (or Expression with resultType string). */ - password?: SecretBaseUnion; + newClusterVersion?: any; /** - * The hostname of the SAP Message Server. Type: string (or Expression with resultType string). + * If not using an existing interactive cluster, this specifies the number of worker nodes to use + * for the new job cluster or instance pool. For new job clusters, this a string-formatted Int32, + * like '1' means numOfWorker is 1 or '1:10' means auto-scale from 1 (min) to 10 (max). For + * instance pools, this is a string-formatted Int32, and can only specify a fixed number of + * worker nodes, such as '2'. Required if newClusterVersion is specified. Type: string (or + * Expression with resultType string). */ - messageServer?: any; + newClusterNumOfWorker?: any; /** - * The service name or port number of the Message Server. Type: string (or Expression with - * resultType string). + * The node type of the new job cluster. This property is required if newClusterVersion is + * specified and instancePoolId is not specified. If instancePoolId is specified, this property + * is ignored. Type: string (or Expression with resultType string). */ - messageServerService?: any; + newClusterNodeType?: any; /** - * SNC activation indicator to access the SAP server where the table is located. Must be either 0 - * (off) or 1 (on). Type: string (or Expression with resultType string). + * A set of optional, user-specified Spark configuration key-value pairs. */ - sncMode?: any; + newClusterSparkConf?: { [propertyName: string]: any }; /** - * Initiator's SNC name to access the SAP server where the table is located. Type: string (or - * Expression with resultType string). + * A set of optional, user-specified Spark environment variables key-value pairs. */ - sncMyName?: any; + newClusterSparkEnvVars?: { [propertyName: string]: any }; /** - * Communication partner's SNC name to access the SAP server where the table is located. Type: - * string (or Expression with resultType string). + * Additional tags for cluster resources. This property is ignored in instance pool + * configurations. */ - sncPartnerName?: any; + newClusterCustomTags?: { [propertyName: string]: any }; /** - * External security product's library to access the SAP server where the table is located. Type: - * string (or Expression with resultType string). + * The driver node type for the new job cluster. This property is ignored in instance pool + * configurations. Type: string (or Expression with resultType string). */ - sncLibraryPath?: any; + newClusterDriverNodeType?: any; /** - * SNC Quality of Protection. Allowed value include: 1, 2, 3, 8, 9. Type: string (or Expression - * with resultType string). + * User-defined initialization scripts for the new cluster. Type: array of strings (or Expression + * with resultType array of strings). */ - sncQop?: any; + newClusterInitScripts?: any; /** - * The Logon Group for the SAP System. Type: string (or Expression with resultType string). + * Enable the elastic disk on the new cluster. This property is now ignored, and takes the + * default elastic disk behavior in Databricks (elastic disks are always enabled). Type: boolean + * (or Expression with resultType boolean). */ - logonGroup?: any; + newClusterEnableElasticDisk?: any; /** * The encrypted credential used for authentication. Credentials are encrypted using the * integration runtime credential manager. Type: string (or Expression with resultType string). @@ -2672,13 +2578,13 @@ export interface SapTableLinkedService { } /** - * Google AdWords service linked service. + * Azure Data Lake Analytics linked service. */ -export interface GoogleAdWordsLinkedService { +export interface AzureDataLakeAnalyticsLinkedService { /** * Polymorphic Discriminator */ - type: "GoogleAdWords"; + type: "AzureDataLakeAnalytics"; /** * The integration runtime reference. */ @@ -2696,70 +2602,75 @@ export interface GoogleAdWordsLinkedService { */ annotations?: any[]; /** - * The Client customer ID of the AdWords account that you want to fetch report data for. + * The Azure Data Lake Analytics account name. Type: string (or Expression with resultType + * string). */ - clientCustomerID: any; + accountName: any; /** - * The developer token associated with the manager account that you use to grant access to the - * AdWords API. + * The ID of the application used to authenticate against the Azure Data Lake Analytics account. + * Type: string (or Expression with resultType string). */ - developerToken: SecretBaseUnion; + servicePrincipalId?: any; /** - * The OAuth 2.0 authentication mechanism used for authentication. ServiceAuthentication can only - * be used on self-hosted IR. Possible values include: 'ServiceAuthentication', - * 'UserAuthentication' + * The Key of the application used to authenticate against the Azure Data Lake Analytics account. */ - authenticationType: GoogleAdWordsAuthenticationType; + servicePrincipalKey?: SecretBaseUnion; /** - * The refresh token obtained from Google for authorizing access to AdWords for - * UserAuthentication. + * The name or ID of the tenant to which the service principal belongs. Type: string (or + * Expression with resultType string). */ - refreshToken?: SecretBaseUnion; + tenant: any; /** - * The client id of the google application used to acquire the refresh token. Type: string (or - * Expression with resultType string). + * Data Lake Analytics account subscription ID (if different from Data Factory account). Type: + * string (or Expression with resultType string). */ - clientId?: any; + subscriptionId?: any; /** - * The client secret of the google application used to acquire the refresh token. + * Data Lake Analytics account resource group name (if different from Data Factory account). + * Type: string (or Expression with resultType string). */ - clientSecret?: SecretBaseUnion; + resourceGroupName?: any; /** - * The service account email ID that is used for ServiceAuthentication and can only be used on - * self-hosted IR. + * Azure Data Lake Analytics URI Type: string (or Expression with resultType string). */ - email?: any; + dataLakeAnalyticsUri?: any; /** - * The full path to the .p12 key file that is used to authenticate the service account email - * address and can only be used on self-hosted IR. + * The encrypted credential used for authentication. Credentials are encrypted using the + * integration runtime credential manager. Type: string (or Expression with resultType string). */ - keyFilePath?: any; + encryptedCredential?: any; +} + +/** + * Custom script action to run on HDI ondemand cluster once it's up. + */ +export interface ScriptAction { /** - * The full path of the .pem file containing trusted CA certificates for verifying the server - * when connecting over SSL. This property can only be set when using SSL on self-hosted IR. The - * default value is the cacerts.pem file installed with the IR. + * The user provided name of the script action. */ - trustedCertPath?: any; + name: string; /** - * Specifies whether to use a CA certificate from the system trust store or from a specified PEM - * file. The default value is false. + * The URI for the script action. */ - useSystemTrustStore?: any; + uri: string; /** - * The encrypted credential used for authentication. Credentials are encrypted using the - * integration runtime credential manager. Type: string (or Expression with resultType string). + * The node types on which the script action should be executed. */ - encryptedCredential?: any; + roles: any; + /** + * The parameters for the script action. + */ + parameters?: string; } /** - * Oracle Service Cloud linked service. + * HDInsight ondemand linked service. */ -export interface OracleServiceCloudLinkedService { +export interface HDInsightOnDemandLinkedService { /** * Polymorphic Discriminator */ - type: "OracleServiceCloud"; + type: "HDInsightOnDemand"; /** * The integration runtime reference. */ @@ -2777,165 +2688,166 @@ export interface OracleServiceCloudLinkedService { */ annotations?: any[]; /** - * The URL of the Oracle Service Cloud instance. + * Number of worker/data nodes in the cluster. Suggestion value: 4. Type: string (or Expression + * with resultType string). */ - host: any; + clusterSize: any; /** - * The user name that you use to access Oracle Service Cloud server. + * The allowed idle time for the on-demand HDInsight cluster. Specifies how long the on-demand + * HDInsight cluster stays alive after completion of an activity run if there are no other active + * jobs in the cluster. The minimum value is 5 mins. Type: string (or Expression with resultType + * string). */ - username: any; + timeToLive: any; /** - * The password corresponding to the user name that you provided in the username key. + * Version of the HDInsight cluster.  Type: string (or Expression with resultType string). */ - password: SecretBaseUnion; + version: any; /** - * Specifies whether the data source endpoints are encrypted using HTTPS. The default value is - * true. Type: boolean (or Expression with resultType boolean). + * Azure Storage linked service to be used by the on-demand cluster for storing and processing + * data. */ - useEncryptedEndpoints?: any; + linkedServiceName: LinkedServiceReference; /** - * Specifies whether to require the host name in the server's certificate to match the host name - * of the server when connecting over SSL. The default value is true. Type: boolean (or - * Expression with resultType boolean). + * The customer’s subscription to host the cluster. Type: string (or Expression with resultType + * string). */ - useHostVerification?: any; + hostSubscriptionId: any; /** - * Specifies whether to verify the identity of the server when connecting over SSL. The default - * value is true. Type: boolean (or Expression with resultType boolean). + * The service principal id for the hostSubscriptionId. Type: string (or Expression with + * resultType string). */ - usePeerVerification?: any; + servicePrincipalId?: any; /** - * The encrypted credential used for authentication. Credentials are encrypted using the - * integration runtime credential manager. Type: string (or Expression with resultType string). + * The key for the service principal id. */ - encryptedCredential?: any; -} - -/** - * Dynamics AX linked service. - */ -export interface DynamicsAXLinkedService { + servicePrincipalKey?: SecretBaseUnion; /** - * Polymorphic Discriminator + * The Tenant id/name to which the service principal belongs. Type: string (or Expression with + * resultType string). */ - type: "DynamicsAX"; + tenant: any; + /** + * The resource group where the cluster belongs. Type: string (or Expression with resultType + * string). + */ + clusterResourceGroup: any; + /** + * The prefix of cluster name, postfix will be distinct with timestamp. Type: string (or + * Expression with resultType string). + */ + clusterNamePrefix?: any; + /** + * The username to access the cluster. Type: string (or Expression with resultType string). + */ + clusterUserName?: any; /** - * The integration runtime reference. + * The password to access the cluster. */ - connectVia?: IntegrationRuntimeReference; + clusterPassword?: SecretBaseUnion; /** - * Linked service description. + * The username to SSH remotely connect to cluster’s node (for Linux). Type: string (or + * Expression with resultType string). */ - description?: string; + clusterSshUserName?: any; /** - * Parameters for linked service. + * The password to SSH remotely connect cluster’s node (for Linux). */ - parameters?: { [propertyName: string]: ParameterSpecification }; + clusterSshPassword?: SecretBaseUnion; /** - * List of tags that can be used for describing the linked service. + * Specifies additional storage accounts for the HDInsight linked service so that the Data + * Factory service can register them on your behalf. */ - annotations?: any[]; + additionalLinkedServiceNames?: LinkedServiceReference[]; /** - * The Dynamics AX (or Dynamics 365 Finance and Operations) instance OData endpoint. + * The name of Azure SQL linked service that point to the HCatalog database. The on-demand + * HDInsight cluster is created by using the Azure SQL database as the metastore. */ - url: any; + hcatalogLinkedServiceName?: LinkedServiceReference; /** - * Specify the application's client ID. Type: string (or Expression with resultType string). + * The cluster type. Type: string (or Expression with resultType string). */ - servicePrincipalId: any; + clusterType?: any; /** - * Specify the application's key. Mark this field as a SecureString to store it securely in Data - * Factory, or reference a secret stored in Azure Key Vault. Type: string (or Expression with + * The version of spark if the cluster type is 'spark'. Type: string (or Expression with * resultType string). */ - servicePrincipalKey: SecretBaseUnion; + sparkVersion?: any; /** - * Specify the tenant information (domain name or tenant ID) under which your application - * resides. Retrieve it by hovering the mouse in the top-right corner of the Azure portal. Type: - * string (or Expression with resultType string). + * Specifies the core configuration parameters (as in core-site.xml) for the HDInsight cluster to + * be created. */ - tenant: any; + coreConfiguration?: any; /** - * Specify the resource you are requesting authorization. Type: string (or Expression with - * resultType string). + * Specifies the HBase configuration parameters (hbase-site.xml) for the HDInsight cluster. */ - aadResourceId: any; + hBaseConfiguration?: any; /** - * The encrypted credential used for authentication. Credentials are encrypted using the - * integration runtime credential manager. Type: string (or Expression with resultType string). + * Specifies the HDFS configuration parameters (hdfs-site.xml) for the HDInsight cluster. */ - encryptedCredential?: any; -} - -/** - * Responsys linked service. - */ -export interface ResponsysLinkedService { + hdfsConfiguration?: any; /** - * Polymorphic Discriminator + * Specifies the hive configuration parameters (hive-site.xml) for the HDInsight cluster. */ - type: "Responsys"; + hiveConfiguration?: any; /** - * The integration runtime reference. + * Specifies the MapReduce configuration parameters (mapred-site.xml) for the HDInsight cluster. */ - connectVia?: IntegrationRuntimeReference; + mapReduceConfiguration?: any; /** - * Linked service description. + * Specifies the Oozie configuration parameters (oozie-site.xml) for the HDInsight cluster. */ - description?: string; + oozieConfiguration?: any; /** - * Parameters for linked service. + * Specifies the Storm configuration parameters (storm-site.xml) for the HDInsight cluster. */ - parameters?: { [propertyName: string]: ParameterSpecification }; + stormConfiguration?: any; /** - * List of tags that can be used for describing the linked service. + * Specifies the Yarn configuration parameters (yarn-site.xml) for the HDInsight cluster. */ - annotations?: any[]; + yarnConfiguration?: any; /** - * The endpoint of the Responsys server. + * The encrypted credential used for authentication. Credentials are encrypted using the + * integration runtime credential manager. Type: string (or Expression with resultType string). */ - endpoint: any; + encryptedCredential?: any; /** - * The client ID associated with the Responsys application. Type: string (or Expression with - * resultType string). + * Specifies the size of the head node for the HDInsight cluster. */ - clientId: any; + headNodeSize?: any; /** - * The client secret associated with the Responsys application. Type: string (or Expression with - * resultType string). + * Specifies the size of the data node for the HDInsight cluster. */ - clientSecret?: SecretBaseUnion; + dataNodeSize?: any; /** - * Specifies whether the data source endpoints are encrypted using HTTPS. The default value is - * true. Type: boolean (or Expression with resultType boolean). + * Specifies the size of the Zoo Keeper node for the HDInsight cluster. */ - useEncryptedEndpoints?: any; + zookeeperNodeSize?: any; /** - * Specifies whether to require the host name in the server's certificate to match the host name - * of the server when connecting over SSL. The default value is true. Type: boolean (or - * Expression with resultType boolean). + * Custom script actions to run on HDI ondemand cluster once it's up. Please refer to + * https://docs.microsoft.com/en-us/azure/hdinsight/hdinsight-hadoop-customize-cluster-linux?toc=%2Fen-us%2Fazure%2Fhdinsight%2Fr-server%2FTOC.json&bc=%2Fen-us%2Fazure%2Fbread%2Ftoc.json#understanding-script-actions. */ - useHostVerification?: any; + scriptActions?: ScriptAction[]; /** - * Specifies whether to verify the identity of the server when connecting over SSL. The default - * value is true. Type: boolean (or Expression with resultType boolean). + * The ARM resource ID for the vNet to which the cluster should be joined after creation. Type: + * string (or Expression with resultType string). */ - usePeerVerification?: any; + virtualNetworkId?: any; /** - * The encrypted credential used for authentication. Credentials are encrypted using the - * integration runtime credential manager. Type: string (or Expression with resultType string). + * The ARM resource ID for the subnet in the vNet. If virtualNetworkId was specified, then this + * property is required. Type: string (or Expression with resultType string). */ - encryptedCredential?: any; + subnetName?: any; } /** - * Azure Databricks linked service. + * Salesforce Marketing Cloud linked service. */ -export interface AzureDatabricksLinkedService { +export interface SalesforceMarketingCloudLinkedService { /** * Polymorphic Discriminator */ - type: "AzureDatabricks"; + type: "SalesforceMarketingCloud"; /** * The integration runtime reference. */ @@ -2953,76 +2865,31 @@ export interface AzureDatabricksLinkedService { */ annotations?: any[]; /** - * .azuredatabricks.net, domain name of your Databricks deployment. Type: string (or + * The client ID associated with the Salesforce Marketing Cloud application. Type: string (or * Expression with resultType string). */ - domain: any; - /** - * Access token for databricks REST API. Refer to - * https://docs.azuredatabricks.net/api/latest/authentication.html. Type: string (or Expression - * with resultType string). - */ - accessToken: SecretBaseUnion; - /** - * The id of an existing interactive cluster that will be used for all runs of this activity. - * Type: string (or Expression with resultType string). - */ - existingClusterId?: any; - /** - * The id of an existing instance pool that will be used for all runs of this activity. Type: - * string (or Expression with resultType string). - */ - instancePoolId?: any; - /** - * If not using an existing interactive cluster, this specifies the Spark version of a new job - * cluster or instance pool nodes created for each run of this activity. Required if - * instancePoolId is specified. Type: string (or Expression with resultType string). - */ - newClusterVersion?: any; + clientId: any; /** - * If not using an existing interactive cluster, this specifies the number of worker nodes to use - * for the new job cluster or instance pool. For new job clusters, this a string-formatted Int32, - * like '1' means numOfWorker is 1 or '1:10' means auto-scale from 1 (min) to 10 (max). For - * instance pools, this is a string-formatted Int32, and can only specify a fixed number of - * worker nodes, such as '2'. Required if newClusterVersion is specified. Type: string (or + * The client secret associated with the Salesforce Marketing Cloud application. Type: string (or * Expression with resultType string). */ - newClusterNumOfWorker?: any; - /** - * The node type of the new job cluster. This property is required if newClusterVersion is - * specified and instancePoolId is not specified. If instancePoolId is specified, this property - * is ignored. Type: string (or Expression with resultType string). - */ - newClusterNodeType?: any; - /** - * A set of optional, user-specified Spark configuration key-value pairs. - */ - newClusterSparkConf?: { [propertyName: string]: any }; - /** - * A set of optional, user-specified Spark environment variables key-value pairs. - */ - newClusterSparkEnvVars?: { [propertyName: string]: any }; - /** - * Additional tags for cluster resources. This property is ignored in instance pool - * configurations. - */ - newClusterCustomTags?: { [propertyName: string]: any }; + clientSecret?: SecretBaseUnion; /** - * The driver node type for the new job cluster. This property is ignored in instance pool - * configurations. Type: string (or Expression with resultType string). + * Specifies whether the data source endpoints are encrypted using HTTPS. The default value is + * true. Type: boolean (or Expression with resultType boolean). */ - newClusterDriverNodeType?: any; + useEncryptedEndpoints?: any; /** - * User-defined initialization scripts for the new cluster. Type: array of strings (or Expression - * with resultType array of strings). + * Specifies whether to require the host name in the server's certificate to match the host name + * of the server when connecting over SSL. The default value is true. Type: boolean (or + * Expression with resultType boolean). */ - newClusterInitScripts?: any; + useHostVerification?: any; /** - * Enable the elastic disk on the new cluster. This property is now ignored, and takes the - * default elastic disk behavior in Databricks (elastic disks are always enabled). Type: boolean - * (or Expression with resultType boolean). + * Specifies whether to verify the identity of the server when connecting over SSL. The default + * value is true. Type: boolean (or Expression with resultType boolean). */ - newClusterEnableElasticDisk?: any; + usePeerVerification?: any; /** * The encrypted credential used for authentication. Credentials are encrypted using the * integration runtime credential manager. Type: string (or Expression with resultType string). @@ -3031,13 +2898,13 @@ export interface AzureDatabricksLinkedService { } /** - * Azure Data Lake Analytics linked service. + * Netezza linked service. */ -export interface AzureDataLakeAnalyticsLinkedService { +export interface NetezzaLinkedService { /** * Polymorphic Discriminator */ - type: "AzureDataLakeAnalytics"; + type: "Netezza"; /** * The integration runtime reference. */ @@ -3055,38 +2922,13 @@ export interface AzureDataLakeAnalyticsLinkedService { */ annotations?: any[]; /** - * The Azure Data Lake Analytics account name. Type: string (or Expression with resultType - * string). - */ - accountName: any; - /** - * The ID of the application used to authenticate against the Azure Data Lake Analytics account. - * Type: string (or Expression with resultType string). - */ - servicePrincipalId?: any; - /** - * The Key of the application used to authenticate against the Azure Data Lake Analytics account. - */ - servicePrincipalKey?: SecretBaseUnion; - /** - * The name or ID of the tenant to which the service principal belongs. Type: string (or - * Expression with resultType string). - */ - tenant: any; - /** - * Data Lake Analytics account subscription ID (if different from Data Factory account). Type: - * string (or Expression with resultType string). - */ - subscriptionId?: any; - /** - * Data Lake Analytics account resource group name (if different from Data Factory account). - * Type: string (or Expression with resultType string). + * An ODBC connection string. Type: string, SecureString or AzureKeyVaultSecretReference. */ - resourceGroupName?: any; + connectionString?: any; /** - * Azure Data Lake Analytics URI Type: string (or Expression with resultType string). + * The Azure key vault secret reference of password in connection string. */ - dataLakeAnalyticsUri?: any; + pwd?: AzureKeyVaultSecretReference; /** * The encrypted credential used for authentication. Credentials are encrypted using the * integration runtime credential manager. Type: string (or Expression with resultType string). @@ -3095,35 +2937,13 @@ export interface AzureDataLakeAnalyticsLinkedService { } /** - * Custom script action to run on HDI ondemand cluster once it's up. - */ -export interface ScriptAction { - /** - * The user provided name of the script action. - */ - name: string; - /** - * The URI for the script action. - */ - uri: string; - /** - * The node types on which the script action should be executed. - */ - roles: any; - /** - * The parameters for the script action. - */ - parameters?: string; -} - -/** - * HDInsight ondemand linked service. + * Vertica linked service. */ -export interface HDInsightOnDemandLinkedService { +export interface VerticaLinkedService { /** * Polymorphic Discriminator */ - type: "HDInsightOnDemand"; + type: "Vertica"; /** * The integration runtime reference. */ @@ -3141,166 +2961,142 @@ export interface HDInsightOnDemandLinkedService { */ annotations?: any[]; /** - * Number of worker/data nodes in the cluster. Suggestion value: 4. Type: string (or Expression - * with resultType string). - */ - clusterSize: any; - /** - * The allowed idle time for the on-demand HDInsight cluster. Specifies how long the on-demand - * HDInsight cluster stays alive after completion of an activity run if there are no other active - * jobs in the cluster. The minimum value is 5 mins. Type: string (or Expression with resultType - * string). + * An ODBC connection string. Type: string, SecureString or AzureKeyVaultSecretReference. */ - timeToLive: any; + connectionString?: any; /** - * Version of the HDInsight cluster.  Type: string (or Expression with resultType string). + * The Azure key vault secret reference of password in connection string. */ - version: any; + pwd?: AzureKeyVaultSecretReference; /** - * Azure Storage linked service to be used by the on-demand cluster for storing and processing - * data. + * The encrypted credential used for authentication. Credentials are encrypted using the + * integration runtime credential manager. Type: string (or Expression with resultType string). */ - linkedServiceName: LinkedServiceReference; + encryptedCredential?: any; +} + +/** + * Zoho server linked service. + */ +export interface ZohoLinkedService { /** - * The customer’s subscription to host the cluster. Type: string (or Expression with resultType - * string). + * Polymorphic Discriminator */ - hostSubscriptionId: any; + type: "Zoho"; /** - * The service principal id for the hostSubscriptionId. Type: string (or Expression with - * resultType string). + * The integration runtime reference. */ - servicePrincipalId?: any; + connectVia?: IntegrationRuntimeReference; /** - * The key for the service principal id. + * Linked service description. */ - servicePrincipalKey?: SecretBaseUnion; + description?: string; /** - * The Tenant id/name to which the service principal belongs. Type: string (or Expression with - * resultType string). + * Parameters for linked service. */ - tenant: any; + parameters?: { [propertyName: string]: ParameterSpecification }; /** - * The resource group where the cluster belongs. Type: string (or Expression with resultType - * string). + * List of tags that can be used for describing the linked service. */ - clusterResourceGroup: any; + annotations?: any[]; /** - * The prefix of cluster name, postfix will be distinct with timestamp. Type: string (or - * Expression with resultType string). + * The endpoint of the Zoho server. (i.e. crm.zoho.com/crm/private) */ - clusterNamePrefix?: any; + endpoint: any; /** - * The username to access the cluster. Type: string (or Expression with resultType string). + * The access token for Zoho authentication. */ - clusterUserName?: any; + accessToken?: SecretBaseUnion; /** - * The password to access the cluster. + * Specifies whether the data source endpoints are encrypted using HTTPS. The default value is + * true. */ - clusterPassword?: SecretBaseUnion; + useEncryptedEndpoints?: any; /** - * The username to SSH remotely connect to cluster’s node (for Linux). Type: string (or - * Expression with resultType string). + * Specifies whether to require the host name in the server's certificate to match the host name + * of the server when connecting over SSL. The default value is true. */ - clusterSshUserName?: any; + useHostVerification?: any; /** - * The password to SSH remotely connect cluster’s node (for Linux). + * Specifies whether to verify the identity of the server when connecting over SSL. The default + * value is true. */ - clusterSshPassword?: SecretBaseUnion; + usePeerVerification?: any; /** - * Specifies additional storage accounts for the HDInsight linked service so that the Data - * Factory service can register them on your behalf. + * The encrypted credential used for authentication. Credentials are encrypted using the + * integration runtime credential manager. Type: string (or Expression with resultType string). */ - additionalLinkedServiceNames?: LinkedServiceReference[]; + encryptedCredential?: any; +} + +/** + * Xero Service linked service. + */ +export interface XeroLinkedService { /** - * The name of Azure SQL linked service that point to the HCatalog database. The on-demand - * HDInsight cluster is created by using the Azure SQL database as the metastore. + * Polymorphic Discriminator */ - hcatalogLinkedServiceName?: LinkedServiceReference; + type: "Xero"; /** - * The cluster type. Type: string (or Expression with resultType string). + * The integration runtime reference. */ - clusterType?: any; + connectVia?: IntegrationRuntimeReference; /** - * The version of spark if the cluster type is 'spark'. Type: string (or Expression with - * resultType string). + * Linked service description. */ - sparkVersion?: any; + description?: string; /** - * Specifies the core configuration parameters (as in core-site.xml) for the HDInsight cluster to - * be created. + * Parameters for linked service. */ - coreConfiguration?: any; + parameters?: { [propertyName: string]: ParameterSpecification }; /** - * Specifies the HBase configuration parameters (hbase-site.xml) for the HDInsight cluster. + * List of tags that can be used for describing the linked service. */ - hBaseConfiguration?: any; + annotations?: any[]; /** - * Specifies the HDFS configuration parameters (hdfs-site.xml) for the HDInsight cluster. + * The endpoint of the Xero server. (i.e. api.xero.com) */ - hdfsConfiguration?: any; + host: any; /** - * Specifies the hive configuration parameters (hive-site.xml) for the HDInsight cluster. + * The consumer key associated with the Xero application. */ - hiveConfiguration?: any; + consumerKey?: SecretBaseUnion; /** - * Specifies the MapReduce configuration parameters (mapred-site.xml) for the HDInsight cluster. + * The private key from the .pem file that was generated for your Xero private application. You + * must include all the text from the .pem file, including the Unix line endings( + * ). */ - mapReduceConfiguration?: any; + privateKey?: SecretBaseUnion; /** - * Specifies the Oozie configuration parameters (oozie-site.xml) for the HDInsight cluster. + * Specifies whether the data source endpoints are encrypted using HTTPS. The default value is + * true. */ - oozieConfiguration?: any; + useEncryptedEndpoints?: any; /** - * Specifies the Storm configuration parameters (storm-site.xml) for the HDInsight cluster. + * Specifies whether to require the host name in the server's certificate to match the host name + * of the server when connecting over SSL. The default value is true. */ - stormConfiguration?: any; + useHostVerification?: any; /** - * Specifies the Yarn configuration parameters (yarn-site.xml) for the HDInsight cluster. + * Specifies whether to verify the identity of the server when connecting over SSL. The default + * value is true. */ - yarnConfiguration?: any; + usePeerVerification?: any; /** * The encrypted credential used for authentication. Credentials are encrypted using the * integration runtime credential manager. Type: string (or Expression with resultType string). */ encryptedCredential?: any; - /** - * Specifies the size of the head node for the HDInsight cluster. - */ - headNodeSize?: any; - /** - * Specifies the size of the data node for the HDInsight cluster. - */ - dataNodeSize?: any; - /** - * Specifies the size of the Zoo Keeper node for the HDInsight cluster. - */ - zookeeperNodeSize?: any; - /** - * Custom script actions to run on HDI ondemand cluster once it's up. Please refer to - * https://docs.microsoft.com/en-us/azure/hdinsight/hdinsight-hadoop-customize-cluster-linux?toc=%2Fen-us%2Fazure%2Fhdinsight%2Fr-server%2FTOC.json&bc=%2Fen-us%2Fazure%2Fbread%2Ftoc.json#understanding-script-actions. - */ - scriptActions?: ScriptAction[]; - /** - * The ARM resource ID for the vNet to which the cluster should be joined after creation. Type: - * string (or Expression with resultType string). - */ - virtualNetworkId?: any; - /** - * The ARM resource ID for the subnet in the vNet. If virtualNetworkId was specified, then this - * property is required. Type: string (or Expression with resultType string). - */ - subnetName?: any; } /** - * Salesforce Marketing Cloud linked service. + * Square Service linked service. */ -export interface SalesforceMarketingCloudLinkedService { +export interface SquareLinkedService { /** * Polymorphic Discriminator */ - type: "SalesforceMarketingCloud"; + type: "Square"; /** * The integration runtime reference. */ @@ -3318,29 +3114,34 @@ export interface SalesforceMarketingCloudLinkedService { */ annotations?: any[]; /** - * The client ID associated with the Salesforce Marketing Cloud application. Type: string (or - * Expression with resultType string). + * The URL of the Square instance. (i.e. mystore.mysquare.com) + */ + host: any; + /** + * The client ID associated with your Square application. */ clientId: any; /** - * The client secret associated with the Salesforce Marketing Cloud application. Type: string (or - * Expression with resultType string). + * The client secret associated with your Square application. */ clientSecret?: SecretBaseUnion; + /** + * The redirect URL assigned in the Square application dashboard. (i.e. http://localhost:2500) + */ + redirectUri: any; /** * Specifies whether the data source endpoints are encrypted using HTTPS. The default value is - * true. Type: boolean (or Expression with resultType boolean). + * true. */ useEncryptedEndpoints?: any; /** * Specifies whether to require the host name in the server's certificate to match the host name - * of the server when connecting over SSL. The default value is true. Type: boolean (or - * Expression with resultType boolean). + * of the server when connecting over SSL. The default value is true. */ useHostVerification?: any; /** * Specifies whether to verify the identity of the server when connecting over SSL. The default - * value is true. Type: boolean (or Expression with resultType boolean). + * value is true. */ usePeerVerification?: any; /** @@ -3351,13 +3152,13 @@ export interface SalesforceMarketingCloudLinkedService { } /** - * Netezza linked service. + * Spark Server linked service. */ -export interface NetezzaLinkedService { +export interface SparkLinkedService { /** * Polymorphic Discriminator */ - type: "Netezza"; + type: "Spark"; /** * The integration runtime reference. */ @@ -3373,15 +3174,68 @@ export interface NetezzaLinkedService { /** * List of tags that can be used for describing the linked service. */ - annotations?: any[]; + annotations?: any[]; + /** + * IP address or host name of the Spark server + */ + host: any; + /** + * The TCP port that the Spark server uses to listen for client connections. + */ + port: any; + /** + * The type of Spark server. Possible values include: 'SharkServer', 'SharkServer2', + * 'SparkThriftServer' + */ + serverType?: SparkServerType; + /** + * The transport protocol to use in the Thrift layer. Possible values include: 'Binary', 'SASL', + * 'HTTP ' + */ + thriftTransportProtocol?: SparkThriftTransportProtocol; + /** + * The authentication method used to access the Spark server. Possible values include: + * 'Anonymous', 'Username', 'UsernameAndPassword', 'WindowsAzureHDInsightService' + */ + authenticationType: SparkAuthenticationType; + /** + * The user name that you use to access Spark Server. + */ + username?: any; + /** + * The password corresponding to the user name that you provided in the Username field + */ + password?: SecretBaseUnion; + /** + * The partial URL corresponding to the Spark server. + */ + httpPath?: any; + /** + * Specifies whether the connections to the server are encrypted using SSL. The default value is + * false. + */ + enableSsl?: any; + /** + * The full path of the .pem file containing trusted CA certificates for verifying the server + * when connecting over SSL. This property can only be set when using SSL on self-hosted IR. The + * default value is the cacerts.pem file installed with the IR. + */ + trustedCertPath?: any; + /** + * Specifies whether to use a CA certificate from the system trust store or from a specified PEM + * file. The default value is false. + */ + useSystemTrustStore?: any; /** - * An ODBC connection string. Type: string, SecureString or AzureKeyVaultSecretReference. + * Specifies whether to require a CA-issued SSL certificate name to match the host name of the + * server when connecting over SSL. The default value is false. */ - connectionString?: any; + allowHostNameCNMismatch?: any; /** - * The Azure key vault secret reference of password in connection string. + * Specifies whether to allow self-signed certificates from the server. The default value is + * false. */ - pwd?: AzureKeyVaultSecretReference; + allowSelfSignedServerCert?: any; /** * The encrypted credential used for authentication. Credentials are encrypted using the * integration runtime credential manager. Type: string (or Expression with resultType string). @@ -3390,13 +3244,13 @@ export interface NetezzaLinkedService { } /** - * Vertica linked service. + * Shopify Service linked service. */ -export interface VerticaLinkedService { +export interface ShopifyLinkedService { /** * Polymorphic Discriminator */ - type: "Vertica"; + type: "Shopify"; /** * The integration runtime reference. */ @@ -3414,13 +3268,29 @@ export interface VerticaLinkedService { */ annotations?: any[]; /** - * An ODBC connection string. Type: string, SecureString or AzureKeyVaultSecretReference. + * The endpoint of the Shopify server. (i.e. mystore.myshopify.com) */ - connectionString?: any; + host: any; /** - * The Azure key vault secret reference of password in connection string. + * The API access token that can be used to access Shopify’s data. The token won't expire if it + * is offline mode. */ - pwd?: AzureKeyVaultSecretReference; + accessToken?: SecretBaseUnion; + /** + * Specifies whether the data source endpoints are encrypted using HTTPS. The default value is + * true. + */ + useEncryptedEndpoints?: any; + /** + * Specifies whether to require the host name in the server's certificate to match the host name + * of the server when connecting over SSL. The default value is true. + */ + useHostVerification?: any; + /** + * Specifies whether to verify the identity of the server when connecting over SSL. The default + * value is true. + */ + usePeerVerification?: any; /** * The encrypted credential used for authentication. Credentials are encrypted using the * integration runtime credential manager. Type: string (or Expression with resultType string). @@ -3429,13 +3299,13 @@ export interface VerticaLinkedService { } /** - * Zoho server linked service. + * ServiceNow server linked service. */ -export interface ZohoLinkedService { +export interface ServiceNowLinkedService { /** * Polymorphic Discriminator */ - type: "Zoho"; + type: "ServiceNow"; /** * The integration runtime reference. */ @@ -3453,13 +3323,29 @@ export interface ZohoLinkedService { */ annotations?: any[]; /** - * The endpoint of the Zoho server. (i.e. crm.zoho.com/crm/private) + * The endpoint of the ServiceNow server. (i.e. .service-now.com) */ endpoint: any; /** - * The access token for Zoho authentication. + * The authentication type to use. Possible values include: 'Basic', 'OAuth2' */ - accessToken?: SecretBaseUnion; + authenticationType: ServiceNowAuthenticationType; + /** + * The user name used to connect to the ServiceNow server for Basic and OAuth2 authentication. + */ + username?: any; + /** + * The password corresponding to the user name for Basic and OAuth2 authentication. + */ + password?: SecretBaseUnion; + /** + * The client id for OAuth2 authentication. + */ + clientId?: any; + /** + * The client secret for OAuth2 authentication. + */ + clientSecret?: SecretBaseUnion; /** * Specifies whether the data source endpoints are encrypted using HTTPS. The default value is * true. @@ -3483,13 +3369,13 @@ export interface ZohoLinkedService { } /** - * Xero Service linked service. + * QuickBooks server linked service. */ -export interface XeroLinkedService { +export interface QuickBooksLinkedService { /** * Polymorphic Discriminator */ - type: "Xero"; + type: "QuickBooks"; /** * The integration runtime reference. */ @@ -3507,34 +3393,34 @@ export interface XeroLinkedService { */ annotations?: any[]; /** - * The endpoint of the Xero server. (i.e. api.xero.com) + * The endpoint of the QuickBooks server. (i.e. quickbooks.api.intuit.com) */ - host: any; + endpoint: any; /** - * The consumer key associated with the Xero application. + * The company ID of the QuickBooks company to authorize. */ - consumerKey?: SecretBaseUnion; + companyId: any; /** - * The private key from the .pem file that was generated for your Xero private application. You - * must include all the text from the .pem file, including the Unix line endings( - * ). + * The consumer key for OAuth 1.0 authentication. */ - privateKey?: SecretBaseUnion; + consumerKey: any; /** - * Specifies whether the data source endpoints are encrypted using HTTPS. The default value is - * true. + * The consumer secret for OAuth 1.0 authentication. */ - useEncryptedEndpoints?: any; + consumerSecret: SecretBaseUnion; /** - * Specifies whether to require the host name in the server's certificate to match the host name - * of the server when connecting over SSL. The default value is true. + * The access token for OAuth 1.0 authentication. */ - useHostVerification?: any; + accessToken: SecretBaseUnion; /** - * Specifies whether to verify the identity of the server when connecting over SSL. The default - * value is true. + * The access token secret for OAuth 1.0 authentication. */ - usePeerVerification?: any; + accessTokenSecret: SecretBaseUnion; + /** + * Specifies whether the data source endpoints are encrypted using HTTPS. The default value is + * true. + */ + useEncryptedEndpoints?: any; /** * The encrypted credential used for authentication. Credentials are encrypted using the * integration runtime credential manager. Type: string (or Expression with resultType string). @@ -3543,13 +3429,13 @@ export interface XeroLinkedService { } /** - * Square Service linked service. + * Presto server linked service. */ -export interface SquareLinkedService { +export interface PrestoLinkedService { /** * Polymorphic Discriminator */ - type: "Square"; + type: "Presto"; /** * The integration runtime reference. */ @@ -3567,36 +3453,66 @@ export interface SquareLinkedService { */ annotations?: any[]; /** - * The URL of the Square instance. (i.e. mystore.mysquare.com) + * The IP address or host name of the Presto server. (i.e. 192.168.222.160) */ host: any; /** - * The client ID associated with your Square application. + * The version of the Presto server. (i.e. 0.148-t) */ - clientId: any; + serverVersion: any; /** - * The client secret associated with your Square application. + * The catalog context for all request against the server. */ - clientSecret?: SecretBaseUnion; + catalog: any; /** - * The redirect URL assigned in the Square application dashboard. (i.e. http://localhost:2500) + * The TCP port that the Presto server uses to listen for client connections. The default value + * is 8080. */ - redirectUri: any; + port?: any; /** - * Specifies whether the data source endpoints are encrypted using HTTPS. The default value is - * true. + * The authentication mechanism used to connect to the Presto server. Possible values include: + * 'Anonymous', 'LDAP' */ - useEncryptedEndpoints?: any; + authenticationType: PrestoAuthenticationType; /** - * Specifies whether to require the host name in the server's certificate to match the host name - * of the server when connecting over SSL. The default value is true. + * The user name used to connect to the Presto server. */ - useHostVerification?: any; + username?: any; /** - * Specifies whether to verify the identity of the server when connecting over SSL. The default - * value is true. + * The password corresponding to the user name. */ - usePeerVerification?: any; + password?: SecretBaseUnion; + /** + * Specifies whether the connections to the server are encrypted using SSL. The default value is + * false. + */ + enableSsl?: any; + /** + * The full path of the .pem file containing trusted CA certificates for verifying the server + * when connecting over SSL. This property can only be set when using SSL on self-hosted IR. The + * default value is the cacerts.pem file installed with the IR. + */ + trustedCertPath?: any; + /** + * Specifies whether to use a CA certificate from the system trust store or from a specified PEM + * file. The default value is false. + */ + useSystemTrustStore?: any; + /** + * Specifies whether to require a CA-issued SSL certificate name to match the host name of the + * server when connecting over SSL. The default value is false. + */ + allowHostNameCNMismatch?: any; + /** + * Specifies whether to allow self-signed certificates from the server. The default value is + * false. + */ + allowSelfSignedServerCert?: any; + /** + * The local time zone used by the connection. Valid values for this option are specified in the + * IANA Time Zone Database. The default value is the system time zone. + */ + timeZoneID?: any; /** * The encrypted credential used for authentication. Credentials are encrypted using the * integration runtime credential manager. Type: string (or Expression with resultType string). @@ -3605,13 +3521,13 @@ export interface SquareLinkedService { } /** - * Spark Server linked service. + * Phoenix server linked service. */ -export interface SparkLinkedService { +export interface PhoenixLinkedService { /** * Polymorphic Discriminator */ - type: "Spark"; + type: "Phoenix"; /** * The integration runtime reference. */ @@ -3629,40 +3545,32 @@ export interface SparkLinkedService { */ annotations?: any[]; /** - * IP address or host name of the Spark server + * The IP address or host name of the Phoenix server. (i.e. 192.168.222.160) */ host: any; /** - * The TCP port that the Spark server uses to listen for client connections. - */ - port: any; - /** - * The type of Spark server. Possible values include: 'SharkServer', 'SharkServer2', - * 'SparkThriftServer' + * The TCP port that the Phoenix server uses to listen for client connections. The default value + * is 8765. */ - serverType?: SparkServerType; + port?: any; /** - * The transport protocol to use in the Thrift layer. Possible values include: 'Binary', 'SASL', - * 'HTTP ' + * The partial URL corresponding to the Phoenix server. (i.e. /gateway/sandbox/phoenix/version). + * The default value is hbasephoenix if using WindowsAzureHDInsightService. */ - thriftTransportProtocol?: SparkThriftTransportProtocol; + httpPath?: any; /** - * The authentication method used to access the Spark server. Possible values include: - * 'Anonymous', 'Username', 'UsernameAndPassword', 'WindowsAzureHDInsightService' + * The authentication mechanism used to connect to the Phoenix server. Possible values include: + * 'Anonymous', 'UsernameAndPassword', 'WindowsAzureHDInsightService' */ - authenticationType: SparkAuthenticationType; + authenticationType: PhoenixAuthenticationType; /** - * The user name that you use to access Spark Server. + * The user name used to connect to the Phoenix server. */ username?: any; /** - * The password corresponding to the user name that you provided in the Username field + * The password corresponding to the user name. */ password?: SecretBaseUnion; - /** - * The partial URL corresponding to the Spark server. - */ - httpPath?: any; /** * Specifies whether the connections to the server are encrypted using SSL. The default value is * false. @@ -3697,13 +3605,13 @@ export interface SparkLinkedService { } /** - * Shopify Service linked service. + * Paypal Service linked service. */ -export interface ShopifyLinkedService { +export interface PaypalLinkedService { /** * Polymorphic Discriminator */ - type: "Shopify"; + type: "Paypal"; /** * The integration runtime reference. */ @@ -3721,14 +3629,17 @@ export interface ShopifyLinkedService { */ annotations?: any[]; /** - * The endpoint of the Shopify server. (i.e. mystore.myshopify.com) + * The URL of the PayPal instance. (i.e. api.sandbox.paypal.com) */ host: any; /** - * The API access token that can be used to access Shopify’s data. The token won't expire if it - * is offline mode. + * The client ID associated with your PayPal application. */ - accessToken?: SecretBaseUnion; + clientId: any; + /** + * The client secret associated with your PayPal application. + */ + clientSecret?: SecretBaseUnion; /** * Specifies whether the data source endpoints are encrypted using HTTPS. The default value is * true. @@ -3752,13 +3663,13 @@ export interface ShopifyLinkedService { } /** - * ServiceNow server linked service. + * Marketo server linked service. */ -export interface ServiceNowLinkedService { +export interface MarketoLinkedService { /** * Polymorphic Discriminator */ - type: "ServiceNow"; + type: "Marketo"; /** * The integration runtime reference. */ @@ -3776,27 +3687,15 @@ export interface ServiceNowLinkedService { */ annotations?: any[]; /** - * The endpoint of the ServiceNow server. (i.e. .service-now.com) + * The endpoint of the Marketo server. (i.e. 123-ABC-321.mktorest.com) */ endpoint: any; /** - * The authentication type to use. Possible values include: 'Basic', 'OAuth2' - */ - authenticationType: ServiceNowAuthenticationType; - /** - * The user name used to connect to the ServiceNow server for Basic and OAuth2 authentication. - */ - username?: any; - /** - * The password corresponding to the user name for Basic and OAuth2 authentication. - */ - password?: SecretBaseUnion; - /** - * The client id for OAuth2 authentication. + * The client Id of your Marketo service. */ - clientId?: any; + clientId: any; /** - * The client secret for OAuth2 authentication. + * The client secret of your Marketo service. */ clientSecret?: SecretBaseUnion; /** @@ -3822,13 +3721,13 @@ export interface ServiceNowLinkedService { } /** - * QuickBooks server linked service. + * Azure Database for MariaDB linked service. */ -export interface QuickBooksLinkedService { +export interface AzureMariaDBLinkedService { /** * Polymorphic Discriminator */ - type: "QuickBooks"; + type: "AzureMariaDB"; /** * The integration runtime reference. */ @@ -3846,34 +3745,13 @@ export interface QuickBooksLinkedService { */ annotations?: any[]; /** - * The endpoint of the QuickBooks server. (i.e. quickbooks.api.intuit.com) - */ - endpoint: any; - /** - * The company ID of the QuickBooks company to authorize. - */ - companyId: any; - /** - * The consumer key for OAuth 1.0 authentication. - */ - consumerKey: any; - /** - * The consumer secret for OAuth 1.0 authentication. - */ - consumerSecret: SecretBaseUnion; - /** - * The access token for OAuth 1.0 authentication. - */ - accessToken: SecretBaseUnion; - /** - * The access token secret for OAuth 1.0 authentication. + * An ODBC connection string. Type: string, SecureString or AzureKeyVaultSecretReference. */ - accessTokenSecret: SecretBaseUnion; + connectionString?: any; /** - * Specifies whether the data source endpoints are encrypted using HTTPS. The default value is - * true. + * The Azure key vault secret reference of password in connection string. */ - useEncryptedEndpoints?: any; + pwd?: AzureKeyVaultSecretReference; /** * The encrypted credential used for authentication. Credentials are encrypted using the * integration runtime credential manager. Type: string (or Expression with resultType string). @@ -3882,13 +3760,13 @@ export interface QuickBooksLinkedService { } /** - * Presto server linked service. + * MariaDB server linked service. */ -export interface PrestoLinkedService { +export interface MariaDBLinkedService { /** * Polymorphic Discriminator */ - type: "Presto"; + type: "MariaDB"; /** * The integration runtime reference. */ @@ -3906,66 +3784,67 @@ export interface PrestoLinkedService { */ annotations?: any[]; /** - * The IP address or host name of the Presto server. (i.e. 192.168.222.160) + * An ODBC connection string. Type: string, SecureString or AzureKeyVaultSecretReference. */ - host: any; + connectionString?: any; /** - * The version of the Presto server. (i.e. 0.148-t) + * The Azure key vault secret reference of password in connection string. */ - serverVersion: any; + pwd?: AzureKeyVaultSecretReference; /** - * The catalog context for all request against the server. + * The encrypted credential used for authentication. Credentials are encrypted using the + * integration runtime credential manager. Type: string (or Expression with resultType string). */ - catalog: any; + encryptedCredential?: any; +} + +/** + * Magento server linked service. + */ +export interface MagentoLinkedService { /** - * The TCP port that the Presto server uses to listen for client connections. The default value - * is 8080. + * Polymorphic Discriminator */ - port?: any; + type: "Magento"; /** - * The authentication mechanism used to connect to the Presto server. Possible values include: - * 'Anonymous', 'LDAP' + * The integration runtime reference. */ - authenticationType: PrestoAuthenticationType; + connectVia?: IntegrationRuntimeReference; /** - * The user name used to connect to the Presto server. + * Linked service description. */ - username?: any; + description?: string; /** - * The password corresponding to the user name. + * Parameters for linked service. */ - password?: SecretBaseUnion; + parameters?: { [propertyName: string]: ParameterSpecification }; /** - * Specifies whether the connections to the server are encrypted using SSL. The default value is - * false. + * List of tags that can be used for describing the linked service. */ - enableSsl?: any; + annotations?: any[]; /** - * The full path of the .pem file containing trusted CA certificates for verifying the server - * when connecting over SSL. This property can only be set when using SSL on self-hosted IR. The - * default value is the cacerts.pem file installed with the IR. + * The URL of the Magento instance. (i.e. 192.168.222.110/magento3) */ - trustedCertPath?: any; + host: any; /** - * Specifies whether to use a CA certificate from the system trust store or from a specified PEM - * file. The default value is false. + * The access token from Magento. */ - useSystemTrustStore?: any; + accessToken?: SecretBaseUnion; /** - * Specifies whether to require a CA-issued SSL certificate name to match the host name of the - * server when connecting over SSL. The default value is false. + * Specifies whether the data source endpoints are encrypted using HTTPS. The default value is + * true. */ - allowHostNameCNMismatch?: any; + useEncryptedEndpoints?: any; /** - * Specifies whether to allow self-signed certificates from the server. The default value is - * false. + * Specifies whether to require the host name in the server's certificate to match the host name + * of the server when connecting over SSL. The default value is true. */ - allowSelfSignedServerCert?: any; + useHostVerification?: any; /** - * The local time zone used by the connection. Valid values for this option are specified in the - * IANA Time Zone Database. The default value is the system time zone. + * Specifies whether to verify the identity of the server when connecting over SSL. The default + * value is true. */ - timeZoneID?: any; + usePeerVerification?: any; /** * The encrypted credential used for authentication. Credentials are encrypted using the * integration runtime credential manager. Type: string (or Expression with resultType string). @@ -3974,13 +3853,13 @@ export interface PrestoLinkedService { } /** - * Phoenix server linked service. + * Jira Service linked service. */ -export interface PhoenixLinkedService { +export interface JiraLinkedService { /** * Polymorphic Discriminator */ - type: "Phoenix"; + type: "Jira"; /** * The integration runtime reference. */ @@ -3998,58 +3877,37 @@ export interface PhoenixLinkedService { */ annotations?: any[]; /** - * The IP address or host name of the Phoenix server. (i.e. 192.168.222.160) + * The IP address or host name of the Jira service. (e.g. jira.example.com) */ host: any; /** - * The TCP port that the Phoenix server uses to listen for client connections. The default value - * is 8765. + * The TCP port that the Jira server uses to listen for client connections. The default value is + * 443 if connecting through HTTPS, or 8080 if connecting through HTTP. */ port?: any; /** - * The partial URL corresponding to the Phoenix server. (i.e. /gateway/sandbox/phoenix/version). - * The default value is hbasephoenix if using WindowsAzureHDInsightService. - */ - httpPath?: any; - /** - * The authentication mechanism used to connect to the Phoenix server. Possible values include: - * 'Anonymous', 'UsernameAndPassword', 'WindowsAzureHDInsightService' - */ - authenticationType: PhoenixAuthenticationType; - /** - * The user name used to connect to the Phoenix server. + * The user name that you use to access Jira Service. */ - username?: any; + username: any; /** - * The password corresponding to the user name. + * The password corresponding to the user name that you provided in the username field. */ password?: SecretBaseUnion; /** - * Specifies whether the connections to the server are encrypted using SSL. The default value is - * false. - */ - enableSsl?: any; - /** - * The full path of the .pem file containing trusted CA certificates for verifying the server - * when connecting over SSL. This property can only be set when using SSL on self-hosted IR. The - * default value is the cacerts.pem file installed with the IR. - */ - trustedCertPath?: any; - /** - * Specifies whether to use a CA certificate from the system trust store or from a specified PEM - * file. The default value is false. + * Specifies whether the data source endpoints are encrypted using HTTPS. The default value is + * true. */ - useSystemTrustStore?: any; + useEncryptedEndpoints?: any; /** - * Specifies whether to require a CA-issued SSL certificate name to match the host name of the - * server when connecting over SSL. The default value is false. + * Specifies whether to require the host name in the server's certificate to match the host name + * of the server when connecting over SSL. The default value is true. */ - allowHostNameCNMismatch?: any; + useHostVerification?: any; /** - * Specifies whether to allow self-signed certificates from the server. The default value is - * false. + * Specifies whether to verify the identity of the server when connecting over SSL. The default + * value is true. */ - allowSelfSignedServerCert?: any; + usePeerVerification?: any; /** * The encrypted credential used for authentication. Credentials are encrypted using the * integration runtime credential manager. Type: string (or Expression with resultType string). @@ -4058,13 +3916,13 @@ export interface PhoenixLinkedService { } /** - * Paypal Service linked service. + * Impala server linked service. */ -export interface PaypalLinkedService { +export interface ImpalaLinkedService { /** * Polymorphic Discriminator */ - type: "Paypal"; + type: "Impala"; /** * The integration runtime reference. */ @@ -4082,32 +3940,54 @@ export interface PaypalLinkedService { */ annotations?: any[]; /** - * The URL of the PayPal instance. (i.e. api.sandbox.paypal.com) + * The IP address or host name of the Impala server. (i.e. 192.168.222.160) */ host: any; /** - * The client ID associated with your PayPal application. + * The TCP port that the Impala server uses to listen for client connections. The default value + * is 21050. */ - clientId: any; + port?: any; /** - * The client secret associated with your PayPal application. + * The authentication type to use. Possible values include: 'Anonymous', 'SASLUsername', + * 'UsernameAndPassword' */ - clientSecret?: SecretBaseUnion; + authenticationType: ImpalaAuthenticationType; /** - * Specifies whether the data source endpoints are encrypted using HTTPS. The default value is - * true. + * The user name used to access the Impala server. The default value is anonymous when using + * SASLUsername. */ - useEncryptedEndpoints?: any; + username?: any; /** - * Specifies whether to require the host name in the server's certificate to match the host name - * of the server when connecting over SSL. The default value is true. + * The password corresponding to the user name when using UsernameAndPassword. */ - useHostVerification?: any; + password?: SecretBaseUnion; /** - * Specifies whether to verify the identity of the server when connecting over SSL. The default - * value is true. + * Specifies whether the connections to the server are encrypted using SSL. The default value is + * false. */ - usePeerVerification?: any; + enableSsl?: any; + /** + * The full path of the .pem file containing trusted CA certificates for verifying the server + * when connecting over SSL. This property can only be set when using SSL on self-hosted IR. The + * default value is the cacerts.pem file installed with the IR. + */ + trustedCertPath?: any; + /** + * Specifies whether to use a CA certificate from the system trust store or from a specified PEM + * file. The default value is false. + */ + useSystemTrustStore?: any; + /** + * Specifies whether to require a CA-issued SSL certificate name to match the host name of the + * server when connecting over SSL. The default value is false. + */ + allowHostNameCNMismatch?: any; + /** + * Specifies whether to allow self-signed certificates from the server. The default value is + * false. + */ + allowSelfSignedServerCert?: any; /** * The encrypted credential used for authentication. Credentials are encrypted using the * integration runtime credential manager. Type: string (or Expression with resultType string). @@ -4116,13 +3996,13 @@ export interface PaypalLinkedService { } /** - * Marketo server linked service. + * Hubspot Service linked service. */ -export interface MarketoLinkedService { +export interface HubspotLinkedService { /** * Polymorphic Discriminator */ - type: "Marketo"; + type: "Hubspot"; /** * The integration runtime reference. */ @@ -4140,17 +4020,21 @@ export interface MarketoLinkedService { */ annotations?: any[]; /** - * The endpoint of the Marketo server. (i.e. 123-ABC-321.mktorest.com) - */ - endpoint: any; - /** - * The client Id of your Marketo service. + * The client ID associated with your Hubspot application. */ clientId: any; /** - * The client secret of your Marketo service. + * The client secret associated with your Hubspot application. */ clientSecret?: SecretBaseUnion; + /** + * The access token obtained when initially authenticating your OAuth integration. + */ + accessToken?: SecretBaseUnion; + /** + * The refresh token obtained when initially authenticating your OAuth integration. + */ + refreshToken?: SecretBaseUnion; /** * Specifies whether the data source endpoints are encrypted using HTTPS. The default value is * true. @@ -4174,13 +4058,13 @@ export interface MarketoLinkedService { } /** - * Azure Database for MariaDB linked service. + * Hive Server linked service. */ -export interface AzureMariaDBLinkedService { +export interface HiveLinkedService { /** * Polymorphic Discriminator */ - type: "AzureMariaDB"; + type: "Hive"; /** * The integration runtime reference. */ @@ -4198,52 +4082,80 @@ export interface AzureMariaDBLinkedService { */ annotations?: any[]; /** - * An ODBC connection string. Type: string, SecureString or AzureKeyVaultSecretReference. + * IP address or host name of the Hive server, separated by ';' for multiple hosts (only when + * serviceDiscoveryMode is enable). */ - connectionString?: any; + host: any; /** - * The Azure key vault secret reference of password in connection string. + * The TCP port that the Hive server uses to listen for client connections. */ - pwd?: AzureKeyVaultSecretReference; + port?: any; /** - * The encrypted credential used for authentication. Credentials are encrypted using the - * integration runtime credential manager. Type: string (or Expression with resultType string). + * The type of Hive server. Possible values include: 'HiveServer1', 'HiveServer2', + * 'HiveThriftServer' */ - encryptedCredential?: any; -} - -/** - * MariaDB server linked service. - */ -export interface MariaDBLinkedService { + serverType?: HiveServerType; /** - * Polymorphic Discriminator + * The transport protocol to use in the Thrift layer. Possible values include: 'Binary', 'SASL', + * 'HTTP ' */ - type: "MariaDB"; + thriftTransportProtocol?: HiveThriftTransportProtocol; /** - * The integration runtime reference. + * The authentication method used to access the Hive server. Possible values include: + * 'Anonymous', 'Username', 'UsernameAndPassword', 'WindowsAzureHDInsightService' */ - connectVia?: IntegrationRuntimeReference; + authenticationType: HiveAuthenticationType; /** - * Linked service description. + * true to indicate using the ZooKeeper service, false not. */ - description?: string; + serviceDiscoveryMode?: any; /** - * Parameters for linked service. + * The namespace on ZooKeeper under which Hive Server 2 nodes are added. */ - parameters?: { [propertyName: string]: ParameterSpecification }; + zooKeeperNameSpace?: any; /** - * List of tags that can be used for describing the linked service. + * Specifies whether the driver uses native HiveQL queries,or converts them into an equivalent + * form in HiveQL. */ - annotations?: any[]; + useNativeQuery?: any; /** - * An ODBC connection string. Type: string, SecureString or AzureKeyVaultSecretReference. + * The user name that you use to access Hive Server. */ - connectionString?: any; + username?: any; /** - * The Azure key vault secret reference of password in connection string. + * The password corresponding to the user name that you provided in the Username field */ - pwd?: AzureKeyVaultSecretReference; + password?: SecretBaseUnion; + /** + * The partial URL corresponding to the Hive server. + */ + httpPath?: any; + /** + * Specifies whether the connections to the server are encrypted using SSL. The default value is + * false. + */ + enableSsl?: any; + /** + * The full path of the .pem file containing trusted CA certificates for verifying the server + * when connecting over SSL. This property can only be set when using SSL on self-hosted IR. The + * default value is the cacerts.pem file installed with the IR. + */ + trustedCertPath?: any; + /** + * Specifies whether to use a CA certificate from the system trust store or from a specified PEM + * file. The default value is false. + */ + useSystemTrustStore?: any; + /** + * Specifies whether to require a CA-issued SSL certificate name to match the host name of the + * server when connecting over SSL. The default value is false. + */ + allowHostNameCNMismatch?: any; + /** + * Specifies whether to allow self-signed certificates from the server. The default value is + * false. + */ + allowSelfSignedServerCert?: any; /** * The encrypted credential used for authentication. Credentials are encrypted using the * integration runtime credential manager. Type: string (or Expression with resultType string). @@ -4252,13 +4164,13 @@ export interface MariaDBLinkedService { } /** - * Magento server linked service. + * HBase server linked service. */ -export interface MagentoLinkedService { +export interface HBaseLinkedService { /** * Polymorphic Discriminator */ - type: "Magento"; + type: "HBase"; /** * The integration runtime reference. */ @@ -4276,28 +4188,52 @@ export interface MagentoLinkedService { */ annotations?: any[]; /** - * The URL of the Magento instance. (i.e. 192.168.222.110/magento3) + * The IP address or host name of the HBase server. (i.e. 192.168.222.160) */ host: any; /** - * The access token from Magento. + * The TCP port that the HBase instance uses to listen for client connections. The default value + * is 9090. */ - accessToken?: SecretBaseUnion; + port?: any; /** - * Specifies whether the data source endpoints are encrypted using HTTPS. The default value is - * true. + * The partial URL corresponding to the HBase server. (i.e. /gateway/sandbox/hbase/version) */ - useEncryptedEndpoints?: any; + httpPath?: any; /** - * Specifies whether to require the host name in the server's certificate to match the host name - * of the server when connecting over SSL. The default value is true. + * The authentication mechanism to use to connect to the HBase server. Possible values include: + * 'Anonymous', 'Basic' */ - useHostVerification?: any; + authenticationType: HBaseAuthenticationType; /** - * Specifies whether to verify the identity of the server when connecting over SSL. The default - * value is true. + * The user name used to connect to the HBase instance. */ - usePeerVerification?: any; + username?: any; + /** + * The password corresponding to the user name. + */ + password?: SecretBaseUnion; + /** + * Specifies whether the connections to the server are encrypted using SSL. The default value is + * false. + */ + enableSsl?: any; + /** + * The full path of the .pem file containing trusted CA certificates for verifying the server + * when connecting over SSL. This property can only be set when using SSL on self-hosted IR. The + * default value is the cacerts.pem file installed with the IR. + */ + trustedCertPath?: any; + /** + * Specifies whether to require a CA-issued SSL certificate name to match the host name of the + * server when connecting over SSL. The default value is false. + */ + allowHostNameCNMismatch?: any; + /** + * Specifies whether to allow self-signed certificates from the server. The default value is + * false. + */ + allowSelfSignedServerCert?: any; /** * The encrypted credential used for authentication. Credentials are encrypted using the * integration runtime credential manager. Type: string (or Expression with resultType string). @@ -4306,13 +4242,13 @@ export interface MagentoLinkedService { } /** - * Jira Service linked service. + * Greenplum Database linked service. */ -export interface JiraLinkedService { +export interface GreenplumLinkedService { /** * Polymorphic Discriminator */ - type: "Jira"; + type: "Greenplum"; /** * The integration runtime reference. */ @@ -4330,37 +4266,13 @@ export interface JiraLinkedService { */ annotations?: any[]; /** - * The IP address or host name of the Jira service. (e.g. jira.example.com) - */ - host: any; - /** - * The TCP port that the Jira server uses to listen for client connections. The default value is - * 443 if connecting through HTTPS, or 8080 if connecting through HTTP. - */ - port?: any; - /** - * The user name that you use to access Jira Service. - */ - username: any; - /** - * The password corresponding to the user name that you provided in the username field. - */ - password?: SecretBaseUnion; - /** - * Specifies whether the data source endpoints are encrypted using HTTPS. The default value is - * true. - */ - useEncryptedEndpoints?: any; - /** - * Specifies whether to require the host name in the server's certificate to match the host name - * of the server when connecting over SSL. The default value is true. + * An ODBC connection string. Type: string, SecureString or AzureKeyVaultSecretReference. */ - useHostVerification?: any; + connectionString?: any; /** - * Specifies whether to verify the identity of the server when connecting over SSL. The default - * value is true. + * The Azure key vault secret reference of password in connection string. */ - usePeerVerification?: any; + pwd?: AzureKeyVaultSecretReference; /** * The encrypted credential used for authentication. Credentials are encrypted using the * integration runtime credential manager. Type: string (or Expression with resultType string). @@ -4369,13 +4281,13 @@ export interface JiraLinkedService { } /** - * Impala server linked service. + * Google BigQuery service linked service. */ -export interface ImpalaLinkedService { +export interface GoogleBigQueryLinkedService { /** * Polymorphic Discriminator */ - type: "Impala"; + type: "GoogleBigQuery"; /** * The integration runtime reference. */ @@ -4393,33 +4305,49 @@ export interface ImpalaLinkedService { */ annotations?: any[]; /** - * The IP address or host name of the Impala server. (i.e. 192.168.222.160) + * The default BigQuery project to query against. */ - host: any; + project: any; /** - * The TCP port that the Impala server uses to listen for client connections. The default value - * is 21050. + * A comma-separated list of public BigQuery projects to access. */ - port?: any; + additionalProjects?: any; /** - * The authentication type to use. Possible values include: 'Anonymous', 'SASLUsername', - * 'UsernameAndPassword' + * Whether to request access to Google Drive. Allowing Google Drive access enables support for + * federated tables that combine BigQuery data with data from Google Drive. The default value is + * false. */ - authenticationType: ImpalaAuthenticationType; + requestGoogleDriveScope?: any; /** - * The user name used to access the Impala server. The default value is anonymous when using - * SASLUsername. + * The OAuth 2.0 authentication mechanism used for authentication. ServiceAuthentication can only + * be used on self-hosted IR. Possible values include: 'ServiceAuthentication', + * 'UserAuthentication' */ - username?: any; + authenticationType: GoogleBigQueryAuthenticationType; /** - * The password corresponding to the user name when using UsernameAndPassword. + * The refresh token obtained from Google for authorizing access to BigQuery for + * UserAuthentication. */ - password?: SecretBaseUnion; + refreshToken?: SecretBaseUnion; /** - * Specifies whether the connections to the server are encrypted using SSL. The default value is - * false. + * The client id of the google application used to acquire the refresh token. Type: string (or + * Expression with resultType string). */ - enableSsl?: any; + clientId?: any; + /** + * The client secret of the google application used to acquire the refresh token. + */ + clientSecret?: SecretBaseUnion; + /** + * The service account email ID that is used for ServiceAuthentication and can only be used on + * self-hosted IR. + */ + email?: any; + /** + * The full path to the .p12 key file that is used to authenticate the service account email + * address and can only be used on self-hosted IR. + */ + keyFilePath?: any; /** * The full path of the .pem file containing trusted CA certificates for verifying the server * when connecting over SSL. This property can only be set when using SSL on self-hosted IR. The @@ -4431,16 +4359,6 @@ export interface ImpalaLinkedService { * file. The default value is false. */ useSystemTrustStore?: any; - /** - * Specifies whether to require a CA-issued SSL certificate name to match the host name of the - * server when connecting over SSL. The default value is false. - */ - allowHostNameCNMismatch?: any; - /** - * Specifies whether to allow self-signed certificates from the server. The default value is - * false. - */ - allowSelfSignedServerCert?: any; /** * The encrypted credential used for authentication. Credentials are encrypted using the * integration runtime credential manager. Type: string (or Expression with resultType string). @@ -4449,13 +4367,13 @@ export interface ImpalaLinkedService { } /** - * Hubspot Service linked service. + * Eloqua server linked service. */ -export interface HubspotLinkedService { +export interface EloquaLinkedService { /** * Polymorphic Discriminator */ - type: "Hubspot"; + type: "Eloqua"; /** * The integration runtime reference. */ @@ -4473,21 +4391,18 @@ export interface HubspotLinkedService { */ annotations?: any[]; /** - * The client ID associated with your Hubspot application. - */ - clientId: any; - /** - * The client secret associated with your Hubspot application. + * The endpoint of the Eloqua server. (i.e. eloqua.example.com) */ - clientSecret?: SecretBaseUnion; + endpoint: any; /** - * The access token obtained when initially authenticating your OAuth integration. + * The site name and user name of your Eloqua account in the form: sitename/username. (i.e. + * Eloqua/Alice) */ - accessToken?: SecretBaseUnion; + username: any; /** - * The refresh token obtained when initially authenticating your OAuth integration. + * The password corresponding to the user name. */ - refreshToken?: SecretBaseUnion; + password?: SecretBaseUnion; /** * Specifies whether the data source endpoints are encrypted using HTTPS. The default value is * true. @@ -4511,13 +4426,13 @@ export interface HubspotLinkedService { } /** - * Hive Server linked service. + * Drill server linked service. */ -export interface HiveLinkedService { +export interface DrillLinkedService { /** * Polymorphic Discriminator */ - type: "Hive"; + type: "Drill"; /** * The integration runtime reference. */ @@ -4535,80 +4450,52 @@ export interface HiveLinkedService { */ annotations?: any[]; /** - * IP address or host name of the Hive server, separated by ';' for multiple hosts (only when - * serviceDiscoveryMode is enable). - */ - host: any; - /** - * The TCP port that the Hive server uses to listen for client connections. - */ - port?: any; - /** - * The type of Hive server. Possible values include: 'HiveServer1', 'HiveServer2', - * 'HiveThriftServer' - */ - serverType?: HiveServerType; - /** - * The transport protocol to use in the Thrift layer. Possible values include: 'Binary', 'SASL', - * 'HTTP ' - */ - thriftTransportProtocol?: HiveThriftTransportProtocol; - /** - * The authentication method used to access the Hive server. Possible values include: - * 'Anonymous', 'Username', 'UsernameAndPassword', 'WindowsAzureHDInsightService' - */ - authenticationType: HiveAuthenticationType; - /** - * true to indicate using the ZooKeeper service, false not. - */ - serviceDiscoveryMode?: any; - /** - * The namespace on ZooKeeper under which Hive Server 2 nodes are added. + * An ODBC connection string. Type: string, SecureString or AzureKeyVaultSecretReference. */ - zooKeeperNameSpace?: any; + connectionString?: any; /** - * Specifies whether the driver uses native HiveQL queries,or converts them into an equivalent - * form in HiveQL. + * The Azure key vault secret reference of password in connection string. */ - useNativeQuery?: any; + pwd?: AzureKeyVaultSecretReference; /** - * The user name that you use to access Hive Server. + * The encrypted credential used for authentication. Credentials are encrypted using the + * integration runtime credential manager. Type: string (or Expression with resultType string). */ - username?: any; + encryptedCredential?: any; +} + +/** + * Couchbase server linked service. + */ +export interface CouchbaseLinkedService { /** - * The password corresponding to the user name that you provided in the Username field + * Polymorphic Discriminator */ - password?: SecretBaseUnion; + type: "Couchbase"; /** - * The partial URL corresponding to the Hive server. + * The integration runtime reference. */ - httpPath?: any; + connectVia?: IntegrationRuntimeReference; /** - * Specifies whether the connections to the server are encrypted using SSL. The default value is - * false. + * Linked service description. */ - enableSsl?: any; + description?: string; /** - * The full path of the .pem file containing trusted CA certificates for verifying the server - * when connecting over SSL. This property can only be set when using SSL on self-hosted IR. The - * default value is the cacerts.pem file installed with the IR. + * Parameters for linked service. */ - trustedCertPath?: any; + parameters?: { [propertyName: string]: ParameterSpecification }; /** - * Specifies whether to use a CA certificate from the system trust store or from a specified PEM - * file. The default value is false. + * List of tags that can be used for describing the linked service. */ - useSystemTrustStore?: any; + annotations?: any[]; /** - * Specifies whether to require a CA-issued SSL certificate name to match the host name of the - * server when connecting over SSL. The default value is false. + * An ODBC connection string. Type: string, SecureString or AzureKeyVaultSecretReference. */ - allowHostNameCNMismatch?: any; + connectionString?: any; /** - * Specifies whether to allow self-signed certificates from the server. The default value is - * false. + * The Azure key vault secret reference of credString in connection string. */ - allowSelfSignedServerCert?: any; + credString?: AzureKeyVaultSecretReference; /** * The encrypted credential used for authentication. Credentials are encrypted using the * integration runtime credential manager. Type: string (or Expression with resultType string). @@ -4617,13 +4504,13 @@ export interface HiveLinkedService { } /** - * HBase server linked service. + * Concur Service linked service. */ -export interface HBaseLinkedService { +export interface ConcurLinkedService { /** * Polymorphic Discriminator */ - type: "HBase"; + type: "Concur"; /** * The integration runtime reference. */ @@ -4641,52 +4528,32 @@ export interface HBaseLinkedService { */ annotations?: any[]; /** - * The IP address or host name of the HBase server. (i.e. 192.168.222.160) - */ - host: any; - /** - * The TCP port that the HBase instance uses to listen for client connections. The default value - * is 9090. - */ - port?: any; - /** - * The partial URL corresponding to the HBase server. (i.e. /gateway/sandbox/hbase/version) - */ - httpPath?: any; - /** - * The authentication mechanism to use to connect to the HBase server. Possible values include: - * 'Anonymous', 'Basic' + * Application client_id supplied by Concur App Management. */ - authenticationType: HBaseAuthenticationType; + clientId: any; /** - * The user name used to connect to the HBase instance. + * The user name that you use to access Concur Service. */ - username?: any; + username: any; /** - * The password corresponding to the user name. + * The password corresponding to the user name that you provided in the username field. */ password?: SecretBaseUnion; /** - * Specifies whether the connections to the server are encrypted using SSL. The default value is - * false. - */ - enableSsl?: any; - /** - * The full path of the .pem file containing trusted CA certificates for verifying the server - * when connecting over SSL. This property can only be set when using SSL on self-hosted IR. The - * default value is the cacerts.pem file installed with the IR. + * Specifies whether the data source endpoints are encrypted using HTTPS. The default value is + * true. */ - trustedCertPath?: any; + useEncryptedEndpoints?: any; /** - * Specifies whether to require a CA-issued SSL certificate name to match the host name of the - * server when connecting over SSL. The default value is false. + * Specifies whether to require the host name in the server's certificate to match the host name + * of the server when connecting over SSL. The default value is true. */ - allowHostNameCNMismatch?: any; + useHostVerification?: any; /** - * Specifies whether to allow self-signed certificates from the server. The default value is - * false. + * Specifies whether to verify the identity of the server when connecting over SSL. The default + * value is true. */ - allowSelfSignedServerCert?: any; + usePeerVerification?: any; /** * The encrypted credential used for authentication. Credentials are encrypted using the * integration runtime credential manager. Type: string (or Expression with resultType string). @@ -4695,13 +4562,13 @@ export interface HBaseLinkedService { } /** - * Greenplum Database linked service. + * Azure PostgreSQL linked service. */ -export interface GreenplumLinkedService { +export interface AzurePostgreSqlLinkedService { /** * Polymorphic Discriminator */ - type: "Greenplum"; + type: "AzurePostgreSql"; /** * The integration runtime reference. */ @@ -4725,7 +4592,7 @@ export interface GreenplumLinkedService { /** * The Azure key vault secret reference of password in connection string. */ - pwd?: AzureKeyVaultSecretReference; + password?: AzureKeyVaultSecretReference; /** * The encrypted credential used for authentication. Credentials are encrypted using the * integration runtime credential manager. Type: string (or Expression with resultType string). @@ -4734,13 +4601,13 @@ export interface GreenplumLinkedService { } /** - * Google BigQuery service linked service. + * Amazon Marketplace Web Service linked service. */ -export interface GoogleBigQueryLinkedService { +export interface AmazonMWSLinkedService { /** * Polymorphic Discriminator */ - type: "GoogleBigQuery"; + type: "AmazonMWS"; /** * The integration runtime reference. */ @@ -4758,60 +4625,97 @@ export interface GoogleBigQueryLinkedService { */ annotations?: any[]; /** - * The default BigQuery project to query against. + * The endpoint of the Amazon MWS server, (i.e. mws.amazonservices.com) */ - project: any; + endpoint: any; /** - * A comma-separated list of public BigQuery projects to access. + * The Amazon Marketplace ID you want to retrieve data from. To retrieve data from multiple + * Marketplace IDs, separate them with a comma (,). (i.e. A2EUQ1WTGCTBG2) */ - additionalProjects?: any; + marketplaceID: any; /** - * Whether to request access to Google Drive. Allowing Google Drive access enables support for - * federated tables that combine BigQuery data with data from Google Drive. The default value is - * false. + * The Amazon seller ID. */ - requestGoogleDriveScope?: any; + sellerID: any; /** - * The OAuth 2.0 authentication mechanism used for authentication. ServiceAuthentication can only - * be used on self-hosted IR. Possible values include: 'ServiceAuthentication', - * 'UserAuthentication' + * The Amazon MWS authentication token. */ - authenticationType: GoogleBigQueryAuthenticationType; + mwsAuthToken?: SecretBaseUnion; /** - * The refresh token obtained from Google for authorizing access to BigQuery for - * UserAuthentication. + * The access key id used to access data. */ - refreshToken?: SecretBaseUnion; + accessKeyId: any; /** - * The client id of the google application used to acquire the refresh token. Type: string (or - * Expression with resultType string). + * The secret key used to access data. */ - clientId?: any; + secretKey?: SecretBaseUnion; /** - * The client secret of the google application used to acquire the refresh token. + * Specifies whether the data source endpoints are encrypted using HTTPS. The default value is + * true. */ - clientSecret?: SecretBaseUnion; + useEncryptedEndpoints?: any; /** - * The service account email ID that is used for ServiceAuthentication and can only be used on - * self-hosted IR. + * Specifies whether to require the host name in the server's certificate to match the host name + * of the server when connecting over SSL. The default value is true. */ - email?: any; + useHostVerification?: any; /** - * The full path to the .p12 key file that is used to authenticate the service account email - * address and can only be used on self-hosted IR. + * Specifies whether to verify the identity of the server when connecting over SSL. The default + * value is true. */ - keyFilePath?: any; + usePeerVerification?: any; /** - * The full path of the .pem file containing trusted CA certificates for verifying the server - * when connecting over SSL. This property can only be set when using SSL on self-hosted IR. The - * default value is the cacerts.pem file installed with the IR. + * The encrypted credential used for authentication. Credentials are encrypted using the + * integration runtime credential manager. Type: string (or Expression with resultType string). + */ + encryptedCredential?: any; +} + +/** + * SAP HANA Linked Service. + */ +export interface SapHanaLinkedService { + /** + * Polymorphic Discriminator + */ + type: "SapHana"; + /** + * The integration runtime reference. + */ + connectVia?: IntegrationRuntimeReference; + /** + * Linked service description. + */ + description?: string; + /** + * Parameters for linked service. + */ + parameters?: { [propertyName: string]: ParameterSpecification }; + /** + * List of tags that can be used for describing the linked service. + */ + annotations?: any[]; + /** + * SAP HANA ODBC connection string. Type: string, SecureString or AzureKeyVaultSecretReference. + */ + connectionString?: any; + /** + * Host name of the SAP HANA server. Type: string (or Expression with resultType string). + */ + server: any; + /** + * The authentication type to be used to connect to the SAP HANA server. Possible values include: + * 'Basic', 'Windows' + */ + authenticationType?: SapHanaAuthenticationType; + /** + * Username to access the SAP HANA server. Type: string (or Expression with resultType string). */ - trustedCertPath?: any; + userName?: any; /** - * Specifies whether to use a CA certificate from the system trust store or from a specified PEM - * file. The default value is false. + * Password to access the SAP HANA server. */ - useSystemTrustStore?: any; + password?: SecretBaseUnion; /** * The encrypted credential used for authentication. Credentials are encrypted using the * integration runtime credential manager. Type: string (or Expression with resultType string). @@ -4820,13 +4724,13 @@ export interface GoogleBigQueryLinkedService { } /** - * Eloqua server linked service. + * SAP Business Warehouse Linked Service. */ -export interface EloquaLinkedService { +export interface SapBWLinkedService { /** * Polymorphic Discriminator */ - type: "Eloqua"; + type: "SapBW"; /** * The integration runtime reference. */ @@ -4844,33 +4748,27 @@ export interface EloquaLinkedService { */ annotations?: any[]; /** - * The endpoint of the Eloqua server. (i.e. eloqua.example.com) - */ - endpoint: any; - /** - * The site name and user name of your Eloqua account in the form: sitename/username. (i.e. - * Eloqua/Alice) + * Host name of the SAP BW instance. Type: string (or Expression with resultType string). */ - username: any; + server: any; /** - * The password corresponding to the user name. + * System number of the BW system. (Usually a two-digit decimal number represented as a string.) + * Type: string (or Expression with resultType string). */ - password?: SecretBaseUnion; + systemNumber: any; /** - * Specifies whether the data source endpoints are encrypted using HTTPS. The default value is - * true. + * Client ID of the client on the BW system. (Usually a three-digit decimal number represented as + * a string) Type: string (or Expression with resultType string). */ - useEncryptedEndpoints?: any; + clientId: any; /** - * Specifies whether to require the host name in the server's certificate to match the host name - * of the server when connecting over SSL. The default value is true. + * Username to access the SAP BW server. Type: string (or Expression with resultType string). */ - useHostVerification?: any; + userName?: any; /** - * Specifies whether to verify the identity of the server when connecting over SSL. The default - * value is true. + * Password to access the SAP BW server. */ - usePeerVerification?: any; + password?: SecretBaseUnion; /** * The encrypted credential used for authentication. Credentials are encrypted using the * integration runtime credential manager. Type: string (or Expression with resultType string). @@ -4879,13 +4777,13 @@ export interface EloquaLinkedService { } /** - * Drill server linked service. + * A linked service for an SSH File Transfer Protocol (SFTP) server. */ -export interface DrillLinkedService { +export interface SftpServerLinkedService { /** * Polymorphic Discriminator */ - type: "Drill"; + type: "Sftp"; /** * The integration runtime reference. */ @@ -4903,28 +4801,70 @@ export interface DrillLinkedService { */ annotations?: any[]; /** - * An ODBC connection string. Type: string, SecureString or AzureKeyVaultSecretReference. + * The SFTP server host name. Type: string (or Expression with resultType string). */ - connectionString?: any; + host: any; /** - * The Azure key vault secret reference of password in connection string. + * The TCP port number that the SFTP server uses to listen for client connections. Default value + * is 22. Type: integer (or Expression with resultType integer), minimum: 0. */ - pwd?: AzureKeyVaultSecretReference; + port?: any; + /** + * The authentication type to be used to connect to the FTP server. Possible values include: + * 'Basic', 'SshPublicKey' + */ + authenticationType?: SftpAuthenticationType; + /** + * The username used to log on to the SFTP server. Type: string (or Expression with resultType + * string). + */ + userName?: any; + /** + * Password to logon the SFTP server for Basic authentication. + */ + password?: SecretBaseUnion; /** * The encrypted credential used for authentication. Credentials are encrypted using the * integration runtime credential manager. Type: string (or Expression with resultType string). */ encryptedCredential?: any; + /** + * The SSH private key file path for SshPublicKey authentication. Only valid for on-premises + * copy. For on-premises copy with SshPublicKey authentication, either PrivateKeyPath or + * PrivateKeyContent should be specified. SSH private key should be OpenSSH format. Type: string + * (or Expression with resultType string). + */ + privateKeyPath?: any; + /** + * Base64 encoded SSH private key content for SshPublicKey authentication. For on-premises copy + * with SshPublicKey authentication, either PrivateKeyPath or PrivateKeyContent should be + * specified. SSH private key should be OpenSSH format. + */ + privateKeyContent?: SecretBaseUnion; + /** + * The password to decrypt the SSH private key if the SSH private key is encrypted. + */ + passPhrase?: SecretBaseUnion; + /** + * If true, skip the SSH host key validation. Default value is false. Type: boolean (or + * Expression with resultType boolean). + */ + skipHostKeyValidation?: any; + /** + * The host key finger-print of the SFTP server. When SkipHostKeyValidation is false, + * HostKeyFingerprint should be specified. Type: string (or Expression with resultType string). + */ + hostKeyFingerprint?: any; } /** - * Couchbase server linked service. + * A FTP server Linked Service. */ -export interface CouchbaseLinkedService { +export interface FtpServerLinkedService { /** * Polymorphic Discriminator */ - type: "Couchbase"; + type: "FtpServer"; /** * The integration runtime reference. */ @@ -4942,28 +4882,52 @@ export interface CouchbaseLinkedService { */ annotations?: any[]; /** - * An ODBC connection string. Type: string, SecureString or AzureKeyVaultSecretReference. + * Host name of the FTP server. Type: string (or Expression with resultType string). */ - connectionString?: any; + host: any; /** - * The Azure key vault secret reference of credString in connection string. + * The TCP port number that the FTP server uses to listen for client connections. Default value + * is 21. Type: integer (or Expression with resultType integer), minimum: 0. */ - credString?: AzureKeyVaultSecretReference; + port?: any; + /** + * The authentication type to be used to connect to the FTP server. Possible values include: + * 'Basic', 'Anonymous' + */ + authenticationType?: FtpAuthenticationType; + /** + * Username to logon the FTP server. Type: string (or Expression with resultType string). + */ + userName?: any; + /** + * Password to logon the FTP server. + */ + password?: SecretBaseUnion; /** * The encrypted credential used for authentication. Credentials are encrypted using the * integration runtime credential manager. Type: string (or Expression with resultType string). */ encryptedCredential?: any; + /** + * If true, connect to the FTP server over SSL/TLS channel. Default value is true. Type: boolean + * (or Expression with resultType boolean). + */ + enableSsl?: any; + /** + * If true, validate the FTP server SSL certificate when connect over SSL/TLS channel. Default + * value is true. Type: boolean (or Expression with resultType boolean). + */ + enableServerCertificateValidation?: any; } /** - * Concur Service linked service. + * Linked service for an HTTP source. */ -export interface ConcurLinkedService { +export interface HttpLinkedService { /** * Polymorphic Discriminator */ - type: "Concur"; + type: "HttpServer"; /** * The integration runtime reference. */ @@ -4981,47 +4945,58 @@ export interface ConcurLinkedService { */ annotations?: any[]; /** - * Application client_id supplied by Concur App Management. + * The base URL of the HTTP endpoint, e.g. http://www.microsoft.com. Type: string (or Expression + * with resultType string). */ - clientId: any; + url: any; /** - * The user name that you use to access Concur Service. + * The authentication type to be used to connect to the HTTP server. Possible values include: + * 'Basic', 'Anonymous', 'Digest', 'Windows', 'ClientCertificate' */ - username: any; + authenticationType?: HttpAuthenticationType; /** - * The password corresponding to the user name that you provided in the username field. + * User name for Basic, Digest, or Windows authentication. Type: string (or Expression with + * resultType string). */ - password?: SecretBaseUnion; + userName?: any; /** - * Specifies whether the data source endpoints are encrypted using HTTPS. The default value is - * true. + * Password for Basic, Digest, Windows, or ClientCertificate with EmbeddedCertData + * authentication. */ - useEncryptedEndpoints?: any; + password?: SecretBaseUnion; /** - * Specifies whether to require the host name in the server's certificate to match the host name - * of the server when connecting over SSL. The default value is true. + * Base64 encoded certificate data for ClientCertificate authentication. For on-premises copy + * with ClientCertificate authentication, either CertThumbprint or EmbeddedCertData/Password + * should be specified. Type: string (or Expression with resultType string). */ - useHostVerification?: any; + embeddedCertData?: any; /** - * Specifies whether to verify the identity of the server when connecting over SSL. The default - * value is true. + * Thumbprint of certificate for ClientCertificate authentication. Only valid for on-premises + * copy. For on-premises copy with ClientCertificate authentication, either CertThumbprint or + * EmbeddedCertData/Password should be specified. Type: string (or Expression with resultType + * string). */ - usePeerVerification?: any; + certThumbprint?: any; /** * The encrypted credential used for authentication. Credentials are encrypted using the * integration runtime credential manager. Type: string (or Expression with resultType string). */ encryptedCredential?: any; + /** + * If true, validate the HTTPS server SSL certificate. Default value is true. Type: boolean (or + * Expression with resultType boolean). + */ + enableServerCertificateValidation?: any; } /** - * Azure PostgreSQL linked service. + * Linked service for Windows Azure Search Service. */ -export interface AzurePostgreSqlLinkedService { +export interface AzureSearchLinkedService { /** * Polymorphic Discriminator */ - type: "AzurePostgreSql"; + type: "AzureSearch"; /** * The integration runtime reference. */ @@ -5039,13 +5014,13 @@ export interface AzurePostgreSqlLinkedService { */ annotations?: any[]; /** - * An ODBC connection string. Type: string, SecureString or AzureKeyVaultSecretReference. + * URL for Azure Search service. Type: string (or Expression with resultType string). */ - connectionString?: any; + url: any; /** - * The Azure key vault secret reference of password in connection string. + * Admin Key for Azure Search service */ - password?: AzureKeyVaultSecretReference; + key?: SecretBaseUnion; /** * The encrypted credential used for authentication. Credentials are encrypted using the * integration runtime credential manager. Type: string (or Expression with resultType string). @@ -5054,13 +5029,13 @@ export interface AzurePostgreSqlLinkedService { } /** - * Amazon Marketplace Web Service linked service. + * Custom linked service. */ -export interface AmazonMWSLinkedService { +export interface CustomDataSourceLinkedService { /** * Polymorphic Discriminator */ - type: "AmazonMWS"; + type: "CustomDataSource"; /** * The integration runtime reference. */ @@ -5078,45 +5053,58 @@ export interface AmazonMWSLinkedService { */ annotations?: any[]; /** - * The endpoint of the Amazon MWS server, (i.e. mws.amazonservices.com) + * Custom linked service properties. */ - endpoint: any; + typeProperties: any; +} + +/** + * Linked service for Amazon Redshift. + */ +export interface AmazonRedshiftLinkedService { /** - * The Amazon Marketplace ID you want to retrieve data from. To retrieve data from multiple - * Marketplace IDs, separate them with a comma (,). (i.e. A2EUQ1WTGCTBG2) + * Polymorphic Discriminator */ - marketplaceID: any; + type: "AmazonRedshift"; /** - * The Amazon seller ID. + * The integration runtime reference. */ - sellerID: any; + connectVia?: IntegrationRuntimeReference; /** - * The Amazon MWS authentication token. + * Linked service description. */ - mwsAuthToken?: SecretBaseUnion; + description?: string; /** - * The access key id used to access data. + * Parameters for linked service. */ - accessKeyId: any; + parameters?: { [propertyName: string]: ParameterSpecification }; /** - * The secret key used to access data. + * List of tags that can be used for describing the linked service. */ - secretKey?: SecretBaseUnion; + annotations?: any[]; /** - * Specifies whether the data source endpoints are encrypted using HTTPS. The default value is - * true. + * The name of the Amazon Redshift server. Type: string (or Expression with resultType string). */ - useEncryptedEndpoints?: any; + server: any; /** - * Specifies whether to require the host name in the server's certificate to match the host name - * of the server when connecting over SSL. The default value is true. + * The username of the Amazon Redshift source. Type: string (or Expression with resultType + * string). */ - useHostVerification?: any; + username?: any; + /** + * The password of the Amazon Redshift source. + */ + password?: SecretBaseUnion; + /** + * The database name of the Amazon Redshift source. Type: string (or Expression with resultType + * string). + */ + database: any; /** - * Specifies whether to verify the identity of the server when connecting over SSL. The default - * value is true. + * The TCP port number that the Amazon Redshift server uses to listen for client connections. The + * default value is 5439. Type: integer (or Expression with resultType integer). */ - usePeerVerification?: any; + port?: any; /** * The encrypted credential used for authentication. Credentials are encrypted using the * integration runtime credential manager. Type: string (or Expression with resultType string). @@ -5125,13 +5113,13 @@ export interface AmazonMWSLinkedService { } /** - * SAP HANA Linked Service. + * Linked service for Amazon S3. */ -export interface SapHanaLinkedService { +export interface AmazonS3LinkedService { /** * Polymorphic Discriminator */ - type: "SapHana"; + type: "AmazonS3"; /** * The integration runtime reference. */ @@ -5149,26 +5137,20 @@ export interface SapHanaLinkedService { */ annotations?: any[]; /** - * SAP HANA ODBC connection string. Type: string, SecureString or AzureKeyVaultSecretReference. - */ - connectionString?: any; - /** - * Host name of the SAP HANA server. Type: string (or Expression with resultType string). - */ - server: any; - /** - * The authentication type to be used to connect to the SAP HANA server. Possible values include: - * 'Basic', 'Windows' + * The access key identifier of the Amazon S3 Identity and Access Management (IAM) user. Type: + * string (or Expression with resultType string). */ - authenticationType?: SapHanaAuthenticationType; + accessKeyId?: any; /** - * Username to access the SAP HANA server. Type: string (or Expression with resultType string). + * The secret access key of the Amazon S3 Identity and Access Management (IAM) user. */ - userName?: any; + secretAccessKey?: SecretBaseUnion; /** - * Password to access the SAP HANA server. + * This value specifies the endpoint to access with the S3 Connector. This is an optional + * property; change it only if you want to try a different service endpoint or want to switch + * between https and http. Type: string (or Expression with resultType string). */ - password?: SecretBaseUnion; + serviceUrl?: any; /** * The encrypted credential used for authentication. Credentials are encrypted using the * integration runtime credential manager. Type: string (or Expression with resultType string). @@ -5177,13 +5159,13 @@ export interface SapHanaLinkedService { } /** - * SAP Business Warehouse Linked Service. + * Rest Service linked service. */ -export interface SapBWLinkedService { +export interface RestServiceLinkedService { /** * Polymorphic Discriminator */ - type: "SapBW"; + type: "RestService"; /** * The integration runtime reference. */ @@ -5201,27 +5183,44 @@ export interface SapBWLinkedService { */ annotations?: any[]; /** - * Host name of the SAP BW instance. Type: string (or Expression with resultType string). + * The base URL of the REST service. */ - server: any; + url: any; /** - * System number of the BW system. (Usually a two-digit decimal number represented as a string.) - * Type: string (or Expression with resultType string). + * Whether to validate server side SSL certificate when connecting to the endpoint.The default + * value is true. Type: boolean (or Expression with resultType boolean). */ - systemNumber: any; + enableServerCertificateValidation?: any; /** - * Client ID of the client on the BW system. (Usually a three-digit decimal number represented as - * a string) Type: string (or Expression with resultType string). + * Type of authentication used to connect to the REST service. Possible values include: + * 'Anonymous', 'Basic', 'AadServicePrincipal', 'ManagedServiceIdentity' */ - clientId: any; + authenticationType: RestServiceAuthenticationType; /** - * Username to access the SAP BW server. Type: string (or Expression with resultType string). + * The user name used in Basic authentication type. */ userName?: any; /** - * Password to access the SAP BW server. + * The password used in Basic authentication type. */ password?: SecretBaseUnion; + /** + * The application's client ID used in AadServicePrincipal authentication type. + */ + servicePrincipalId?: any; + /** + * The application's key used in AadServicePrincipal authentication type. + */ + servicePrincipalKey?: SecretBaseUnion; + /** + * The tenant information (domain name or tenant ID) used in AadServicePrincipal authentication + * type under which your application resides. + */ + tenant?: any; + /** + * The resource you are requesting authorization to use. + */ + aadResourceId?: any; /** * The encrypted credential used for authentication. Credentials are encrypted using the * integration runtime credential manager. Type: string (or Expression with resultType string). @@ -5230,13 +5229,13 @@ export interface SapBWLinkedService { } /** - * A linked service for an SSH File Transfer Protocol (SFTP) server. + * SAP Business Warehouse Open Hub Destination Linked Service. */ -export interface SftpServerLinkedService { +export interface SapOpenHubLinkedService { /** * Polymorphic Discriminator */ - type: "Sftp"; + type: "SapOpenHub"; /** * The integration runtime reference. */ @@ -5254,26 +5253,33 @@ export interface SftpServerLinkedService { */ annotations?: any[]; /** - * The SFTP server host name. Type: string (or Expression with resultType string). + * Host name of the SAP BW instance where the open hub destination is located. Type: string (or + * Expression with resultType string). */ - host: any; + server: any; /** - * The TCP port number that the SFTP server uses to listen for client connections. Default value - * is 22. Type: integer (or Expression with resultType integer), minimum: 0. + * System number of the BW system where the open hub destination is located. (Usually a two-digit + * decimal number represented as a string.) Type: string (or Expression with resultType string). */ - port?: any; + systemNumber: any; /** - * The authentication type to be used to connect to the FTP server. Possible values include: - * 'Basic', 'SshPublicKey' + * Client ID of the client on the BW system where the open hub destination is located. (Usually a + * three-digit decimal number represented as a string) Type: string (or Expression with + * resultType string). */ - authenticationType?: SftpAuthenticationType; + clientId: any; /** - * The username used to log on to the SFTP server. Type: string (or Expression with resultType - * string). + * Language of the BW system where the open hub destination is located. The default value is EN. + * Type: string (or Expression with resultType string). + */ + language?: any; + /** + * Username to access the SAP BW server where the open hub destination is located. Type: string + * (or Expression with resultType string). */ userName?: any; /** - * Password to logon the SFTP server for Basic authentication. + * Password to access the SAP BW server where the open hub destination is located. */ password?: SecretBaseUnion; /** @@ -5281,43 +5287,16 @@ export interface SftpServerLinkedService { * integration runtime credential manager. Type: string (or Expression with resultType string). */ encryptedCredential?: any; - /** - * The SSH private key file path for SshPublicKey authentication. Only valid for on-premises - * copy. For on-premises copy with SshPublicKey authentication, either PrivateKeyPath or - * PrivateKeyContent should be specified. SSH private key should be OpenSSH format. Type: string - * (or Expression with resultType string). - */ - privateKeyPath?: any; - /** - * Base64 encoded SSH private key content for SshPublicKey authentication. For on-premises copy - * with SshPublicKey authentication, either PrivateKeyPath or PrivateKeyContent should be - * specified. SSH private key should be OpenSSH format. - */ - privateKeyContent?: SecretBaseUnion; - /** - * The password to decrypt the SSH private key if the SSH private key is encrypted. - */ - passPhrase?: SecretBaseUnion; - /** - * If true, skip the SSH host key validation. Default value is false. Type: boolean (or - * Expression with resultType boolean). - */ - skipHostKeyValidation?: any; - /** - * The host key finger-print of the SFTP server. When SkipHostKeyValidation is false, - * HostKeyFingerprint should be specified. Type: string (or Expression with resultType string). - */ - hostKeyFingerprint?: any; } /** - * A FTP server Linked Service. + * Linked service for SAP ERP Central Component(SAP ECC). */ -export interface FtpServerLinkedService { +export interface SapEccLinkedService { /** * Polymorphic Discriminator */ - type: "FtpServer"; + type: "SapEcc"; /** * The integration runtime reference. */ @@ -5335,52 +5314,35 @@ export interface FtpServerLinkedService { */ annotations?: any[]; /** - * Host name of the FTP server. Type: string (or Expression with resultType string). - */ - host: any; - /** - * The TCP port number that the FTP server uses to listen for client connections. Default value - * is 21. Type: integer (or Expression with resultType integer), minimum: 0. - */ - port?: any; - /** - * The authentication type to be used to connect to the FTP server. Possible values include: - * 'Basic', 'Anonymous' + * The URL of SAP ECC OData API. For example, + * '[https://hostname:port/sap/opu/odata/sap/servicename/]'. Type: string (or Expression with + * resultType string). */ - authenticationType?: FtpAuthenticationType; + url: string; /** - * Username to logon the FTP server. Type: string (or Expression with resultType string). + * The username for Basic authentication. Type: string (or Expression with resultType string). */ - userName?: any; + username?: string; /** - * Password to logon the FTP server. + * The password for Basic authentication. */ password?: SecretBaseUnion; /** * The encrypted credential used for authentication. Credentials are encrypted using the - * integration runtime credential manager. Type: string (or Expression with resultType string). - */ - encryptedCredential?: any; - /** - * If true, connect to the FTP server over SSL/TLS channel. Default value is true. Type: boolean - * (or Expression with resultType boolean). - */ - enableSsl?: any; - /** - * If true, validate the FTP server SSL certificate when connect over SSL/TLS channel. Default - * value is true. Type: boolean (or Expression with resultType boolean). + * integration runtime credential manager. Either encryptedCredential or username/password must + * be provided. Type: string (or Expression with resultType string). */ - enableServerCertificateValidation?: any; + encryptedCredential?: string; } /** - * Linked service for an HTTP source. + * Linked service for SAP Cloud for Customer. */ -export interface HttpLinkedService { +export interface SapCloudForCustomerLinkedService { /** * Polymorphic Discriminator */ - type: "HttpServer"; + type: "SapCloudForCustomer"; /** * The integration runtime reference. */ @@ -5398,58 +5360,35 @@ export interface HttpLinkedService { */ annotations?: any[]; /** - * The base URL of the HTTP endpoint, e.g. http://www.microsoft.com. Type: string (or Expression - * with resultType string). + * The URL of SAP Cloud for Customer OData API. For example, + * '[https://[tenantname].crm.ondemand.com/sap/c4c/odata/v1]'. Type: string (or Expression with + * resultType string). */ url: any; /** - * The authentication type to be used to connect to the HTTP server. Possible values include: - * 'Basic', 'Anonymous', 'Digest', 'Windows', 'ClientCertificate' - */ - authenticationType?: HttpAuthenticationType; - /** - * User name for Basic, Digest, or Windows authentication. Type: string (or Expression with - * resultType string). + * The username for Basic authentication. Type: string (or Expression with resultType string). */ - userName?: any; + username?: any; /** - * Password for Basic, Digest, Windows, or ClientCertificate with EmbeddedCertData - * authentication. + * The password for Basic authentication. */ password?: SecretBaseUnion; - /** - * Base64 encoded certificate data for ClientCertificate authentication. For on-premises copy - * with ClientCertificate authentication, either CertThumbprint or EmbeddedCertData/Password - * should be specified. Type: string (or Expression with resultType string). - */ - embeddedCertData?: any; - /** - * Thumbprint of certificate for ClientCertificate authentication. Only valid for on-premises - * copy. For on-premises copy with ClientCertificate authentication, either CertThumbprint or - * EmbeddedCertData/Password should be specified. Type: string (or Expression with resultType - * string). - */ - certThumbprint?: any; /** * The encrypted credential used for authentication. Credentials are encrypted using the - * integration runtime credential manager. Type: string (or Expression with resultType string). + * integration runtime credential manager. Either encryptedCredential or username/password must + * be provided. Type: string (or Expression with resultType string). */ encryptedCredential?: any; - /** - * If true, validate the HTTPS server SSL certificate. Default value is true. Type: boolean (or - * Expression with resultType boolean). - */ - enableServerCertificateValidation?: any; } /** - * Linked service for Windows Azure Search Service. + * Linked service for Salesforce Service Cloud. */ -export interface AzureSearchLinkedService { +export interface SalesforceServiceCloudLinkedService { /** * Polymorphic Discriminator */ - type: "AzureSearch"; + type: "SalesforceServiceCloud"; /** * The integration runtime reference. */ @@ -5459,21 +5398,42 @@ export interface AzureSearchLinkedService { */ description?: string; /** - * Parameters for linked service. + * Parameters for linked service. + */ + parameters?: { [propertyName: string]: ParameterSpecification }; + /** + * List of tags that can be used for describing the linked service. + */ + annotations?: any[]; + /** + * The URL of Salesforce Service Cloud instance. Default is 'https://login.salesforce.com'. To + * copy data from sandbox, specify 'https://test.salesforce.com'. To copy data from custom + * domain, specify, for example, 'https://[domain].my.salesforce.com'. Type: string (or + * Expression with resultType string). + */ + environmentUrl?: any; + /** + * The username for Basic authentication of the Salesforce instance. Type: string (or Expression + * with resultType string). + */ + username?: any; + /** + * The password for Basic authentication of the Salesforce instance. */ - parameters?: { [propertyName: string]: ParameterSpecification }; + password?: SecretBaseUnion; /** - * List of tags that can be used for describing the linked service. + * The security token is optional to remotely access Salesforce instance. */ - annotations?: any[]; + securityToken?: SecretBaseUnion; /** - * URL for Azure Search service. Type: string (or Expression with resultType string). + * The Salesforce API version used in ADF. Type: string (or Expression with resultType string). */ - url: any; + apiVersion?: any; /** - * Admin Key for Azure Search service + * Extended properties appended to the connection string. Type: string (or Expression with + * resultType string). */ - key?: SecretBaseUnion; + extendedProperties?: any; /** * The encrypted credential used for authentication. Credentials are encrypted using the * integration runtime credential manager. Type: string (or Expression with resultType string). @@ -5482,13 +5442,13 @@ export interface AzureSearchLinkedService { } /** - * Custom linked service. + * Linked service for Salesforce. */ -export interface CustomDataSourceLinkedService { +export interface SalesforceLinkedService { /** * Polymorphic Discriminator */ - type: "CustomDataSource"; + type: "Salesforce"; /** * The integration runtime reference. */ @@ -5506,19 +5466,44 @@ export interface CustomDataSourceLinkedService { */ annotations?: any[]; /** - * Custom linked service properties. + * The URL of Salesforce instance. Default is 'https://login.salesforce.com'. To copy data from + * sandbox, specify 'https://test.salesforce.com'. To copy data from custom domain, specify, for + * example, 'https://[domain].my.salesforce.com'. Type: string (or Expression with resultType + * string). */ - typeProperties: any; + environmentUrl?: any; + /** + * The username for Basic authentication of the Salesforce instance. Type: string (or Expression + * with resultType string). + */ + username?: any; + /** + * The password for Basic authentication of the Salesforce instance. + */ + password?: SecretBaseUnion; + /** + * The security token is optional to remotely access Salesforce instance. + */ + securityToken?: SecretBaseUnion; + /** + * The Salesforce API version used in ADF. Type: string (or Expression with resultType string). + */ + apiVersion?: any; + /** + * The encrypted credential used for authentication. Credentials are encrypted using the + * integration runtime credential manager. Type: string (or Expression with resultType string). + */ + encryptedCredential?: any; } /** - * Linked service for Amazon Redshift. + * Office365 linked service. */ -export interface AmazonRedshiftLinkedService { +export interface Office365LinkedService { /** * Polymorphic Discriminator */ - type: "AmazonRedshift"; + type: "Office365"; /** * The integration runtime reference. */ @@ -5536,28 +5521,23 @@ export interface AmazonRedshiftLinkedService { */ annotations?: any[]; /** - * The name of the Amazon Redshift server. Type: string (or Expression with resultType string). - */ - server: any; - /** - * The username of the Amazon Redshift source. Type: string (or Expression with resultType - * string). + * Azure tenant ID to which the Office 365 account belongs. Type: string (or Expression with + * resultType string). */ - username?: any; + office365TenantId: any; /** - * The password of the Amazon Redshift source. + * Specify the tenant information under which your Azure AD web application resides. Type: string + * (or Expression with resultType string). */ - password?: SecretBaseUnion; + servicePrincipalTenantId: any; /** - * The database name of the Amazon Redshift source. Type: string (or Expression with resultType - * string). + * Specify the application's client ID. Type: string (or Expression with resultType string). */ - database: any; + servicePrincipalId: any; /** - * The TCP port number that the Amazon Redshift server uses to listen for client connections. The - * default value is 5439. Type: integer (or Expression with resultType integer). + * Specify the application's key. */ - port?: any; + servicePrincipalKey: SecretBaseUnion; /** * The encrypted credential used for authentication. Credentials are encrypted using the * integration runtime credential manager. Type: string (or Expression with resultType string). @@ -5566,13 +5546,13 @@ export interface AmazonRedshiftLinkedService { } /** - * Linked service for Amazon S3. + * Azure Data Lake Storage Gen2 linked service. */ -export interface AmazonS3LinkedService { +export interface AzureBlobFSLinkedService { /** * Polymorphic Discriminator */ - type: "AmazonS3"; + type: "AzureBlobFS"; /** * The integration runtime reference. */ @@ -5590,20 +5570,30 @@ export interface AmazonS3LinkedService { */ annotations?: any[]; /** - * The access key identifier of the Amazon S3 Identity and Access Management (IAM) user. Type: - * string (or Expression with resultType string). + * Endpoint for the Azure Data Lake Storage Gen2 service. Type: string (or Expression with + * resultType string). */ - accessKeyId?: any; + url: any; /** - * The secret access key of the Amazon S3 Identity and Access Management (IAM) user. + * Account key for the Azure Data Lake Storage Gen2 service. Type: string (or Expression with + * resultType string). */ - secretAccessKey?: SecretBaseUnion; + accountKey?: any; /** - * This value specifies the endpoint to access with the S3 Connector. This is an optional - * property; change it only if you want to try a different service endpoint or want to switch - * between https and http. Type: string (or Expression with resultType string). + * The ID of the application used to authenticate against the Azure Data Lake Storage Gen2 + * account. Type: string (or Expression with resultType string). */ - serviceUrl?: any; + servicePrincipalId?: any; + /** + * The Key of the application used to authenticate against the Azure Data Lake Storage Gen2 + * account. + */ + servicePrincipalKey?: SecretBaseUnion; + /** + * The name or ID of the tenant to which the service principal belongs. Type: string (or + * Expression with resultType string). + */ + tenant?: any; /** * The encrypted credential used for authentication. Credentials are encrypted using the * integration runtime credential manager. Type: string (or Expression with resultType string). @@ -5612,13 +5602,13 @@ export interface AmazonS3LinkedService { } /** - * Rest Service linked service. + * Azure Data Lake Store linked service. */ -export interface RestServiceLinkedService { +export interface AzureDataLakeStoreLinkedService { /** * Polymorphic Discriminator */ - type: "RestService"; + type: "AzureDataLakeStore"; /** * The integration runtime reference. */ @@ -5636,44 +5626,37 @@ export interface RestServiceLinkedService { */ annotations?: any[]; /** - * The base URL of the REST service. - */ - url: any; - /** - * Whether to validate server side SSL certificate when connecting to the endpoint.The default - * value is true. Type: boolean (or Expression with resultType boolean). - */ - enableServerCertificateValidation?: any; - /** - * Type of authentication used to connect to the REST service. Possible values include: - * 'Anonymous', 'Basic', 'AadServicePrincipal', 'ManagedServiceIdentity' + * Data Lake Store service URI. Type: string (or Expression with resultType string). */ - authenticationType: RestServiceAuthenticationType; + dataLakeStoreUri: any; /** - * The user name used in Basic authentication type. + * The ID of the application used to authenticate against the Azure Data Lake Store account. + * Type: string (or Expression with resultType string). */ - userName?: any; + servicePrincipalId?: any; /** - * The password used in Basic authentication type. + * The Key of the application used to authenticate against the Azure Data Lake Store account. */ - password?: SecretBaseUnion; + servicePrincipalKey?: SecretBaseUnion; /** - * The application's client ID used in AadServicePrincipal authentication type. + * The name or ID of the tenant to which the service principal belongs. Type: string (or + * Expression with resultType string). */ - servicePrincipalId?: any; + tenant?: any; /** - * The application's key used in AadServicePrincipal authentication type. + * Data Lake Store account name. Type: string (or Expression with resultType string). */ - servicePrincipalKey?: SecretBaseUnion; + accountName?: any; /** - * The tenant information (domain name or tenant ID) used in AadServicePrincipal authentication - * type under which your application resides. + * Data Lake Store account subscription ID (if different from Data Factory account). Type: string + * (or Expression with resultType string). */ - tenant?: any; + subscriptionId?: any; /** - * The resource you are requesting authorization to use. + * Data Lake Store account resource group name (if different from Data Factory account). Type: + * string (or Expression with resultType string). */ - aadResourceId?: any; + resourceGroupName?: any; /** * The encrypted credential used for authentication. Credentials are encrypted using the * integration runtime credential manager. Type: string (or Expression with resultType string). @@ -5682,13 +5665,13 @@ export interface RestServiceLinkedService { } /** - * SAP Business Warehouse Open Hub Destination Linked Service. + * Linked service for CosmosDB (MongoDB API) data source. */ -export interface SapOpenHubLinkedService { +export interface CosmosDbMongoDbApiLinkedService { /** * Polymorphic Discriminator */ - type: "SapOpenHub"; + type: "CosmosDbMongoDbApi"; /** * The integration runtime reference. */ @@ -5706,50 +5689,61 @@ export interface SapOpenHubLinkedService { */ annotations?: any[]; /** - * Host name of the SAP BW instance where the open hub destination is located. Type: string (or + * The CosmosDB (MongoDB API) connection string. Type: string, SecureString or + * AzureKeyVaultSecretReference. Type: string, SecureString or AzureKeyVaultSecretReference. + */ + connectionString: any; + /** + * The name of the CosmosDB (MongoDB API) database that you want to access. Type: string (or * Expression with resultType string). */ - server: any; + database: any; +} + +/** + * Linked service for MongoDB data source. + */ +export interface MongoDbV2LinkedService { /** - * System number of the BW system where the open hub destination is located. (Usually a two-digit - * decimal number represented as a string.) Type: string (or Expression with resultType string). + * Polymorphic Discriminator */ - systemNumber: any; + type: "MongoDbV2"; /** - * Client ID of the client on the BW system where the open hub destination is located. (Usually a - * three-digit decimal number represented as a string) Type: string (or Expression with - * resultType string). + * The integration runtime reference. */ - clientId: any; + connectVia?: IntegrationRuntimeReference; /** - * Language of the BW system where the open hub destination is located. The default value is EN. - * Type: string (or Expression with resultType string). + * Linked service description. */ - language?: any; + description?: string; /** - * Username to access the SAP BW server where the open hub destination is located. Type: string - * (or Expression with resultType string). + * Parameters for linked service. */ - userName?: any; + parameters?: { [propertyName: string]: ParameterSpecification }; /** - * Password to access the SAP BW server where the open hub destination is located. + * List of tags that can be used for describing the linked service. */ - password?: SecretBaseUnion; + annotations?: any[]; /** - * The encrypted credential used for authentication. Credentials are encrypted using the - * integration runtime credential manager. Type: string (or Expression with resultType string). + * The MongoDB connection string. Type: string, SecureString or AzureKeyVaultSecretReference. + * Type: string, SecureString or AzureKeyVaultSecretReference. */ - encryptedCredential?: any; + connectionString: any; + /** + * The name of the MongoDB database that you want to access. Type: string (or Expression with + * resultType string). + */ + database: any; } /** - * Linked service for SAP ERP Central Component(SAP ECC). + * Linked service for MongoDb data source. */ -export interface SapEccLinkedService { +export interface MongoDbLinkedService { /** * Polymorphic Discriminator */ - type: "SapEcc"; + type: "MongoDb"; /** * The integration runtime reference. */ @@ -5767,35 +5761,63 @@ export interface SapEccLinkedService { */ annotations?: any[]; /** - * The URL of SAP ECC OData API. For example, - * '[https://hostname:port/sap/opu/odata/sap/servicename/]'. Type: string (or Expression with + * The IP address or server name of the MongoDB server. Type: string (or Expression with * resultType string). */ - url: string; + server: any; + /** + * The authentication type to be used to connect to the MongoDB database. Possible values + * include: 'Basic', 'Anonymous' + */ + authenticationType?: MongoDbAuthenticationType; + /** + * The name of the MongoDB database that you want to access. Type: string (or Expression with + * resultType string). + */ + databaseName: any; + /** + * Username for authentication. Type: string (or Expression with resultType string). + */ + username?: any; + /** + * Password for authentication. + */ + password?: SecretBaseUnion; + /** + * Database to verify the username and password. Type: string (or Expression with resultType + * string). + */ + authSource?: any; + /** + * The TCP port number that the MongoDB server uses to listen for client connections. The default + * value is 27017. Type: integer (or Expression with resultType integer), minimum: 0. + */ + port?: any; /** - * The username for Basic authentication. Type: string (or Expression with resultType string). + * Specifies whether the connections to the server are encrypted using SSL. The default value is + * false. Type: boolean (or Expression with resultType boolean). */ - username?: string; + enableSsl?: any; /** - * The password for Basic authentication. + * Specifies whether to allow self-signed certificates from the server. The default value is + * false. Type: boolean (or Expression with resultType boolean). */ - password?: SecretBaseUnion; + allowSelfSignedServerCert?: any; /** * The encrypted credential used for authentication. Credentials are encrypted using the - * integration runtime credential manager. Either encryptedCredential or username/password must - * be provided. Type: string (or Expression with resultType string). + * integration runtime credential manager. Type: string (or Expression with resultType string). */ - encryptedCredential?: string; + encryptedCredential?: any; } /** - * Linked service for SAP Cloud for Customer. + * Linked service for Cassandra data source. */ -export interface SapCloudForCustomerLinkedService { +export interface CassandraLinkedService { /** * Polymorphic Discriminator */ - type: "SapCloudForCustomer"; + type: "Cassandra"; /** * The integration runtime reference. */ @@ -5813,91 +5835,125 @@ export interface SapCloudForCustomerLinkedService { */ annotations?: any[]; /** - * The URL of SAP Cloud for Customer OData API. For example, - * '[https://[tenantname].crm.ondemand.com/sap/c4c/odata/v1]'. Type: string (or Expression with - * resultType string). + * Host name for connection. Type: string (or Expression with resultType string). */ - url: any; + host: any; /** - * The username for Basic authentication. Type: string (or Expression with resultType string). + * AuthenticationType to be used for connection. Type: string (or Expression with resultType + * string). + */ + authenticationType?: any; + /** + * The port for the connection. Type: integer (or Expression with resultType integer). + */ + port?: any; + /** + * Username for authentication. Type: string (or Expression with resultType string). */ username?: any; /** - * The password for Basic authentication. + * Password for authentication. */ password?: SecretBaseUnion; /** * The encrypted credential used for authentication. Credentials are encrypted using the - * integration runtime credential manager. Either encryptedCredential or username/password must - * be provided. Type: string (or Expression with resultType string). + * integration runtime credential manager. Type: string (or Expression with resultType string). */ encryptedCredential?: any; } /** - * Linked service for Salesforce Service Cloud. + * Contains the possible cases for WebLinkedServiceTypeProperties. */ -export interface SalesforceServiceCloudLinkedService { +export type WebLinkedServiceTypePropertiesUnion = WebLinkedServiceTypeProperties | WebClientCertificateAuthentication | WebBasicAuthentication | WebAnonymousAuthentication; + +/** + * Base definition of WebLinkedServiceTypeProperties, this typeProperties is polymorphic based on + * authenticationType, so not flattened in SDK models. + */ +export interface WebLinkedServiceTypeProperties { /** * Polymorphic Discriminator */ - type: "SalesforceServiceCloud"; + authenticationType: "WebLinkedServiceTypeProperties"; /** - * The integration runtime reference. + * The URL of the web service endpoint, e.g. http://www.microsoft.com . Type: string (or + * Expression with resultType string). */ - connectVia?: IntegrationRuntimeReference; + url: any; +} + +/** + * A WebLinkedService that uses client certificate based authentication to communicate with an HTTP + * endpoint. This scheme follows mutual authentication; the server must also provide valid + * credentials to the client. + */ +export interface WebClientCertificateAuthentication { /** - * Linked service description. + * Polymorphic Discriminator */ - description?: string; + authenticationType: "ClientCertificate"; /** - * Parameters for linked service. + * The URL of the web service endpoint, e.g. http://www.microsoft.com . Type: string (or + * Expression with resultType string). */ - parameters?: { [propertyName: string]: ParameterSpecification }; + url: any; /** - * List of tags that can be used for describing the linked service. + * Base64-encoded contents of a PFX file. */ - annotations?: any[]; + pfx: SecretBaseUnion; /** - * The URL of Salesforce Service Cloud instance. Default is 'https://login.salesforce.com'. To - * copy data from sandbox, specify 'https://test.salesforce.com'. To copy data from custom - * domain, specify, for example, 'https://[domain].my.salesforce.com'. Type: string (or - * Expression with resultType string). + * Password for the PFX file. */ - environmentUrl?: any; + password: SecretBaseUnion; +} + +/** + * A WebLinkedService that uses basic authentication to communicate with an HTTP endpoint. + */ +export interface WebBasicAuthentication { /** - * The username for Basic authentication of the Salesforce instance. Type: string (or Expression - * with resultType string). + * Polymorphic Discriminator */ - username?: any; + authenticationType: "Basic"; /** - * The password for Basic authentication of the Salesforce instance. + * The URL of the web service endpoint, e.g. http://www.microsoft.com . Type: string (or + * Expression with resultType string). */ - password?: SecretBaseUnion; + url: any; /** - * The security token is required to remotely access Salesforce instance. + * User name for Basic authentication. Type: string (or Expression with resultType string). */ - securityToken?: SecretBaseUnion; + username: any; /** - * Extended properties appended to the connection string. Type: string (or Expression with - * resultType string). + * The password for Basic authentication. */ - extendedProperties?: any; + password: SecretBaseUnion; +} + +/** + * A WebLinkedService that uses anonymous authentication to communicate with an HTTP endpoint. + */ +export interface WebAnonymousAuthentication { /** - * The encrypted credential used for authentication. Credentials are encrypted using the - * integration runtime credential manager. Type: string (or Expression with resultType string). + * Polymorphic Discriminator */ - encryptedCredential?: any; + authenticationType: "Anonymous"; + /** + * The URL of the web service endpoint, e.g. http://www.microsoft.com . Type: string (or + * Expression with resultType string). + */ + url: any; } /** - * Linked service for Salesforce. + * Web linked service. */ -export interface SalesforceLinkedService { +export interface WebLinkedService { /** * Polymorphic Discriminator */ - type: "Salesforce"; + type: "Web"; /** * The integration runtime reference. */ @@ -5915,40 +5971,19 @@ export interface SalesforceLinkedService { */ annotations?: any[]; /** - * The URL of Salesforce instance. Default is 'https://login.salesforce.com'. To copy data from - * sandbox, specify 'https://test.salesforce.com'. To copy data from custom domain, specify, for - * example, 'https://[domain].my.salesforce.com'. Type: string (or Expression with resultType - * string). - */ - environmentUrl?: any; - /** - * The username for Basic authentication of the Salesforce instance. Type: string (or Expression - * with resultType string). - */ - username?: any; - /** - * The password for Basic authentication of the Salesforce instance. - */ - password?: SecretBaseUnion; - /** - * The security token is required to remotely access Salesforce instance. - */ - securityToken?: SecretBaseUnion; - /** - * The encrypted credential used for authentication. Credentials are encrypted using the - * integration runtime credential manager. Type: string (or Expression with resultType string). + * Web linked service properties. */ - encryptedCredential?: any; + typeProperties: WebLinkedServiceTypePropertiesUnion; } /** - * Office365 linked service. + * Open Data Protocol (OData) linked service. */ -export interface Office365LinkedService { +export interface ODataLinkedService { /** * Polymorphic Discriminator */ - type: "Office365"; + type: "OData"; /** * The integration runtime reference. */ @@ -5966,23 +6001,57 @@ export interface Office365LinkedService { */ annotations?: any[]; /** - * Azure tenant ID to which the Office 365 account belongs. Type: string (or Expression with - * resultType string). + * The URL of the OData service endpoint. Type: string (or Expression with resultType string). */ - office365TenantId: any; + url: any; /** - * Specify the tenant information under which your Azure AD web application resides. Type: string - * (or Expression with resultType string). + * Type of authentication used to connect to the OData service. Possible values include: 'Basic', + * 'Anonymous', 'Windows', 'AadServicePrincipal', 'ManagedServiceIdentity' */ - servicePrincipalTenantId: any; + authenticationType?: ODataAuthenticationType; /** - * Specify the application's client ID. Type: string (or Expression with resultType string). + * User name of the OData service. Type: string (or Expression with resultType string). */ - servicePrincipalId: any; + userName?: any; /** - * Specify the application's key. + * Password of the OData service. */ - servicePrincipalKey: SecretBaseUnion; + password?: SecretBaseUnion; + /** + * Specify the tenant information (domain name or tenant ID) under which your application + * resides. Type: string (or Expression with resultType string). + */ + tenant?: any; + /** + * Specify the application id of your application registered in Azure Active Directory. Type: + * string (or Expression with resultType string). + */ + servicePrincipalId?: any; + /** + * Specify the resource you are requesting authorization to use Directory. Type: string (or + * Expression with resultType string). + */ + aadResourceId?: any; + /** + * Specify the credential type (key or cert) is used for service principal. Possible values + * include: 'ServicePrincipalKey', 'ServicePrincipalCert' + */ + aadServicePrincipalCredentialType?: ODataAadServicePrincipalCredentialType; + /** + * Specify the secret of your application registered in Azure Active Directory. Type: string (or + * Expression with resultType string). + */ + servicePrincipalKey?: SecretBaseUnion; + /** + * Specify the base64 encoded certificate of your application registered in Azure Active + * Directory. Type: string (or Expression with resultType string). + */ + servicePrincipalEmbeddedCert?: SecretBaseUnion; + /** + * Specify the password of your certificate if your certificate has a password and you are using + * AadServicePrincipal authentication. Type: string (or Expression with resultType string). + */ + servicePrincipalEmbeddedCertPassword?: SecretBaseUnion; /** * The encrypted credential used for authentication. Credentials are encrypted using the * integration runtime credential manager. Type: string (or Expression with resultType string). @@ -5991,13 +6060,13 @@ export interface Office365LinkedService { } /** - * Azure Data Lake Storage Gen2 linked service. + * Hadoop Distributed File System (HDFS) linked service. */ -export interface AzureBlobFSLinkedService { +export interface HdfsLinkedService { /** * Polymorphic Discriminator */ - type: "AzureBlobFS"; + type: "Hdfs"; /** * The integration runtime reference. */ @@ -6015,45 +6084,38 @@ export interface AzureBlobFSLinkedService { */ annotations?: any[]; /** - * Endpoint for the Azure Data Lake Storage Gen2 service. Type: string (or Expression with - * resultType string). + * The URL of the HDFS service endpoint, e.g. http://myhostname:50070/webhdfs/v1 . Type: string + * (or Expression with resultType string). */ url: any; /** - * Account key for the Azure Data Lake Storage Gen2 service. Type: string (or Expression with - * resultType string). - */ - accountKey?: any; - /** - * The ID of the application used to authenticate against the Azure Data Lake Storage Gen2 - * account. Type: string (or Expression with resultType string). + * Type of authentication used to connect to the HDFS. Possible values are: Anonymous and + * Windows. Type: string (or Expression with resultType string). */ - servicePrincipalId?: any; + authenticationType?: any; /** - * The Key of the application used to authenticate against the Azure Data Lake Storage Gen2 - * account. + * The encrypted credential used for authentication. Credentials are encrypted using the + * integration runtime credential manager. Type: string (or Expression with resultType string). */ - servicePrincipalKey?: SecretBaseUnion; + encryptedCredential?: any; /** - * The name or ID of the tenant to which the service principal belongs. Type: string (or - * Expression with resultType string). + * User name for Windows authentication. Type: string (or Expression with resultType string). */ - tenant?: any; + userName?: any; /** - * The encrypted credential used for authentication. Credentials are encrypted using the - * integration runtime credential manager. Type: string (or Expression with resultType string). + * Password for Windows authentication. */ - encryptedCredential?: any; + password?: SecretBaseUnion; } /** - * Azure Data Lake Store linked service. + * Microsoft Access linked service. */ -export interface AzureDataLakeStoreLinkedService { +export interface MicrosoftAccessLinkedService { /** * Polymorphic Discriminator */ - type: "AzureDataLakeStore"; + type: "MicrosoftAccess"; /** * The integration runtime reference. */ @@ -6071,37 +6133,28 @@ export interface AzureDataLakeStoreLinkedService { */ annotations?: any[]; /** - * Data Lake Store service URI. Type: string (or Expression with resultType string). - */ - dataLakeStoreUri: any; - /** - * The ID of the application used to authenticate against the Azure Data Lake Store account. - * Type: string (or Expression with resultType string). - */ - servicePrincipalId?: any; - /** - * The Key of the application used to authenticate against the Azure Data Lake Store account. + * The non-access credential portion of the connection string as well as an optional encrypted + * credential. Type: string, SecureString or AzureKeyVaultSecretReference. */ - servicePrincipalKey?: SecretBaseUnion; + connectionString: any; /** - * The name or ID of the tenant to which the service principal belongs. Type: string (or - * Expression with resultType string). + * Type of authentication used to connect to the Microsoft Access as ODBC data store. Possible + * values are: Anonymous and Basic. Type: string (or Expression with resultType string). */ - tenant?: any; + authenticationType?: any; /** - * Data Lake Store account name. Type: string (or Expression with resultType string). + * The access credential portion of the connection string specified in driver-specific + * property-value format. */ - accountName?: any; + credential?: SecretBaseUnion; /** - * Data Lake Store account subscription ID (if different from Data Factory account). Type: string - * (or Expression with resultType string). + * User name for Basic authentication. Type: string (or Expression with resultType string). */ - subscriptionId?: any; + userName?: any; /** - * Data Lake Store account resource group name (if different from Data Factory account). Type: - * string (or Expression with resultType string). + * Password for Basic authentication. */ - resourceGroupName?: any; + password?: SecretBaseUnion; /** * The encrypted credential used for authentication. Credentials are encrypted using the * integration runtime credential manager. Type: string (or Expression with resultType string). @@ -6110,13 +6163,13 @@ export interface AzureDataLakeStoreLinkedService { } /** - * Linked service for CosmosDB (MongoDB API) data source. + * Informix linked service. */ -export interface CosmosDbMongoDbApiLinkedService { +export interface InformixLinkedService { /** * Polymorphic Discriminator */ - type: "CosmosDbMongoDbApi"; + type: "Informix"; /** * The integration runtime reference. */ @@ -6134,61 +6187,43 @@ export interface CosmosDbMongoDbApiLinkedService { */ annotations?: any[]; /** - * The CosmosDB (MongoDB API) connection string. Type: string, SecureString or - * AzureKeyVaultSecretReference. Type: string, SecureString or AzureKeyVaultSecretReference. + * The non-access credential portion of the connection string as well as an optional encrypted + * credential. Type: string, SecureString or AzureKeyVaultSecretReference. */ connectionString: any; /** - * The name of the CosmosDB (MongoDB API) database that you want to access. Type: string (or - * Expression with resultType string). - */ - database: any; -} - -/** - * Linked service for MongoDB data source. - */ -export interface MongoDbV2LinkedService { - /** - * Polymorphic Discriminator - */ - type: "MongoDbV2"; - /** - * The integration runtime reference. - */ - connectVia?: IntegrationRuntimeReference; - /** - * Linked service description. + * Type of authentication used to connect to the Informix as ODBC data store. Possible values + * are: Anonymous and Basic. Type: string (or Expression with resultType string). */ - description?: string; + authenticationType?: any; /** - * Parameters for linked service. + * The access credential portion of the connection string specified in driver-specific + * property-value format. */ - parameters?: { [propertyName: string]: ParameterSpecification }; + credential?: SecretBaseUnion; /** - * List of tags that can be used for describing the linked service. + * User name for Basic authentication. Type: string (or Expression with resultType string). */ - annotations?: any[]; + userName?: any; /** - * The MongoDB connection string. Type: string, SecureString or AzureKeyVaultSecretReference. - * Type: string, SecureString or AzureKeyVaultSecretReference. + * Password for Basic authentication. */ - connectionString: any; + password?: SecretBaseUnion; /** - * The name of the MongoDB database that you want to access. Type: string (or Expression with - * resultType string). + * The encrypted credential used for authentication. Credentials are encrypted using the + * integration runtime credential manager. Type: string (or Expression with resultType string). */ - database: any; + encryptedCredential?: any; } /** - * Linked service for MongoDb data source. + * Open Database Connectivity (ODBC) linked service. */ -export interface MongoDbLinkedService { +export interface OdbcLinkedService { /** * Polymorphic Discriminator */ - type: "MongoDb"; + type: "Odbc"; /** * The integration runtime reference. */ @@ -6206,48 +6241,28 @@ export interface MongoDbLinkedService { */ annotations?: any[]; /** - * The IP address or server name of the MongoDB server. Type: string (or Expression with - * resultType string). + * The non-access credential portion of the connection string as well as an optional encrypted + * credential. Type: string, SecureString or AzureKeyVaultSecretReference. */ - server: any; + connectionString: any; /** - * The authentication type to be used to connect to the MongoDB database. Possible values - * include: 'Basic', 'Anonymous' + * Type of authentication used to connect to the ODBC data store. Possible values are: Anonymous + * and Basic. Type: string (or Expression with resultType string). */ - authenticationType?: MongoDbAuthenticationType; + authenticationType?: any; /** - * The name of the MongoDB database that you want to access. Type: string (or Expression with - * resultType string). + * The access credential portion of the connection string specified in driver-specific + * property-value format. */ - databaseName: any; + credential?: SecretBaseUnion; /** - * Username for authentication. Type: string (or Expression with resultType string). + * User name for Basic authentication. Type: string (or Expression with resultType string). */ - username?: any; + userName?: any; /** - * Password for authentication. + * Password for Basic authentication. */ password?: SecretBaseUnion; - /** - * Database to verify the username and password. Type: string (or Expression with resultType - * string). - */ - authSource?: any; - /** - * The TCP port number that the MongoDB server uses to listen for client connections. The default - * value is 27017. Type: integer (or Expression with resultType integer), minimum: 0. - */ - port?: any; - /** - * Specifies whether the connections to the server are encrypted using SSL. The default value is - * false. Type: boolean (or Expression with resultType boolean). - */ - enableSsl?: any; - /** - * Specifies whether to allow self-signed certificates from the server. The default value is - * false. Type: boolean (or Expression with resultType boolean). - */ - allowSelfSignedServerCert?: any; /** * The encrypted credential used for authentication. Credentials are encrypted using the * integration runtime credential manager. Type: string (or Expression with resultType string). @@ -6256,13 +6271,13 @@ export interface MongoDbLinkedService { } /** - * Linked service for Cassandra data source. + * Azure ML Service linked service. */ -export interface CassandraLinkedService { +export interface AzureMLServiceLinkedService { /** * Polymorphic Discriminator */ - type: "Cassandra"; + type: "AzureMLService"; /** * The integration runtime reference. */ @@ -6280,26 +6295,34 @@ export interface CassandraLinkedService { */ annotations?: any[]; /** - * Host name for connection. Type: string (or Expression with resultType string). + * Azure ML Service workspace subscription ID. Type: string (or Expression with resultType + * string). */ - host: any; + subscriptionId: any; /** - * AuthenticationType to be used for connection. Type: string (or Expression with resultType + * Azure ML Service workspace resource group name. Type: string (or Expression with resultType * string). */ - authenticationType?: any; + resourceGroupName: any; /** - * The port for the connection. Type: integer (or Expression with resultType integer). + * Azure ML Service workspace name. Type: string (or Expression with resultType string). */ - port?: any; + mlWorkspaceName: any; /** - * Username for authentication. Type: string (or Expression with resultType string). + * The ID of the service principal used to authenticate against the endpoint of a published Azure + * ML Service pipeline. Type: string (or Expression with resultType string). */ - username?: any; + servicePrincipalId?: any; /** - * Password for authentication. + * The key of the service principal used to authenticate against the endpoint of a published + * Azure ML Service pipeline. */ - password?: SecretBaseUnion; + servicePrincipalKey?: SecretBaseUnion; + /** + * The name or ID of the tenant to which the service principal belongs. Type: string (or + * Expression with resultType string). + */ + tenant?: any; /** * The encrypted credential used for authentication. Credentials are encrypted using the * integration runtime credential manager. Type: string (or Expression with resultType string). @@ -6308,97 +6331,74 @@ export interface CassandraLinkedService { } /** - * Contains the possible cases for WebLinkedServiceTypeProperties. - */ -export type WebLinkedServiceTypePropertiesUnion = WebLinkedServiceTypeProperties | WebClientCertificateAuthentication | WebBasicAuthentication | WebAnonymousAuthentication; - -/** - * Base definition of WebLinkedServiceTypeProperties, this typeProperties is polymorphic based on - * authenticationType, so not flattened in SDK models. + * Azure ML Studio Web Service linked service. */ -export interface WebLinkedServiceTypeProperties { +export interface AzureMLLinkedService { /** * Polymorphic Discriminator */ - authenticationType: "WebLinkedServiceTypeProperties"; + type: "AzureML"; /** - * The URL of the web service endpoint, e.g. http://www.microsoft.com . Type: string (or - * Expression with resultType string). + * The integration runtime reference. */ - url: any; -} - -/** - * A WebLinkedService that uses client certificate based authentication to communicate with an HTTP - * endpoint. This scheme follows mutual authentication; the server must also provide valid - * credentials to the client. - */ -export interface WebClientCertificateAuthentication { + connectVia?: IntegrationRuntimeReference; /** - * Polymorphic Discriminator + * Linked service description. */ - authenticationType: "ClientCertificate"; + description?: string; /** - * The URL of the web service endpoint, e.g. http://www.microsoft.com . Type: string (or - * Expression with resultType string). + * Parameters for linked service. */ - url: any; + parameters?: { [propertyName: string]: ParameterSpecification }; /** - * Base64-encoded contents of a PFX file. + * List of tags that can be used for describing the linked service. */ - pfx: SecretBaseUnion; + annotations?: any[]; /** - * Password for the PFX file. + * The Batch Execution REST URL for an Azure ML Studio Web Service endpoint. Type: string (or + * Expression with resultType string). */ - password: SecretBaseUnion; -} - -/** - * A WebLinkedService that uses basic authentication to communicate with an HTTP endpoint. - */ -export interface WebBasicAuthentication { + mlEndpoint: any; /** - * Polymorphic Discriminator + * The API key for accessing the Azure ML model endpoint. */ - authenticationType: "Basic"; + apiKey: SecretBaseUnion; /** - * The URL of the web service endpoint, e.g. http://www.microsoft.com . Type: string (or + * The Update Resource REST URL for an Azure ML Studio Web Service endpoint. Type: string (or * Expression with resultType string). */ - url: any; - /** - * User name for Basic authentication. Type: string (or Expression with resultType string). - */ - username: any; + updateResourceEndpoint?: any; /** - * The password for Basic authentication. + * The ID of the service principal used to authenticate against the ARM-based + * updateResourceEndpoint of an Azure ML Studio web service. Type: string (or Expression with + * resultType string). */ - password: SecretBaseUnion; -} - -/** - * A WebLinkedService that uses anonymous authentication to communicate with an HTTP endpoint. - */ -export interface WebAnonymousAuthentication { + servicePrincipalId?: any; /** - * Polymorphic Discriminator + * The key of the service principal used to authenticate against the ARM-based + * updateResourceEndpoint of an Azure ML Studio web service. */ - authenticationType: "Anonymous"; + servicePrincipalKey?: SecretBaseUnion; /** - * The URL of the web service endpoint, e.g. http://www.microsoft.com . Type: string (or + * The name or ID of the tenant to which the service principal belongs. Type: string (or * Expression with resultType string). */ - url: any; + tenant?: any; + /** + * The encrypted credential used for authentication. Credentials are encrypted using the + * integration runtime credential manager. Type: string (or Expression with resultType string). + */ + encryptedCredential?: any; } /** - * Web linked service. + * Linked service for Teradata data source. */ -export interface WebLinkedService { +export interface TeradataLinkedService { /** * Polymorphic Discriminator */ - type: "Web"; + type: "Teradata"; /** * The integration runtime reference. */ @@ -6416,19 +6416,40 @@ export interface WebLinkedService { */ annotations?: any[]; /** - * Web linked service properties. + * Teradata ODBC connection string. Type: string, SecureString or AzureKeyVaultSecretReference. */ - typeProperties: WebLinkedServiceTypePropertiesUnion; + connectionString?: any; + /** + * Server name for connection. Type: string (or Expression with resultType string). + */ + server?: any; + /** + * AuthenticationType to be used for connection. Possible values include: 'Basic', 'Windows' + */ + authenticationType?: TeradataAuthenticationType; + /** + * Username for authentication. Type: string (or Expression with resultType string). + */ + username?: any; + /** + * Password for authentication. + */ + password?: SecretBaseUnion; + /** + * The encrypted credential used for authentication. Credentials are encrypted using the + * integration runtime credential manager. Type: string (or Expression with resultType string). + */ + encryptedCredential?: any; } /** - * Open Data Protocol (OData) linked service. + * Linked service for DB2 data source. */ -export interface ODataLinkedService { +export interface Db2LinkedService { /** * Polymorphic Discriminator */ - type: "OData"; + type: "Db2"; /** * The integration runtime reference. */ @@ -6446,72 +6467,61 @@ export interface ODataLinkedService { */ annotations?: any[]; /** - * The URL of the OData service endpoint. Type: string (or Expression with resultType string). - */ - url: any; - /** - * Type of authentication used to connect to the OData service. Possible values include: 'Basic', - * 'Anonymous', 'Windows', 'AadServicePrincipal', 'ManagedServiceIdentity' - */ - authenticationType?: ODataAuthenticationType; - /** - * User name of the OData service. Type: string (or Expression with resultType string). - */ - userName?: any; - /** - * Password of the OData service. + * The connection string. It is mutually exclusive with server, database, authenticationType, + * userName, packageCollection and certificateCommonName property. Type: string, SecureString or + * AzureKeyVaultSecretReference. */ - password?: SecretBaseUnion; + connectionString?: any; /** - * Specify the tenant information (domain name or tenant ID) under which your application - * resides. Type: string (or Expression with resultType string). + * Server name for connection. It is mutually exclusive with connectionString property. Type: + * string (or Expression with resultType string). */ - tenant?: any; + server?: any; /** - * Specify the application id of your application registered in Azure Active Directory. Type: + * Database name for connection. It is mutually exclusive with connectionString property. Type: * string (or Expression with resultType string). */ - servicePrincipalId?: any; + database?: any; /** - * Specify the resource you are requesting authorization to use Directory. Type: string (or - * Expression with resultType string). + * AuthenticationType to be used for connection. It is mutually exclusive with connectionString + * property. Possible values include: 'Basic' */ - aadResourceId?: any; + authenticationType?: Db2AuthenticationType; /** - * Specify the credential type (key or cert) is used for service principal. Possible values - * include: 'ServicePrincipalKey', 'ServicePrincipalCert' + * Username for authentication. It is mutually exclusive with connectionString property. Type: + * string (or Expression with resultType string). */ - aadServicePrincipalCredentialType?: ODataAadServicePrincipalCredentialType; + username?: any; /** - * Specify the secret of your application registered in Azure Active Directory. Type: string (or - * Expression with resultType string). + * Password for authentication. */ - servicePrincipalKey?: SecretBaseUnion; + password?: SecretBaseUnion; /** - * Specify the base64 encoded certificate of your application registered in Azure Active - * Directory. Type: string (or Expression with resultType string). + * Under where packages are created when querying database. It is mutually exclusive with + * connectionString property. Type: string (or Expression with resultType string). */ - servicePrincipalEmbeddedCert?: SecretBaseUnion; + packageCollection?: any; /** - * Specify the password of your certificate if your certificate has a password and you are using - * AadServicePrincipal authentication. Type: string (or Expression with resultType string). + * Certificate Common Name when TLS is enabled. It is mutually exclusive with connectionString + * property. Type: string (or Expression with resultType string). */ - servicePrincipalEmbeddedCertPassword?: SecretBaseUnion; + certificateCommonName?: any; /** * The encrypted credential used for authentication. Credentials are encrypted using the - * integration runtime credential manager. Type: string (or Expression with resultType string). + * integration runtime credential manager. It is mutually exclusive with connectionString + * property. Type: string (or Expression with resultType string). */ encryptedCredential?: any; } /** - * Hadoop Distributed File System (HDFS) linked service. + * Linked service for Sybase data source. */ -export interface HdfsLinkedService { +export interface SybaseLinkedService { /** * Polymorphic Discriminator */ - type: "Hdfs"; + type: "Sybase"; /** * The integration runtime reference. */ @@ -6529,38 +6539,44 @@ export interface HdfsLinkedService { */ annotations?: any[]; /** - * The URL of the HDFS service endpoint, e.g. http://myhostname:50070/webhdfs/v1 . Type: string - * (or Expression with resultType string). + * Server name for connection. Type: string (or Expression with resultType string). */ - url: any; + server: any; /** - * Type of authentication used to connect to the HDFS. Possible values are: Anonymous and - * Windows. Type: string (or Expression with resultType string). + * Database name for connection. Type: string (or Expression with resultType string). */ - authenticationType?: any; + database: any; /** - * The encrypted credential used for authentication. Credentials are encrypted using the - * integration runtime credential manager. Type: string (or Expression with resultType string). + * Schema name for connection. Type: string (or Expression with resultType string). */ - encryptedCredential?: any; + schema?: any; /** - * User name for Windows authentication. Type: string (or Expression with resultType string). + * AuthenticationType to be used for connection. Possible values include: 'Basic', 'Windows' */ - userName?: any; + authenticationType?: SybaseAuthenticationType; /** - * Password for Windows authentication. + * Username for authentication. Type: string (or Expression with resultType string). + */ + username?: any; + /** + * Password for authentication. */ password?: SecretBaseUnion; + /** + * The encrypted credential used for authentication. Credentials are encrypted using the + * integration runtime credential manager. Type: string (or Expression with resultType string). + */ + encryptedCredential?: any; } /** - * Microsoft Access linked service. + * Linked service for PostgreSQL data source. */ -export interface MicrosoftAccessLinkedService { +export interface PostgreSqlLinkedService { /** * Polymorphic Discriminator */ - type: "MicrosoftAccess"; + type: "PostgreSql"; /** * The integration runtime reference. */ @@ -6578,28 +6594,13 @@ export interface MicrosoftAccessLinkedService { */ annotations?: any[]; /** - * The non-access credential portion of the connection string as well as an optional encrypted - * credential. Type: string, SecureString or AzureKeyVaultSecretReference. + * The connection string. */ connectionString: any; /** - * Type of authentication used to connect to the Microsoft Access as ODBC data store. Possible - * values are: Anonymous and Basic. Type: string (or Expression with resultType string). - */ - authenticationType?: any; - /** - * The access credential portion of the connection string specified in driver-specific - * property-value format. - */ - credential?: SecretBaseUnion; - /** - * User name for Basic authentication. Type: string (or Expression with resultType string). - */ - userName?: any; - /** - * Password for Basic authentication. + * The Azure key vault secret reference of password in connection string. */ - password?: SecretBaseUnion; + password?: AzureKeyVaultSecretReference; /** * The encrypted credential used for authentication. Credentials are encrypted using the * integration runtime credential manager. Type: string (or Expression with resultType string). @@ -6608,13 +6609,13 @@ export interface MicrosoftAccessLinkedService { } /** - * Informix linked service. + * Linked service for MySQL data source. */ -export interface InformixLinkedService { +export interface MySqlLinkedService { /** * Polymorphic Discriminator */ - type: "Informix"; + type: "MySql"; /** * The integration runtime reference. */ @@ -6632,28 +6633,13 @@ export interface InformixLinkedService { */ annotations?: any[]; /** - * The non-access credential portion of the connection string as well as an optional encrypted - * credential. Type: string, SecureString or AzureKeyVaultSecretReference. + * The connection string. */ connectionString: any; /** - * Type of authentication used to connect to the Informix as ODBC data store. Possible values - * are: Anonymous and Basic. Type: string (or Expression with resultType string). - */ - authenticationType?: any; - /** - * The access credential portion of the connection string specified in driver-specific - * property-value format. - */ - credential?: SecretBaseUnion; - /** - * User name for Basic authentication. Type: string (or Expression with resultType string). - */ - userName?: any; - /** - * Password for Basic authentication. + * The Azure key vault secret reference of password in connection string. */ - password?: SecretBaseUnion; + password?: AzureKeyVaultSecretReference; /** * The encrypted credential used for authentication. Credentials are encrypted using the * integration runtime credential manager. Type: string (or Expression with resultType string). @@ -6662,13 +6648,13 @@ export interface InformixLinkedService { } /** - * Open Database Connectivity (ODBC) linked service. + * Azure MySQL database linked service. */ -export interface OdbcLinkedService { +export interface AzureMySqlLinkedService { /** * Polymorphic Discriminator */ - type: "Odbc"; + type: "AzureMySql"; /** * The integration runtime reference. */ @@ -6686,28 +6672,13 @@ export interface OdbcLinkedService { */ annotations?: any[]; /** - * The non-access credential portion of the connection string as well as an optional encrypted - * credential. Type: string, SecureString or AzureKeyVaultSecretReference. + * The connection string. Type: string, SecureString or AzureKeyVaultSecretReference. */ connectionString: any; /** - * Type of authentication used to connect to the ODBC data store. Possible values are: Anonymous - * and Basic. Type: string (or Expression with resultType string). - */ - authenticationType?: any; - /** - * The access credential portion of the connection string specified in driver-specific - * property-value format. - */ - credential?: SecretBaseUnion; - /** - * User name for Basic authentication. Type: string (or Expression with resultType string). - */ - userName?: any; - /** - * Password for Basic authentication. + * The Azure key vault secret reference of password in connection string. */ - password?: SecretBaseUnion; + password?: AzureKeyVaultSecretReference; /** * The encrypted credential used for authentication. Credentials are encrypted using the * integration runtime credential manager. Type: string (or Expression with resultType string). @@ -6716,13 +6687,13 @@ export interface OdbcLinkedService { } /** - * Azure ML Service linked service. + * Oracle database. */ -export interface AzureMLServiceLinkedService { +export interface OracleLinkedService { /** * Polymorphic Discriminator */ - type: "AzureMLService"; + type: "Oracle"; /** * The integration runtime reference. */ @@ -6740,34 +6711,59 @@ export interface AzureMLServiceLinkedService { */ annotations?: any[]; /** - * Azure ML Service workspace subscription ID. Type: string (or Expression with resultType - * string). + * The connection string. Type: string, SecureString or AzureKeyVaultSecretReference. */ - subscriptionId: any; + connectionString: any; /** - * Azure ML Service workspace resource group name. Type: string (or Expression with resultType - * string). + * The Azure key vault secret reference of password in connection string. */ - resourceGroupName: any; + password?: AzureKeyVaultSecretReference; /** - * Azure ML Service workspace name. Type: string (or Expression with resultType string). + * The encrypted credential used for authentication. Credentials are encrypted using the + * integration runtime credential manager. Type: string (or Expression with resultType string). + */ + encryptedCredential?: any; +} + +/** + * Linked service for Google Cloud Storage. + */ +export interface GoogleCloudStorageLinkedService { + /** + * Polymorphic Discriminator + */ + type: "GoogleCloudStorage"; + /** + * The integration runtime reference. + */ + connectVia?: IntegrationRuntimeReference; + /** + * Linked service description. + */ + description?: string; + /** + * Parameters for linked service. + */ + parameters?: { [propertyName: string]: ParameterSpecification }; + /** + * List of tags that can be used for describing the linked service. */ - mlWorkspaceName: any; + annotations?: any[]; /** - * The ID of the service principal used to authenticate against the endpoint of a published Azure - * ML Service pipeline. Type: string (or Expression with resultType string). + * The access key identifier of the Google Cloud Storage Identity and Access Management (IAM) + * user. Type: string (or Expression with resultType string). */ - servicePrincipalId?: any; + accessKeyId?: any; /** - * The key of the service principal used to authenticate against the endpoint of a published - * Azure ML Service pipeline. + * The secret access key of the Google Cloud Storage Identity and Access Management (IAM) user. */ - servicePrincipalKey?: SecretBaseUnion; + secretAccessKey?: SecretBaseUnion; /** - * The name or ID of the tenant to which the service principal belongs. Type: string (or - * Expression with resultType string). + * This value specifies the endpoint to access with the Google Cloud Storage Connector. This is + * an optional property; change it only if you want to try a different service endpoint or want + * to switch between https and http. Type: string (or Expression with resultType string). */ - tenant?: any; + serviceUrl?: any; /** * The encrypted credential used for authentication. Credentials are encrypted using the * integration runtime credential manager. Type: string (or Expression with resultType string). @@ -6776,13 +6772,13 @@ export interface AzureMLServiceLinkedService { } /** - * Azure ML Studio Web Service linked service. + * Azure File Storage linked service. */ -export interface AzureMLLinkedService { +export interface AzureFileStorageLinkedService { /** * Polymorphic Discriminator */ - type: "AzureML"; + type: "AzureFileStorage"; /** * The integration runtime reference. */ @@ -6800,35 +6796,17 @@ export interface AzureMLLinkedService { */ annotations?: any[]; /** - * The Batch Execution REST URL for an Azure ML Studio Web Service endpoint. Type: string (or - * Expression with resultType string). - */ - mlEndpoint: any; - /** - * The API key for accessing the Azure ML model endpoint. - */ - apiKey: SecretBaseUnion; - /** - * The Update Resource REST URL for an Azure ML Studio Web Service endpoint. Type: string (or - * Expression with resultType string). - */ - updateResourceEndpoint?: any; - /** - * The ID of the service principal used to authenticate against the ARM-based - * updateResourceEndpoint of an Azure ML Studio web service. Type: string (or Expression with - * resultType string). + * Host name of the server. Type: string (or Expression with resultType string). */ - servicePrincipalId?: any; + host: any; /** - * The key of the service principal used to authenticate against the ARM-based - * updateResourceEndpoint of an Azure ML Studio web service. + * User ID to logon the server. Type: string (or Expression with resultType string). */ - servicePrincipalKey?: SecretBaseUnion; + userId?: any; /** - * The name or ID of the tenant to which the service principal belongs. Type: string (or - * Expression with resultType string). + * Password to logon the server. */ - tenant?: any; + password?: SecretBaseUnion; /** * The encrypted credential used for authentication. Credentials are encrypted using the * integration runtime credential manager. Type: string (or Expression with resultType string). @@ -6837,13 +6815,13 @@ export interface AzureMLLinkedService { } /** - * Linked service for Teradata data source. + * File system linked service. */ -export interface TeradataLinkedService { +export interface FileServerLinkedService { /** * Polymorphic Discriminator */ - type: "Teradata"; + type: "FileServer"; /** * The integration runtime reference. */ @@ -6861,23 +6839,15 @@ export interface TeradataLinkedService { */ annotations?: any[]; /** - * Teradata ODBC connection string. Type: string, SecureString or AzureKeyVaultSecretReference. - */ - connectionString?: any; - /** - * Server name for connection. Type: string (or Expression with resultType string). - */ - server?: any; - /** - * AuthenticationType to be used for connection. Possible values include: 'Basic', 'Windows' + * Host name of the server. Type: string (or Expression with resultType string). */ - authenticationType?: TeradataAuthenticationType; + host: any; /** - * Username for authentication. Type: string (or Expression with resultType string). + * User ID to logon the server. Type: string (or Expression with resultType string). */ - username?: any; + userId?: any; /** - * Password for authentication. + * Password to logon the server. */ password?: SecretBaseUnion; /** @@ -6888,13 +6858,13 @@ export interface TeradataLinkedService { } /** - * Linked service for DB2 data source. + * HDInsight linked service. */ -export interface Db2LinkedService { +export interface HDInsightLinkedService { /** * Polymorphic Discriminator */ - type: "Db2"; + type: "HDInsight"; /** * The integration runtime reference. */ @@ -6912,50 +6882,49 @@ export interface Db2LinkedService { */ annotations?: any[]; /** - * Server name for connection. Type: string (or Expression with resultType string). - */ - server: any; - /** - * Database name for connection. Type: string (or Expression with resultType string). - */ - database: any; - /** - * AuthenticationType to be used for connection. Possible values include: 'Basic' + * HDInsight cluster URI. Type: string (or Expression with resultType string). */ - authenticationType?: Db2AuthenticationType; + clusterUri: any; /** - * Username for authentication. Type: string (or Expression with resultType string). + * HDInsight cluster user name. Type: string (or Expression with resultType string). */ - username?: any; + userName?: any; /** - * Password for authentication. + * HDInsight cluster password. */ password?: SecretBaseUnion; /** - * Under where packages are created when querying database. Type: string (or Expression with - * resultType string). + * The Azure Storage linked service reference. */ - packageCollection?: any; + linkedServiceName?: LinkedServiceReference; /** - * Certificate Common Name when TLS is enabled. Type: string (or Expression with resultType - * string). + * A reference to the Azure SQL linked service that points to the HCatalog database. */ - certificateCommonName?: any; + hcatalogLinkedServiceName?: LinkedServiceReference; /** * The encrypted credential used for authentication. Credentials are encrypted using the * integration runtime credential manager. Type: string (or Expression with resultType string). */ encryptedCredential?: any; + /** + * Specify if the HDInsight is created with ESP (Enterprise Security Package). Type: Boolean. + */ + isEspEnabled?: any; + /** + * Specify the FileSystem if the main storage for the HDInsight is ADLS Gen2. Type: string (or + * Expression with resultType string). + */ + fileSystem?: any; } /** - * Linked service for Sybase data source. + * Common Data Service for Apps linked service. */ -export interface SybaseLinkedService { +export interface CommonDataServiceForAppsLinkedService { /** * Polymorphic Discriminator */ - type: "Sybase"; + type: "CommonDataServiceForApps"; /** * The integration runtime reference. */ @@ -6973,29 +6942,68 @@ export interface SybaseLinkedService { */ annotations?: any[]; /** - * Server name for connection. Type: string (or Expression with resultType string). + * The deployment type of the Common Data Service for Apps instance. 'Online' for Common Data + * Service for Apps Online and 'OnPremisesWithIfd' for Common Data Service for Apps on-premises + * with Ifd. Type: string (or Expression with resultType string). Possible values include: + * 'Online', 'OnPremisesWithIfd' */ - server: any; + deploymentType: DynamicsDeploymentType; /** - * Database name for connection. Type: string (or Expression with resultType string). + * The host name of the on-premises Common Data Service for Apps server. The property is required + * for on-prem and not allowed for online. Type: string (or Expression with resultType string). */ - database: any; + hostName?: any; /** - * Schema name for connection. Type: string (or Expression with resultType string). + * The port of on-premises Common Data Service for Apps server. The property is required for + * on-prem and not allowed for online. Default is 443. Type: integer (or Expression with + * resultType integer), minimum: 0. */ - schema?: any; + port?: any; /** - * AuthenticationType to be used for connection. Possible values include: 'Basic', 'Windows' + * The URL to the Microsoft Common Data Service for Apps server. The property is required for + * on-line and not allowed for on-prem. Type: string (or Expression with resultType string). */ - authenticationType?: SybaseAuthenticationType; + serviceUri?: any; /** - * Username for authentication. Type: string (or Expression with resultType string). + * The organization name of the Common Data Service for Apps instance. The property is required + * for on-prem and required for online when there are more than one Common Data Service for Apps + * instances associated with the user. Type: string (or Expression with resultType string). + */ + organizationName?: any; + /** + * The authentication type to connect to Common Data Service for Apps server. 'Office365' for + * online scenario, 'Ifd' for on-premises with Ifd scenario. 'AADServicePrincipal' for + * Server-To-Server authentication in online scenario. Type: string (or Expression with + * resultType string). Possible values include: 'Office365', 'Ifd', 'AADServicePrincipal' + */ + authenticationType: DynamicsAuthenticationType; + /** + * User name to access the Common Data Service for Apps instance. Type: string (or Expression + * with resultType string). */ username?: any; /** - * Password for authentication. + * Password to access the Common Data Service for Apps instance. */ password?: SecretBaseUnion; + /** + * The client ID of the application in Azure Active Directory used for Server-To-Server + * authentication. Type: string (or Expression with resultType string). + */ + servicePrincipalId?: any; + /** + * The service principal credential type to use in Server-To-Server authentication. + * 'ServicePrincipalKey' for key/secret, 'ServicePrincipalCert' for certificate. Type: string (or + * Expression with resultType string). + */ + servicePrincipalCredentialType?: any; + /** + * The credential of the service principal object in Azure Active Directory. If + * servicePrincipalCredentialType is 'ServicePrincipalKey', servicePrincipalCredential can be + * SecureString or AzureKeyVaultSecretReference. If servicePrincipalCredentialType is + * 'ServicePrincipalCert', servicePrincipalCredential can only be AzureKeyVaultSecretReference. + */ + servicePrincipalCredential?: SecretBaseUnion; /** * The encrypted credential used for authentication. Credentials are encrypted using the * integration runtime credential manager. Type: string (or Expression with resultType string). @@ -7004,13 +7012,13 @@ export interface SybaseLinkedService { } /** - * Linked service for PostgreSQL data source. + * Dynamics CRM linked service. */ -export interface PostgreSqlLinkedService { +export interface DynamicsCrmLinkedService { /** * Polymorphic Discriminator */ - type: "PostgreSql"; + type: "DynamicsCrm"; /** * The integration runtime reference. */ @@ -7028,52 +7036,67 @@ export interface PostgreSqlLinkedService { */ annotations?: any[]; /** - * The connection string. + * The deployment type of the Dynamics CRM instance. 'Online' for Dynamics CRM Online and + * 'OnPremisesWithIfd' for Dynamics CRM on-premises with Ifd. Type: string (or Expression with + * resultType string). Possible values include: 'Online', 'OnPremisesWithIfd' */ - connectionString: any; + deploymentType: DynamicsDeploymentType; /** - * The Azure key vault secret reference of password in connection string. + * The host name of the on-premises Dynamics CRM server. The property is required for on-prem and + * not allowed for online. Type: string (or Expression with resultType string). */ - password?: AzureKeyVaultSecretReference; + hostName?: any; /** - * The encrypted credential used for authentication. Credentials are encrypted using the - * integration runtime credential manager. Type: string (or Expression with resultType string). + * The port of on-premises Dynamics CRM server. The property is required for on-prem and not + * allowed for online. Default is 443. Type: integer (or Expression with resultType integer), + * minimum: 0. */ - encryptedCredential?: any; -} - -/** - * Linked service for MySQL data source. - */ -export interface MySqlLinkedService { + port?: any; /** - * Polymorphic Discriminator + * The URL to the Microsoft Dynamics CRM server. The property is required for on-line and not + * allowed for on-prem. Type: string (or Expression with resultType string). */ - type: "MySql"; + serviceUri?: any; /** - * The integration runtime reference. + * The organization name of the Dynamics CRM instance. The property is required for on-prem and + * required for online when there are more than one Dynamics CRM instances associated with the + * user. Type: string (or Expression with resultType string). */ - connectVia?: IntegrationRuntimeReference; + organizationName?: any; /** - * Linked service description. + * The authentication type to connect to Dynamics CRM server. 'Office365' for online scenario, + * 'Ifd' for on-premises with Ifd scenario, 'AADServicePrincipal' for Server-To-Server + * authentication in online scenario. Type: string (or Expression with resultType string). + * Possible values include: 'Office365', 'Ifd', 'AADServicePrincipal' */ - description?: string; + authenticationType: DynamicsAuthenticationType; /** - * Parameters for linked service. + * User name to access the Dynamics CRM instance. Type: string (or Expression with resultType + * string). */ - parameters?: { [propertyName: string]: ParameterSpecification }; + username?: any; /** - * List of tags that can be used for describing the linked service. + * Password to access the Dynamics CRM instance. */ - annotations?: any[]; + password?: SecretBaseUnion; /** - * The connection string. + * The client ID of the application in Azure Active Directory used for Server-To-Server + * authentication. Type: string (or Expression with resultType string). */ - connectionString: any; + servicePrincipalId?: any; /** - * The Azure key vault secret reference of password in connection string. + * The service principal credential type to use in Server-To-Server authentication. + * 'ServicePrincipalKey' for key/secret, 'ServicePrincipalCert' for certificate. Type: string (or + * Expression with resultType string). + */ + servicePrincipalCredentialType?: any; + /** + * The credential of the service principal object in Azure Active Directory. If + * servicePrincipalCredentialType is 'ServicePrincipalKey', servicePrincipalCredential can be + * SecureString or AzureKeyVaultSecretReference. If servicePrincipalCredentialType is + * 'ServicePrincipalCert', servicePrincipalCredential can only be AzureKeyVaultSecretReference. */ - password?: AzureKeyVaultSecretReference; + servicePrincipalCredential?: SecretBaseUnion; /** * The encrypted credential used for authentication. Credentials are encrypted using the * integration runtime credential manager. Type: string (or Expression with resultType string). @@ -7082,13 +7105,13 @@ export interface MySqlLinkedService { } /** - * Azure MySQL database linked service. + * Dynamics linked service. */ -export interface AzureMySqlLinkedService { +export interface DynamicsLinkedService { /** * Polymorphic Discriminator */ - type: "AzureMySql"; + type: "Dynamics"; /** * The integration runtime reference. */ @@ -7106,52 +7129,65 @@ export interface AzureMySqlLinkedService { */ annotations?: any[]; /** - * The connection string. Type: string, SecureString or AzureKeyVaultSecretReference. + * The deployment type of the Dynamics instance. 'Online' for Dynamics Online and + * 'OnPremisesWithIfd' for Dynamics on-premises with Ifd. Type: string (or Expression with + * resultType string). */ - connectionString: any; + deploymentType: any; /** - * The Azure key vault secret reference of password in connection string. + * The host name of the on-premises Dynamics server. The property is required for on-prem and not + * allowed for online. Type: string (or Expression with resultType string). */ - password?: AzureKeyVaultSecretReference; + hostName?: any; /** - * The encrypted credential used for authentication. Credentials are encrypted using the - * integration runtime credential manager. Type: string (or Expression with resultType string). + * The port of on-premises Dynamics server. The property is required for on-prem and not allowed + * for online. Default is 443. Type: integer (or Expression with resultType integer), minimum: 0. */ - encryptedCredential?: any; -} - -/** - * Oracle database. - */ -export interface OracleLinkedService { + port?: any; /** - * Polymorphic Discriminator + * The URL to the Microsoft Dynamics server. The property is required for on-line and not allowed + * for on-prem. Type: string (or Expression with resultType string). */ - type: "Oracle"; + serviceUri?: any; /** - * The integration runtime reference. + * The organization name of the Dynamics instance. The property is required for on-prem and + * required for online when there are more than one Dynamics instances associated with the user. + * Type: string (or Expression with resultType string). */ - connectVia?: IntegrationRuntimeReference; + organizationName?: any; /** - * Linked service description. + * The authentication type to connect to Dynamics server. 'Office365' for online scenario, 'Ifd' + * for on-premises with Ifd scenario, 'AADServicePrincipal' for Server-To-Server authentication + * in online scenario. Type: string (or Expression with resultType string). */ - description?: string; + authenticationType: any; /** - * Parameters for linked service. + * User name to access the Dynamics instance. Type: string (or Expression with resultType + * string). */ - parameters?: { [propertyName: string]: ParameterSpecification }; + username?: any; /** - * List of tags that can be used for describing the linked service. + * Password to access the Dynamics instance. */ - annotations?: any[]; + password?: SecretBaseUnion; /** - * The connection string. Type: string, SecureString or AzureKeyVaultSecretReference. + * The client ID of the application in Azure Active Directory used for Server-To-Server + * authentication. Type: string (or Expression with resultType string). */ - connectionString: any; + servicePrincipalId?: any; /** - * The Azure key vault secret reference of password in connection string. + * The service principal credential type to use in Server-To-Server authentication. + * 'ServicePrincipalKey' for key/secret, 'ServicePrincipalCert' for certificate. Type: string (or + * Expression with resultType string). */ - password?: AzureKeyVaultSecretReference; + servicePrincipalCredentialType?: any; + /** + * The credential of the service principal object in Azure Active Directory. If + * servicePrincipalCredentialType is 'ServicePrincipalKey', servicePrincipalCredential can be + * SecureString or AzureKeyVaultSecretReference. If servicePrincipalCredentialType is + * 'ServicePrincipalCert', servicePrincipalCredential can only be AzureKeyVaultSecretReference. + */ + servicePrincipalCredential?: SecretBaseUnion; /** * The encrypted credential used for authentication. Credentials are encrypted using the * integration runtime credential manager. Type: string (or Expression with resultType string). @@ -7160,13 +7196,13 @@ export interface OracleLinkedService { } /** - * Linked service for Google Cloud Storage. + * Microsoft Azure Cosmos Database (CosmosDB) linked service. */ -export interface GoogleCloudStorageLinkedService { +export interface CosmosDbLinkedService { /** * Polymorphic Discriminator */ - type: "GoogleCloudStorage"; + type: "CosmosDb"; /** * The integration runtime reference. */ @@ -7184,20 +7220,23 @@ export interface GoogleCloudStorageLinkedService { */ annotations?: any[]; /** - * The access key identifier of the Google Cloud Storage Identity and Access Management (IAM) - * user. Type: string (or Expression with resultType string). + * The connection string. Type: string, SecureString or AzureKeyVaultSecretReference. */ - accessKeyId?: any; + connectionString?: any; /** - * The secret access key of the Google Cloud Storage Identity and Access Management (IAM) user. + * The endpoint of the Azure CosmosDB account. Type: string (or Expression with resultType + * string) */ - secretAccessKey?: SecretBaseUnion; + accountEndpoint?: any; /** - * This value specifies the endpoint to access with the Google Cloud Storage Connector. This is - * an optional property; change it only if you want to try a different service endpoint or want - * to switch between https and http. Type: string (or Expression with resultType string). + * The name of the database. Type: string (or Expression with resultType string) */ - serviceUrl?: any; + database?: any; + /** + * The account key of the Azure CosmosDB account. Type: SecureString or + * AzureKeyVaultSecretReference. + */ + accountKey?: SecretBaseUnion; /** * The encrypted credential used for authentication. Credentials are encrypted using the * integration runtime credential manager. Type: string (or Expression with resultType string). @@ -7206,13 +7245,13 @@ export interface GoogleCloudStorageLinkedService { } /** - * Azure File Storage linked service. + * Azure Key Vault linked service. */ -export interface AzureFileStorageLinkedService { +export interface AzureKeyVaultLinkedService { /** * Polymorphic Discriminator */ - type: "AzureFileStorage"; + type: "AzureKeyVault"; /** * The integration runtime reference. */ @@ -7230,32 +7269,20 @@ export interface AzureFileStorageLinkedService { */ annotations?: any[]; /** - * Host name of the server. Type: string (or Expression with resultType string). - */ - host: any; - /** - * User ID to logon the server. Type: string (or Expression with resultType string). - */ - userId?: any; - /** - * Password to logon the server. - */ - password?: SecretBaseUnion; - /** - * The encrypted credential used for authentication. Credentials are encrypted using the - * integration runtime credential manager. Type: string (or Expression with resultType string). + * The base URL of the Azure Key Vault. e.g. https://myakv.vault.azure.net Type: string (or + * Expression with resultType string). */ - encryptedCredential?: any; + baseUrl: any; } /** - * File system linked service. + * Azure Batch linked service. */ -export interface FileServerLinkedService { +export interface AzureBatchLinkedService { /** * Polymorphic Discriminator */ - type: "FileServer"; + type: "AzureBatch"; /** * The integration runtime reference. */ @@ -7273,17 +7300,25 @@ export interface FileServerLinkedService { */ annotations?: any[]; /** - * Host name of the server. Type: string (or Expression with resultType string). + * The Azure Batch account name. Type: string (or Expression with resultType string). */ - host: any; + accountName: any; /** - * User ID to logon the server. Type: string (or Expression with resultType string). + * The Azure Batch account access key. */ - userId?: any; + accessKey?: SecretBaseUnion; /** - * Password to logon the server. + * The Azure Batch URI. Type: string (or Expression with resultType string). */ - password?: SecretBaseUnion; + batchUri: any; + /** + * The Azure Batch pool name. Type: string (or Expression with resultType string). + */ + poolName: any; + /** + * The Azure Storage linked service reference. + */ + linkedServiceName: LinkedServiceReference; /** * The encrypted credential used for authentication. Credentials are encrypted using the * integration runtime credential manager. Type: string (or Expression with resultType string). @@ -7292,13 +7327,13 @@ export interface FileServerLinkedService { } /** - * HDInsight linked service. + * Azure SQL Managed Instance linked service. */ -export interface HDInsightLinkedService { +export interface AzureSqlMILinkedService { /** * Polymorphic Discriminator */ - type: "HDInsight"; + type: "AzureSqlMI"; /** * The integration runtime reference. */ @@ -7316,49 +7351,42 @@ export interface HDInsightLinkedService { */ annotations?: any[]; /** - * HDInsight cluster URI. Type: string (or Expression with resultType string). + * The connection string. Type: string, SecureString or AzureKeyVaultSecretReference. */ - clusterUri: any; + connectionString: any; /** - * HDInsight cluster user name. Type: string (or Expression with resultType string). + * The Azure key vault secret reference of password in connection string. */ - userName?: any; + password?: AzureKeyVaultSecretReference; /** - * HDInsight cluster password. + * The ID of the service principal used to authenticate against Azure SQL Managed Instance. Type: + * string (or Expression with resultType string). */ - password?: SecretBaseUnion; + servicePrincipalId?: any; /** - * The Azure Storage linked service reference. + * The key of the service principal used to authenticate against Azure SQL Managed Instance. */ - linkedServiceName?: LinkedServiceReference; + servicePrincipalKey?: SecretBaseUnion; /** - * A reference to the Azure SQL linked service that points to the HCatalog database. + * The name or ID of the tenant to which the service principal belongs. Type: string (or + * Expression with resultType string). */ - hcatalogLinkedServiceName?: LinkedServiceReference; + tenant?: any; /** * The encrypted credential used for authentication. Credentials are encrypted using the * integration runtime credential manager. Type: string (or Expression with resultType string). */ encryptedCredential?: any; - /** - * Specify if the HDInsight is created with ESP (Enterprise Security Package). Type: Boolean. - */ - isEspEnabled?: any; - /** - * Specify the FileSystem if the main storage for the HDInsight is ADLS Gen2. Type: string (or - * Expression with resultType string). - */ - fileSystem?: any; } /** - * Common Data Service for Apps linked service. + * Microsoft Azure SQL Database linked service. */ -export interface CommonDataServiceForAppsLinkedService { +export interface AzureSqlDatabaseLinkedService { /** * Polymorphic Discriminator */ - type: "CommonDataServiceForApps"; + type: "AzureSqlDatabase"; /** * The integration runtime reference. */ @@ -7376,68 +7404,71 @@ export interface CommonDataServiceForAppsLinkedService { */ annotations?: any[]; /** - * The deployment type of the Common Data Service for Apps instance. 'Online' for Common Data - * Service for Apps Online and 'OnPremisesWithIfd' for Common Data Service for Apps on-premises - * with Ifd. Type: string (or Expression with resultType string). Possible values include: - * 'Online', 'OnPremisesWithIfd' + * The connection string. Type: string, SecureString or AzureKeyVaultSecretReference. */ - deploymentType: DynamicsDeploymentType; + connectionString: any; /** - * The host name of the on-premises Common Data Service for Apps server. The property is required - * for on-prem and not allowed for online. Type: string (or Expression with resultType string). + * The Azure key vault secret reference of password in connection string. */ - hostName?: any; + password?: AzureKeyVaultSecretReference; /** - * The port of on-premises Common Data Service for Apps server. The property is required for - * on-prem and not allowed for online. Default is 443. Type: integer (or Expression with - * resultType integer), minimum: 0. + * The ID of the service principal used to authenticate against Azure SQL Database. Type: string + * (or Expression with resultType string). */ - port?: any; + servicePrincipalId?: any; /** - * The URL to the Microsoft Common Data Service for Apps server. The property is required for - * on-line and not allowed for on-prem. Type: string (or Expression with resultType string). + * The key of the service principal used to authenticate against Azure SQL Database. */ - serviceUri?: any; + servicePrincipalKey?: SecretBaseUnion; /** - * The organization name of the Common Data Service for Apps instance. The property is required - * for on-prem and required for online when there are more than one Common Data Service for Apps - * instances associated with the user. Type: string (or Expression with resultType string). + * The name or ID of the tenant to which the service principal belongs. Type: string (or + * Expression with resultType string). */ - organizationName?: any; + tenant?: any; /** - * The authentication type to connect to Common Data Service for Apps server. 'Office365' for - * online scenario, 'Ifd' for on-premises with Ifd scenario. 'AADServicePrincipal' for - * Server-To-Server authentication in online scenario. Type: string (or Expression with - * resultType string). Possible values include: 'Office365', 'Ifd', 'AADServicePrincipal' + * The encrypted credential used for authentication. Credentials are encrypted using the + * integration runtime credential manager. Type: string (or Expression with resultType string). */ - authenticationType: DynamicsAuthenticationType; + encryptedCredential?: any; +} + +/** + * SQL Server linked service. + */ +export interface SqlServerLinkedService { /** - * User name to access the Common Data Service for Apps instance. Type: string (or Expression - * with resultType string). + * Polymorphic Discriminator + */ + type: "SqlServer"; + /** + * The integration runtime reference. + */ + connectVia?: IntegrationRuntimeReference; + /** + * Linked service description. + */ + description?: string; + /** + * Parameters for linked service. */ - username?: any; + parameters?: { [propertyName: string]: ParameterSpecification }; /** - * Password to access the Common Data Service for Apps instance. + * List of tags that can be used for describing the linked service. */ - password?: SecretBaseUnion; + annotations?: any[]; /** - * The client ID of the application in Azure Active Directory used for Server-To-Server - * authentication. Type: string (or Expression with resultType string). + * The connection string. Type: string, SecureString or AzureKeyVaultSecretReference. */ - servicePrincipalId?: any; + connectionString: any; /** - * The service principal credential type to use in Server-To-Server authentication. - * 'ServicePrincipalKey' for key/secret, 'ServicePrincipalCert' for certificate. Type: string (or - * Expression with resultType string). + * The on-premises Windows authentication user name. Type: string (or Expression with resultType + * string). */ - servicePrincipalCredentialType?: any; + userName?: any; /** - * The credential of the service principal object in Azure Active Directory. If - * servicePrincipalCredentialType is 'ServicePrincipalKey', servicePrincipalCredential can be - * SecureString or AzureKeyVaultSecretReference. If servicePrincipalCredentialType is - * 'ServicePrincipalCert', servicePrincipalCredential can only be AzureKeyVaultSecretReference. + * The on-premises Windows authentication password. */ - servicePrincipalCredential?: SecretBaseUnion; + password?: SecretBaseUnion; /** * The encrypted credential used for authentication. Credentials are encrypted using the * integration runtime credential manager. Type: string (or Expression with resultType string). @@ -7446,13 +7477,13 @@ export interface CommonDataServiceForAppsLinkedService { } /** - * Dynamics CRM linked service. + * Azure SQL Data Warehouse linked service. */ -export interface DynamicsCrmLinkedService { +export interface AzureSqlDWLinkedService { /** * Polymorphic Discriminator */ - type: "DynamicsCrm"; + type: "AzureSqlDW"; /** * The integration runtime reference. */ @@ -7470,82 +7501,92 @@ export interface DynamicsCrmLinkedService { */ annotations?: any[]; /** - * The deployment type of the Dynamics CRM instance. 'Online' for Dynamics CRM Online and - * 'OnPremisesWithIfd' for Dynamics CRM on-premises with Ifd. Type: string (or Expression with - * resultType string). Possible values include: 'Online', 'OnPremisesWithIfd' + * The connection string. Type: string, SecureString or AzureKeyVaultSecretReference. Type: + * string, SecureString or AzureKeyVaultSecretReference. */ - deploymentType: DynamicsDeploymentType; + connectionString: any; /** - * The host name of the on-premises Dynamics CRM server. The property is required for on-prem and - * not allowed for online. Type: string (or Expression with resultType string). + * The Azure key vault secret reference of password in connection string. */ - hostName?: any; + password?: AzureKeyVaultSecretReference; /** - * The port of on-premises Dynamics CRM server. The property is required for on-prem and not - * allowed for online. Default is 443. Type: integer (or Expression with resultType integer), - * minimum: 0. + * The ID of the service principal used to authenticate against Azure SQL Data Warehouse. Type: + * string (or Expression with resultType string). */ - port?: any; + servicePrincipalId?: any; /** - * The URL to the Microsoft Dynamics CRM server. The property is required for on-line and not - * allowed for on-prem. Type: string (or Expression with resultType string). + * The key of the service principal used to authenticate against Azure SQL Data Warehouse. */ - serviceUri?: any; + servicePrincipalKey?: SecretBaseUnion; /** - * The organization name of the Dynamics CRM instance. The property is required for on-prem and - * required for online when there are more than one Dynamics CRM instances associated with the - * user. Type: string (or Expression with resultType string). + * The name or ID of the tenant to which the service principal belongs. Type: string (or + * Expression with resultType string). */ - organizationName?: any; + tenant?: any; /** - * The authentication type to connect to Dynamics CRM server. 'Office365' for online scenario, - * 'Ifd' for on-premises with Ifd scenario, 'AADServicePrincipal' for Server-To-Server - * authentication in online scenario. Type: string (or Expression with resultType string). - * Possible values include: 'Office365', 'Ifd', 'AADServicePrincipal' + * The encrypted credential used for authentication. Credentials are encrypted using the + * integration runtime credential manager. Type: string (or Expression with resultType string). */ - authenticationType: DynamicsAuthenticationType; + encryptedCredential?: any; +} + +/** + * The azure table storage linked service. + */ +export interface AzureTableStorageLinkedService { /** - * User name to access the Dynamics CRM instance. Type: string (or Expression with resultType - * string). + * Polymorphic Discriminator */ - username?: any; + type: "AzureTableStorage"; /** - * Password to access the Dynamics CRM instance. + * The integration runtime reference. */ - password?: SecretBaseUnion; + connectVia?: IntegrationRuntimeReference; /** - * The client ID of the application in Azure Active Directory used for Server-To-Server - * authentication. Type: string (or Expression with resultType string). + * Linked service description. */ - servicePrincipalId?: any; + description?: string; /** - * The service principal credential type to use in Server-To-Server authentication. - * 'ServicePrincipalKey' for key/secret, 'ServicePrincipalCert' for certificate. Type: string (or - * Expression with resultType string). + * Parameters for linked service. */ - servicePrincipalCredentialType?: any; + parameters?: { [propertyName: string]: ParameterSpecification }; /** - * The credential of the service principal object in Azure Active Directory. If - * servicePrincipalCredentialType is 'ServicePrincipalKey', servicePrincipalCredential can be - * SecureString or AzureKeyVaultSecretReference. If servicePrincipalCredentialType is - * 'ServicePrincipalCert', servicePrincipalCredential can only be AzureKeyVaultSecretReference. + * List of tags that can be used for describing the linked service. */ - servicePrincipalCredential?: SecretBaseUnion; + annotations?: any[]; + /** + * The connection string. It is mutually exclusive with sasUri property. Type: string, + * SecureString or AzureKeyVaultSecretReference. + */ + connectionString?: any; + /** + * The Azure key vault secret reference of accountKey in connection string. + */ + accountKey?: AzureKeyVaultSecretReference; + /** + * SAS URI of the Azure Storage resource. It is mutually exclusive with connectionString + * property. Type: string, SecureString or AzureKeyVaultSecretReference. + */ + sasUri?: any; + /** + * The Azure key vault secret reference of sasToken in sas uri. + */ + sasToken?: AzureKeyVaultSecretReference; /** * The encrypted credential used for authentication. Credentials are encrypted using the * integration runtime credential manager. Type: string (or Expression with resultType string). */ - encryptedCredential?: any; + encryptedCredential?: string; } /** - * Dynamics linked service. + * The azure blob storage linked service. */ -export interface DynamicsLinkedService { +export interface AzureBlobStorageLinkedService { /** * Polymorphic Discriminator */ - type: "Dynamics"; + type: "AzureBlobStorage"; /** * The integration runtime reference. */ @@ -7563,80 +7604,57 @@ export interface DynamicsLinkedService { */ annotations?: any[]; /** - * The deployment type of the Dynamics instance. 'Online' for Dynamics Online and - * 'OnPremisesWithIfd' for Dynamics on-premises with Ifd. Type: string (or Expression with - * resultType string). - */ - deploymentType: any; - /** - * The host name of the on-premises Dynamics server. The property is required for on-prem and not - * allowed for online. Type: string (or Expression with resultType string). - */ - hostName?: any; - /** - * The port of on-premises Dynamics server. The property is required for on-prem and not allowed - * for online. Default is 443. Type: integer (or Expression with resultType integer), minimum: 0. - */ - port?: any; - /** - * The URL to the Microsoft Dynamics server. The property is required for on-line and not allowed - * for on-prem. Type: string (or Expression with resultType string). + * The connection string. It is mutually exclusive with sasUri, serviceEndpoint property. Type: + * string, SecureString or AzureKeyVaultSecretReference. */ - serviceUri?: any; + connectionString?: any; /** - * The organization name of the Dynamics instance. The property is required for on-prem and - * required for online when there are more than one Dynamics instances associated with the user. - * Type: string (or Expression with resultType string). + * The Azure key vault secret reference of accountKey in connection string. */ - organizationName?: any; + accountKey?: AzureKeyVaultSecretReference; /** - * The authentication type to connect to Dynamics server. 'Office365' for online scenario, 'Ifd' - * for on-premises with Ifd scenario, 'AADServicePrincipal' for Server-To-Server authentication - * in online scenario. Type: string (or Expression with resultType string). + * SAS URI of the Azure Blob Storage resource. It is mutually exclusive with connectionString, + * serviceEndpoint property. Type: string, SecureString or AzureKeyVaultSecretReference. */ - authenticationType: any; + sasUri?: any; /** - * User name to access the Dynamics instance. Type: string (or Expression with resultType - * string). + * The Azure key vault secret reference of sasToken in sas uri. */ - username?: any; + sasToken?: AzureKeyVaultSecretReference; /** - * Password to access the Dynamics instance. + * Blob service endpoint of the Azure Blob Storage resource. It is mutually exclusive with + * connectionString, sasUri property. */ - password?: SecretBaseUnion; + serviceEndpoint?: string; /** - * The client ID of the application in Azure Active Directory used for Server-To-Server - * authentication. Type: string (or Expression with resultType string). + * The ID of the service principal used to authenticate against Azure SQL Data Warehouse. Type: + * string (or Expression with resultType string). */ servicePrincipalId?: any; /** - * The service principal credential type to use in Server-To-Server authentication. - * 'ServicePrincipalKey' for key/secret, 'ServicePrincipalCert' for certificate. Type: string (or - * Expression with resultType string). + * The key of the service principal used to authenticate against Azure SQL Data Warehouse. */ - servicePrincipalCredentialType?: any; + servicePrincipalKey?: SecretBaseUnion; /** - * The credential of the service principal object in Azure Active Directory. If - * servicePrincipalCredentialType is 'ServicePrincipalKey', servicePrincipalCredential can be - * SecureString or AzureKeyVaultSecretReference. If servicePrincipalCredentialType is - * 'ServicePrincipalCert', servicePrincipalCredential can only be AzureKeyVaultSecretReference. + * The name or ID of the tenant to which the service principal belongs. Type: string (or + * Expression with resultType string). */ - servicePrincipalCredential?: SecretBaseUnion; + tenant?: any; /** * The encrypted credential used for authentication. Credentials are encrypted using the * integration runtime credential manager. Type: string (or Expression with resultType string). */ - encryptedCredential?: any; + encryptedCredential?: string; } /** - * Microsoft Azure Cosmos Database (CosmosDB) linked service. + * The storage account linked service. */ -export interface CosmosDbLinkedService { +export interface AzureStorageLinkedService { /** * Polymorphic Discriminator */ - type: "CosmosDb"; + type: "AzureStorage"; /** * The integration runtime reference. */ @@ -7654,490 +7672,456 @@ export interface CosmosDbLinkedService { */ annotations?: any[]; /** - * The connection string. Type: string, SecureString or AzureKeyVaultSecretReference. + * The connection string. It is mutually exclusive with sasUri property. Type: string, + * SecureString or AzureKeyVaultSecretReference. */ connectionString?: any; /** - * The endpoint of the Azure CosmosDB account. Type: string (or Expression with resultType - * string) + * The Azure key vault secret reference of accountKey in connection string. */ - accountEndpoint?: any; + accountKey?: AzureKeyVaultSecretReference; /** - * The name of the database. Type: string (or Expression with resultType string) + * SAS URI of the Azure Storage resource. It is mutually exclusive with connectionString + * property. Type: string, SecureString or AzureKeyVaultSecretReference. */ - database?: any; + sasUri?: any; /** - * The account key of the Azure CosmosDB account. Type: SecureString or - * AzureKeyVaultSecretReference. + * The Azure key vault secret reference of sasToken in sas uri. */ - accountKey?: SecretBaseUnion; + sasToken?: AzureKeyVaultSecretReference; /** * The encrypted credential used for authentication. Credentials are encrypted using the * integration runtime credential manager. Type: string (or Expression with resultType string). */ - encryptedCredential?: any; + encryptedCredential?: string; } /** - * Azure Key Vault linked service. + * The snowflake dataset. */ -export interface AzureKeyVaultLinkedService { +export interface SnowflakeDataset { /** * Polymorphic Discriminator */ - type: "AzureKeyVault"; + type: "SnowflakeTable"; /** - * The integration runtime reference. + * Dataset description. */ - connectVia?: IntegrationRuntimeReference; + description?: string; /** - * Linked service description. + * Columns that define the structure of the dataset. Type: array (or Expression with resultType + * array), itemType: DatasetDataElement. */ - description?: string; + structure?: any; /** - * Parameters for linked service. + * Columns that define the physical type schema of the dataset. Type: array (or Expression with + * resultType array), itemType: DatasetSchemaDataElement. + */ + schema?: any; + /** + * Linked service reference. + */ + linkedServiceName: LinkedServiceReference; + /** + * Parameters for dataset. */ parameters?: { [propertyName: string]: ParameterSpecification }; /** - * List of tags that can be used for describing the linked service. + * List of tags that can be used for describing the Dataset. */ annotations?: any[]; /** - * The base URL of the Azure Key Vault. e.g. https://myakv.vault.azure.net Type: string (or - * Expression with resultType string). + * The folder that this Dataset is in. If not specified, Dataset will appear at the root level. */ - baseUrl: any; + folder?: DatasetFolder; + /** + * The schema name of the Snowflake database. Type: string (or Expression with resultType + * string). + */ + snowflakeDatasetSchema?: any; + /** + * The table name of the Snowflake database. Type: string (or Expression with resultType string). + */ + table?: any; } /** - * Azure Batch linked service. + * Google AdWords service dataset. */ -export interface AzureBatchLinkedService { +export interface GoogleAdWordsObjectDataset { /** * Polymorphic Discriminator */ - type: "AzureBatch"; - /** - * The integration runtime reference. - */ - connectVia?: IntegrationRuntimeReference; + type: "GoogleAdWordsObject"; /** - * Linked service description. + * Dataset description. */ description?: string; /** - * Parameters for linked service. - */ - parameters?: { [propertyName: string]: ParameterSpecification }; - /** - * List of tags that can be used for describing the linked service. + * Columns that define the structure of the dataset. Type: array (or Expression with resultType + * array), itemType: DatasetDataElement. */ - annotations?: any[]; + structure?: any; /** - * The Azure Batch account name. Type: string (or Expression with resultType string). + * Columns that define the physical type schema of the dataset. Type: array (or Expression with + * resultType array), itemType: DatasetSchemaDataElement. */ - accountName: any; + schema?: any; /** - * The Azure Batch account access key. + * Linked service reference. */ - accessKey?: SecretBaseUnion; + linkedServiceName: LinkedServiceReference; /** - * The Azure Batch URI. Type: string (or Expression with resultType string). + * Parameters for dataset. */ - batchUri: any; + parameters?: { [propertyName: string]: ParameterSpecification }; /** - * The Azure Batch pool name. Type: string (or Expression with resultType string). + * List of tags that can be used for describing the Dataset. */ - poolName: any; + annotations?: any[]; /** - * The Azure Storage linked service reference. + * The folder that this Dataset is in. If not specified, Dataset will appear at the root level. */ - linkedServiceName: LinkedServiceReference; + folder?: DatasetFolder; /** - * The encrypted credential used for authentication. Credentials are encrypted using the - * integration runtime credential manager. Type: string (or Expression with resultType string). + * The table name. Type: string (or Expression with resultType string). */ - encryptedCredential?: any; + tableName?: any; } /** - * Azure SQL Managed Instance linked service. + * The Azure Data Explorer (Kusto) dataset. */ -export interface AzureSqlMILinkedService { +export interface AzureDataExplorerTableDataset { /** * Polymorphic Discriminator */ - type: "AzureSqlMI"; - /** - * The integration runtime reference. - */ - connectVia?: IntegrationRuntimeReference; + type: "AzureDataExplorerTable"; /** - * Linked service description. + * Dataset description. */ description?: string; /** - * Parameters for linked service. - */ - parameters?: { [propertyName: string]: ParameterSpecification }; - /** - * List of tags that can be used for describing the linked service. + * Columns that define the structure of the dataset. Type: array (or Expression with resultType + * array), itemType: DatasetDataElement. */ - annotations?: any[]; + structure?: any; /** - * The connection string. Type: string, SecureString or AzureKeyVaultSecretReference. + * Columns that define the physical type schema of the dataset. Type: array (or Expression with + * resultType array), itemType: DatasetSchemaDataElement. */ - connectionString: any; + schema?: any; /** - * The Azure key vault secret reference of password in connection string. + * Linked service reference. */ - password?: AzureKeyVaultSecretReference; + linkedServiceName: LinkedServiceReference; /** - * The ID of the service principal used to authenticate against Azure SQL Managed Instance. Type: - * string (or Expression with resultType string). + * Parameters for dataset. */ - servicePrincipalId?: any; + parameters?: { [propertyName: string]: ParameterSpecification }; /** - * The key of the service principal used to authenticate against Azure SQL Managed Instance. + * List of tags that can be used for describing the Dataset. */ - servicePrincipalKey?: SecretBaseUnion; + annotations?: any[]; /** - * The name or ID of the tenant to which the service principal belongs. Type: string (or - * Expression with resultType string). + * The folder that this Dataset is in. If not specified, Dataset will appear at the root level. */ - tenant?: any; + folder?: DatasetFolder; /** - * The encrypted credential used for authentication. Credentials are encrypted using the - * integration runtime credential manager. Type: string (or Expression with resultType string). + * The table name of the Azure Data Explorer database. Type: string (or Expression with + * resultType string). */ - encryptedCredential?: any; + table?: any; } /** - * Microsoft Azure SQL Database linked service. + * Oracle Service Cloud dataset. */ -export interface AzureSqlDatabaseLinkedService { +export interface OracleServiceCloudObjectDataset { /** * Polymorphic Discriminator */ - type: "AzureSqlDatabase"; - /** - * The integration runtime reference. - */ - connectVia?: IntegrationRuntimeReference; + type: "OracleServiceCloudObject"; /** - * Linked service description. + * Dataset description. */ description?: string; /** - * Parameters for linked service. - */ - parameters?: { [propertyName: string]: ParameterSpecification }; - /** - * List of tags that can be used for describing the linked service. + * Columns that define the structure of the dataset. Type: array (or Expression with resultType + * array), itemType: DatasetDataElement. */ - annotations?: any[]; + structure?: any; /** - * The connection string. Type: string, SecureString or AzureKeyVaultSecretReference. + * Columns that define the physical type schema of the dataset. Type: array (or Expression with + * resultType array), itemType: DatasetSchemaDataElement. */ - connectionString: any; + schema?: any; /** - * The Azure key vault secret reference of password in connection string. + * Linked service reference. */ - password?: AzureKeyVaultSecretReference; + linkedServiceName: LinkedServiceReference; /** - * The ID of the service principal used to authenticate against Azure SQL Database. Type: string - * (or Expression with resultType string). + * Parameters for dataset. */ - servicePrincipalId?: any; + parameters?: { [propertyName: string]: ParameterSpecification }; /** - * The key of the service principal used to authenticate against Azure SQL Database. + * List of tags that can be used for describing the Dataset. */ - servicePrincipalKey?: SecretBaseUnion; + annotations?: any[]; /** - * The name or ID of the tenant to which the service principal belongs. Type: string (or - * Expression with resultType string). + * The folder that this Dataset is in. If not specified, Dataset will appear at the root level. */ - tenant?: any; + folder?: DatasetFolder; /** - * The encrypted credential used for authentication. Credentials are encrypted using the - * integration runtime credential manager. Type: string (or Expression with resultType string). + * The table name. Type: string (or Expression with resultType string). */ - encryptedCredential?: any; + tableName?: any; } /** - * SQL Server linked service. + * The path of the Dynamics AX OData entity. */ -export interface SqlServerLinkedService { +export interface DynamicsAXResourceDataset { /** * Polymorphic Discriminator */ - type: "SqlServer"; + type: "DynamicsAXResource"; /** - * The integration runtime reference. + * Dataset description. */ - connectVia?: IntegrationRuntimeReference; + description?: string; /** - * Linked service description. + * Columns that define the structure of the dataset. Type: array (or Expression with resultType + * array), itemType: DatasetDataElement. */ - description?: string; + structure?: any; /** - * Parameters for linked service. + * Columns that define the physical type schema of the dataset. Type: array (or Expression with + * resultType array), itemType: DatasetSchemaDataElement. */ - parameters?: { [propertyName: string]: ParameterSpecification }; + schema?: any; /** - * List of tags that can be used for describing the linked service. + * Linked service reference. */ - annotations?: any[]; + linkedServiceName: LinkedServiceReference; /** - * The connection string. Type: string, SecureString or AzureKeyVaultSecretReference. + * Parameters for dataset. */ - connectionString: any; + parameters?: { [propertyName: string]: ParameterSpecification }; /** - * The on-premises Windows authentication user name. Type: string (or Expression with resultType - * string). + * List of tags that can be used for describing the Dataset. */ - userName?: any; + annotations?: any[]; /** - * The on-premises Windows authentication password. + * The folder that this Dataset is in. If not specified, Dataset will appear at the root level. */ - password?: SecretBaseUnion; + folder?: DatasetFolder; /** - * The encrypted credential used for authentication. Credentials are encrypted using the - * integration runtime credential manager. Type: string (or Expression with resultType string). + * The path of the Dynamics AX OData entity. Type: string (or Expression with resultType string). */ - encryptedCredential?: any; + path: any; } /** - * Azure SQL Data Warehouse linked service. + * Responsys dataset. */ -export interface AzureSqlDWLinkedService { +export interface ResponsysObjectDataset { /** * Polymorphic Discriminator */ - type: "AzureSqlDW"; - /** - * The integration runtime reference. - */ - connectVia?: IntegrationRuntimeReference; + type: "ResponsysObject"; /** - * Linked service description. + * Dataset description. */ description?: string; /** - * Parameters for linked service. - */ - parameters?: { [propertyName: string]: ParameterSpecification }; - /** - * List of tags that can be used for describing the linked service. + * Columns that define the structure of the dataset. Type: array (or Expression with resultType + * array), itemType: DatasetDataElement. */ - annotations?: any[]; + structure?: any; /** - * The connection string. Type: string, SecureString or AzureKeyVaultSecretReference. Type: - * string, SecureString or AzureKeyVaultSecretReference. + * Columns that define the physical type schema of the dataset. Type: array (or Expression with + * resultType array), itemType: DatasetSchemaDataElement. */ - connectionString: any; + schema?: any; /** - * The Azure key vault secret reference of password in connection string. + * Linked service reference. */ - password?: AzureKeyVaultSecretReference; + linkedServiceName: LinkedServiceReference; /** - * The ID of the service principal used to authenticate against Azure SQL Data Warehouse. Type: - * string (or Expression with resultType string). + * Parameters for dataset. */ - servicePrincipalId?: any; + parameters?: { [propertyName: string]: ParameterSpecification }; /** - * The key of the service principal used to authenticate against Azure SQL Data Warehouse. + * List of tags that can be used for describing the Dataset. */ - servicePrincipalKey?: SecretBaseUnion; + annotations?: any[]; /** - * The name or ID of the tenant to which the service principal belongs. Type: string (or - * Expression with resultType string). + * The folder that this Dataset is in. If not specified, Dataset will appear at the root level. */ - tenant?: any; + folder?: DatasetFolder; /** - * The encrypted credential used for authentication. Credentials are encrypted using the - * integration runtime credential manager. Type: string (or Expression with resultType string). + * The table name. Type: string (or Expression with resultType string). */ - encryptedCredential?: any; + tableName?: any; } /** - * The azure table storage linked service. + * Salesforce Marketing Cloud dataset. */ -export interface AzureTableStorageLinkedService { +export interface SalesforceMarketingCloudObjectDataset { /** * Polymorphic Discriminator */ - type: "AzureTableStorage"; - /** - * The integration runtime reference. - */ - connectVia?: IntegrationRuntimeReference; + type: "SalesforceMarketingCloudObject"; /** - * Linked service description. + * Dataset description. */ description?: string; /** - * Parameters for linked service. + * Columns that define the structure of the dataset. Type: array (or Expression with resultType + * array), itemType: DatasetDataElement. */ - parameters?: { [propertyName: string]: ParameterSpecification }; + structure?: any; /** - * List of tags that can be used for describing the linked service. + * Columns that define the physical type schema of the dataset. Type: array (or Expression with + * resultType array), itemType: DatasetSchemaDataElement. */ - annotations?: any[]; + schema?: any; /** - * The connection string. It is mutually exclusive with sasUri property. Type: string, - * SecureString or AzureKeyVaultSecretReference. + * Linked service reference. */ - connectionString?: any; + linkedServiceName: LinkedServiceReference; /** - * The Azure key vault secret reference of accountKey in connection string. + * Parameters for dataset. */ - accountKey?: AzureKeyVaultSecretReference; + parameters?: { [propertyName: string]: ParameterSpecification }; /** - * SAS URI of the Azure Storage resource. It is mutually exclusive with connectionString - * property. Type: string, SecureString or AzureKeyVaultSecretReference. + * List of tags that can be used for describing the Dataset. */ - sasUri?: any; + annotations?: any[]; /** - * The Azure key vault secret reference of sasToken in sas uri. + * The folder that this Dataset is in. If not specified, Dataset will appear at the root level. */ - sasToken?: AzureKeyVaultSecretReference; + folder?: DatasetFolder; /** - * The encrypted credential used for authentication. Credentials are encrypted using the - * integration runtime credential manager. Type: string (or Expression with resultType string). + * The table name. Type: string (or Expression with resultType string). */ - encryptedCredential?: string; + tableName?: any; } /** - * The azure blob storage linked service. + * Vertica dataset. */ -export interface AzureBlobStorageLinkedService { +export interface VerticaTableDataset { /** * Polymorphic Discriminator */ - type: "AzureBlobStorage"; - /** - * The integration runtime reference. - */ - connectVia?: IntegrationRuntimeReference; + type: "VerticaTable"; /** - * Linked service description. + * Dataset description. */ description?: string; /** - * Parameters for linked service. - */ - parameters?: { [propertyName: string]: ParameterSpecification }; - /** - * List of tags that can be used for describing the linked service. - */ - annotations?: any[]; - /** - * The connection string. It is mutually exclusive with sasUri, serviceEndpoint property. Type: - * string, SecureString or AzureKeyVaultSecretReference. + * Columns that define the structure of the dataset. Type: array (or Expression with resultType + * array), itemType: DatasetDataElement. */ - connectionString?: any; + structure?: any; /** - * The Azure key vault secret reference of accountKey in connection string. + * Columns that define the physical type schema of the dataset. Type: array (or Expression with + * resultType array), itemType: DatasetSchemaDataElement. */ - accountKey?: AzureKeyVaultSecretReference; + schema?: any; /** - * SAS URI of the Azure Blob Storage resource. It is mutually exclusive with connectionString, - * serviceEndpoint property. Type: string, SecureString or AzureKeyVaultSecretReference. + * Linked service reference. */ - sasUri?: any; + linkedServiceName: LinkedServiceReference; /** - * The Azure key vault secret reference of sasToken in sas uri. + * Parameters for dataset. */ - sasToken?: AzureKeyVaultSecretReference; + parameters?: { [propertyName: string]: ParameterSpecification }; /** - * Blob service endpoint of the Azure Blob Storage resource. It is mutually exclusive with - * connectionString, sasUri property. + * List of tags that can be used for describing the Dataset. */ - serviceEndpoint?: string; + annotations?: any[]; /** - * The ID of the service principal used to authenticate against Azure SQL Data Warehouse. Type: - * string (or Expression with resultType string). + * The folder that this Dataset is in. If not specified, Dataset will appear at the root level. */ - servicePrincipalId?: any; + folder?: DatasetFolder; /** - * The key of the service principal used to authenticate against Azure SQL Data Warehouse. + * This property will be retired. Please consider using schema + table properties instead. */ - servicePrincipalKey?: SecretBaseUnion; + tableName?: any; /** - * The name or ID of the tenant to which the service principal belongs. Type: string (or - * Expression with resultType string). + * The table name of the Vertica. Type: string (or Expression with resultType string). */ - tenant?: any; + table?: any; /** - * The encrypted credential used for authentication. Credentials are encrypted using the - * integration runtime credential manager. Type: string (or Expression with resultType string). + * The schema name of the Vertica. Type: string (or Expression with resultType string). */ - encryptedCredential?: string; + verticaTableDatasetSchema?: any; } /** - * The storage account linked service. + * Netezza dataset. */ -export interface AzureStorageLinkedService { +export interface NetezzaTableDataset { /** * Polymorphic Discriminator */ - type: "AzureStorage"; + type: "NetezzaTable"; /** - * The integration runtime reference. + * Dataset description. */ - connectVia?: IntegrationRuntimeReference; + description?: string; /** - * Linked service description. + * Columns that define the structure of the dataset. Type: array (or Expression with resultType + * array), itemType: DatasetDataElement. */ - description?: string; + structure?: any; /** - * Parameters for linked service. + * Columns that define the physical type schema of the dataset. Type: array (or Expression with + * resultType array), itemType: DatasetSchemaDataElement. */ - parameters?: { [propertyName: string]: ParameterSpecification }; + schema?: any; /** - * List of tags that can be used for describing the linked service. + * Linked service reference. */ - annotations?: any[]; + linkedServiceName: LinkedServiceReference; /** - * The connection string. It is mutually exclusive with sasUri property. Type: string, - * SecureString or AzureKeyVaultSecretReference. + * Parameters for dataset. */ - connectionString?: any; + parameters?: { [propertyName: string]: ParameterSpecification }; /** - * The Azure key vault secret reference of accountKey in connection string. + * List of tags that can be used for describing the Dataset. */ - accountKey?: AzureKeyVaultSecretReference; + annotations?: any[]; /** - * SAS URI of the Azure Storage resource. It is mutually exclusive with connectionString - * property. Type: string, SecureString or AzureKeyVaultSecretReference. + * The folder that this Dataset is in. If not specified, Dataset will appear at the root level. */ - sasUri?: any; + folder?: DatasetFolder; /** - * The Azure key vault secret reference of sasToken in sas uri. + * This property will be retired. Please consider using schema + table properties instead. */ - sasToken?: AzureKeyVaultSecretReference; + tableName?: any; /** - * The encrypted credential used for authentication. Credentials are encrypted using the - * integration runtime credential manager. Type: string (or Expression with resultType string). + * The table name of the Netezza. Type: string (or Expression with resultType string). */ - encryptedCredential?: string; + table?: any; + /** + * The schema name of the Netezza. Type: string (or Expression with resultType string). + */ + netezzaTableDatasetSchema?: any; } /** - * Google AdWords service dataset. + * Zoho server dataset. */ -export interface GoogleAdWordsObjectDataset { +export interface ZohoObjectDataset { /** * Polymorphic Discriminator */ - type: "GoogleAdWordsObject"; + type: "ZohoObject"; /** * Dataset description. */ @@ -8175,13 +8159,13 @@ export interface GoogleAdWordsObjectDataset { } /** - * The Azure Data Explorer (Kusto) dataset. + * Xero Service dataset. */ -export interface AzureDataExplorerTableDataset { +export interface XeroObjectDataset { /** * Polymorphic Discriminator */ - type: "AzureDataExplorerTable"; + type: "XeroObject"; /** * Dataset description. */ @@ -8213,20 +8197,19 @@ export interface AzureDataExplorerTableDataset { */ folder?: DatasetFolder; /** - * The table name of the Azure Data Explorer database. Type: string (or Expression with - * resultType string). + * The table name. Type: string (or Expression with resultType string). */ - table?: any; + tableName?: any; } /** - * Oracle Service Cloud dataset. + * Square Service dataset. */ -export interface OracleServiceCloudObjectDataset { +export interface SquareObjectDataset { /** * Polymorphic Discriminator */ - type: "OracleServiceCloudObject"; + type: "SquareObject"; /** * Dataset description. */ @@ -8264,13 +8247,13 @@ export interface OracleServiceCloudObjectDataset { } /** - * The path of the Dynamics AX OData entity. + * Spark Server dataset. */ -export interface DynamicsAXResourceDataset { +export interface SparkObjectDataset { /** * Polymorphic Discriminator */ - type: "DynamicsAXResource"; + type: "SparkObject"; /** * Dataset description. */ @@ -8302,19 +8285,27 @@ export interface DynamicsAXResourceDataset { */ folder?: DatasetFolder; /** - * The path of the Dynamics AX OData entity. Type: string (or Expression with resultType string). + * This property will be retired. Please consider using schema + table properties instead. */ - path: any; + tableName?: any; + /** + * The table name of the Spark. Type: string (or Expression with resultType string). + */ + table?: any; + /** + * The schema name of the Spark. Type: string (or Expression with resultType string). + */ + sparkObjectDatasetSchema?: any; } /** - * Responsys dataset. + * Shopify Service dataset. */ -export interface ResponsysObjectDataset { +export interface ShopifyObjectDataset { /** * Polymorphic Discriminator */ - type: "ResponsysObject"; + type: "ShopifyObject"; /** * Dataset description. */ @@ -8352,13 +8343,13 @@ export interface ResponsysObjectDataset { } /** - * Salesforce Marketing Cloud dataset. + * ServiceNow server dataset. */ -export interface SalesforceMarketingCloudObjectDataset { +export interface ServiceNowObjectDataset { /** * Polymorphic Discriminator */ - type: "SalesforceMarketingCloudObject"; + type: "ServiceNowObject"; /** * Dataset description. */ @@ -8396,13 +8387,13 @@ export interface SalesforceMarketingCloudObjectDataset { } /** - * Vertica dataset. + * QuickBooks server dataset. */ -export interface VerticaTableDataset { +export interface QuickBooksObjectDataset { /** * Polymorphic Discriminator */ - type: "VerticaTable"; + type: "QuickBooksObject"; /** * Dataset description. */ @@ -8434,27 +8425,19 @@ export interface VerticaTableDataset { */ folder?: DatasetFolder; /** - * This property will be retired. Please consider using schema + table properties instead. + * The table name. Type: string (or Expression with resultType string). */ tableName?: any; - /** - * The table name of the Vertica. Type: string (or Expression with resultType string). - */ - table?: any; - /** - * The schema name of the Vertica. Type: string (or Expression with resultType string). - */ - verticaTableDatasetSchema?: any; } /** - * Netezza dataset. + * Presto server dataset. */ -export interface NetezzaTableDataset { +export interface PrestoObjectDataset { /** * Polymorphic Discriminator */ - type: "NetezzaTable"; + type: "PrestoObject"; /** * Dataset description. */ @@ -8490,23 +8473,23 @@ export interface NetezzaTableDataset { */ tableName?: any; /** - * The table name of the Netezza. Type: string (or Expression with resultType string). + * The table name of the Presto. Type: string (or Expression with resultType string). */ table?: any; /** - * The schema name of the Netezza. Type: string (or Expression with resultType string). + * The schema name of the Presto. Type: string (or Expression with resultType string). */ - netezzaTableDatasetSchema?: any; + prestoObjectDatasetSchema?: any; } /** - * Zoho server dataset. + * Phoenix server dataset. */ -export interface ZohoObjectDataset { +export interface PhoenixObjectDataset { /** * Polymorphic Discriminator */ - type: "ZohoObject"; + type: "PhoenixObject"; /** * Dataset description. */ @@ -8538,19 +8521,27 @@ export interface ZohoObjectDataset { */ folder?: DatasetFolder; /** - * The table name. Type: string (or Expression with resultType string). + * This property will be retired. Please consider using schema + table properties instead. */ tableName?: any; + /** + * The table name of the Phoenix. Type: string (or Expression with resultType string). + */ + table?: any; + /** + * The schema name of the Phoenix. Type: string (or Expression with resultType string). + */ + phoenixObjectDatasetSchema?: any; } /** - * Xero Service dataset. + * Paypal Service dataset. */ -export interface XeroObjectDataset { +export interface PaypalObjectDataset { /** * Polymorphic Discriminator */ - type: "XeroObject"; + type: "PaypalObject"; /** * Dataset description. */ @@ -8588,13 +8579,13 @@ export interface XeroObjectDataset { } /** - * Square Service dataset. + * Marketo server dataset. */ -export interface SquareObjectDataset { +export interface MarketoObjectDataset { /** * Polymorphic Discriminator */ - type: "SquareObject"; + type: "MarketoObject"; /** * Dataset description. */ @@ -8632,13 +8623,13 @@ export interface SquareObjectDataset { } /** - * Spark Server dataset. + * Azure Database for MariaDB dataset. */ -export interface SparkObjectDataset { +export interface AzureMariaDBTableDataset { /** * Polymorphic Discriminator */ - type: "SparkObject"; + type: "AzureMariaDBTable"; /** * Dataset description. */ @@ -8670,27 +8661,19 @@ export interface SparkObjectDataset { */ folder?: DatasetFolder; /** - * This property will be retired. Please consider using schema + table properties instead. + * The table name. Type: string (or Expression with resultType string). */ tableName?: any; - /** - * The table name of the Spark. Type: string (or Expression with resultType string). - */ - table?: any; - /** - * The schema name of the Spark. Type: string (or Expression with resultType string). - */ - sparkObjectDatasetSchema?: any; } /** - * Shopify Service dataset. + * MariaDB server dataset. */ -export interface ShopifyObjectDataset { +export interface MariaDBTableDataset { /** * Polymorphic Discriminator */ - type: "ShopifyObject"; + type: "MariaDBTable"; /** * Dataset description. */ @@ -8728,13 +8711,13 @@ export interface ShopifyObjectDataset { } /** - * ServiceNow server dataset. + * Magento server dataset. */ -export interface ServiceNowObjectDataset { +export interface MagentoObjectDataset { /** * Polymorphic Discriminator */ - type: "ServiceNowObject"; + type: "MagentoObject"; /** * Dataset description. */ @@ -8772,13 +8755,13 @@ export interface ServiceNowObjectDataset { } /** - * QuickBooks server dataset. + * Jira Service dataset. */ -export interface QuickBooksObjectDataset { +export interface JiraObjectDataset { /** * Polymorphic Discriminator */ - type: "QuickBooksObject"; + type: "JiraObject"; /** * Dataset description. */ @@ -8816,13 +8799,13 @@ export interface QuickBooksObjectDataset { } /** - * Presto server dataset. + * Impala server dataset. */ -export interface PrestoObjectDataset { +export interface ImpalaObjectDataset { /** * Polymorphic Discriminator */ - type: "PrestoObject"; + type: "ImpalaObject"; /** * Dataset description. */ @@ -8858,23 +8841,23 @@ export interface PrestoObjectDataset { */ tableName?: any; /** - * The table name of the Presto. Type: string (or Expression with resultType string). + * The table name of the Impala. Type: string (or Expression with resultType string). */ table?: any; /** - * The schema name of the Presto. Type: string (or Expression with resultType string). + * The schema name of the Impala. Type: string (or Expression with resultType string). */ - prestoObjectDatasetSchema?: any; + impalaObjectDatasetSchema?: any; } /** - * Phoenix server dataset. + * Hubspot Service dataset. */ -export interface PhoenixObjectDataset { +export interface HubspotObjectDataset { /** * Polymorphic Discriminator */ - type: "PhoenixObject"; + type: "HubspotObject"; /** * Dataset description. */ @@ -8906,27 +8889,19 @@ export interface PhoenixObjectDataset { */ folder?: DatasetFolder; /** - * This property will be retired. Please consider using schema + table properties instead. + * The table name. Type: string (or Expression with resultType string). */ tableName?: any; - /** - * The table name of the Phoenix. Type: string (or Expression with resultType string). - */ - table?: any; - /** - * The schema name of the Phoenix. Type: string (or Expression with resultType string). - */ - phoenixObjectDatasetSchema?: any; } /** - * Paypal Service dataset. + * Hive Server dataset. */ -export interface PaypalObjectDataset { +export interface HiveObjectDataset { /** * Polymorphic Discriminator */ - type: "PaypalObject"; + type: "HiveObject"; /** * Dataset description. */ @@ -8958,19 +8933,27 @@ export interface PaypalObjectDataset { */ folder?: DatasetFolder; /** - * The table name. Type: string (or Expression with resultType string). + * This property will be retired. Please consider using schema + table properties instead. */ tableName?: any; + /** + * The table name of the Hive. Type: string (or Expression with resultType string). + */ + table?: any; + /** + * The schema name of the Hive. Type: string (or Expression with resultType string). + */ + hiveObjectDatasetSchema?: any; } /** - * Marketo server dataset. + * HBase server dataset. */ -export interface MarketoObjectDataset { +export interface HBaseObjectDataset { /** * Polymorphic Discriminator */ - type: "MarketoObject"; + type: "HBaseObject"; /** * Dataset description. */ @@ -9008,13 +8991,13 @@ export interface MarketoObjectDataset { } /** - * Azure Database for MariaDB dataset. + * Greenplum Database dataset. */ -export interface AzureMariaDBTableDataset { +export interface GreenplumTableDataset { /** * Polymorphic Discriminator */ - type: "AzureMariaDBTable"; + type: "GreenplumTable"; /** * Dataset description. */ @@ -9046,19 +9029,27 @@ export interface AzureMariaDBTableDataset { */ folder?: DatasetFolder; /** - * The table name. Type: string (or Expression with resultType string). + * This property will be retired. Please consider using schema + table properties instead. */ tableName?: any; + /** + * The table name of Greenplum. Type: string (or Expression with resultType string). + */ + table?: any; + /** + * The schema name of Greenplum. Type: string (or Expression with resultType string). + */ + greenplumTableDatasetSchema?: any; } /** - * MariaDB server dataset. + * Google BigQuery service dataset. */ -export interface MariaDBTableDataset { +export interface GoogleBigQueryObjectDataset { /** * Polymorphic Discriminator */ - type: "MariaDBTable"; + type: "GoogleBigQueryObject"; /** * Dataset description. */ @@ -9090,19 +9081,27 @@ export interface MariaDBTableDataset { */ folder?: DatasetFolder; /** - * The table name. Type: string (or Expression with resultType string). + * This property will be retired. Please consider using database + table properties instead. */ tableName?: any; + /** + * The table name of the Google BigQuery. Type: string (or Expression with resultType string). + */ + table?: any; + /** + * The database name of the Google BigQuery. Type: string (or Expression with resultType string). + */ + dataset?: any; } /** - * Magento server dataset. + * Eloqua server dataset. */ -export interface MagentoObjectDataset { +export interface EloquaObjectDataset { /** * Polymorphic Discriminator */ - type: "MagentoObject"; + type: "EloquaObject"; /** * Dataset description. */ @@ -9140,13 +9139,13 @@ export interface MagentoObjectDataset { } /** - * Jira Service dataset. + * Drill server dataset. */ -export interface JiraObjectDataset { +export interface DrillTableDataset { /** * Polymorphic Discriminator */ - type: "JiraObject"; + type: "DrillTable"; /** * Dataset description. */ @@ -9178,19 +9177,27 @@ export interface JiraObjectDataset { */ folder?: DatasetFolder; /** - * The table name. Type: string (or Expression with resultType string). + * This property will be retired. Please consider using schema + table properties instead. */ tableName?: any; + /** + * The table name of the Drill. Type: string (or Expression with resultType string). + */ + table?: any; + /** + * The schema name of the Drill. Type: string (or Expression with resultType string). + */ + drillTableDatasetSchema?: any; } /** - * Impala server dataset. + * Couchbase server dataset. */ -export interface ImpalaObjectDataset { +export interface CouchbaseTableDataset { /** * Polymorphic Discriminator */ - type: "ImpalaObject"; + type: "CouchbaseTable"; /** * Dataset description. */ @@ -9222,27 +9229,19 @@ export interface ImpalaObjectDataset { */ folder?: DatasetFolder; /** - * This property will be retired. Please consider using schema + table properties instead. + * The table name. Type: string (or Expression with resultType string). */ tableName?: any; - /** - * The table name of the Impala. Type: string (or Expression with resultType string). - */ - table?: any; - /** - * The schema name of the Impala. Type: string (or Expression with resultType string). - */ - impalaObjectDatasetSchema?: any; } /** - * Hubspot Service dataset. + * Concur Service dataset. */ -export interface HubspotObjectDataset { +export interface ConcurObjectDataset { /** * Polymorphic Discriminator */ - type: "HubspotObject"; + type: "ConcurObject"; /** * Dataset description. */ @@ -9280,13 +9279,13 @@ export interface HubspotObjectDataset { } /** - * Hive Server dataset. + * Azure PostgreSQL dataset. */ -export interface HiveObjectDataset { +export interface AzurePostgreSqlTableDataset { /** * Polymorphic Discriminator */ - type: "HiveObject"; + type: "AzurePostgreSqlTable"; /** * Dataset description. */ @@ -9318,27 +9317,30 @@ export interface HiveObjectDataset { */ folder?: DatasetFolder; /** - * This property will be retired. Please consider using schema + table properties instead. + * The table name of the Azure PostgreSQL database which includes both schema and table. Type: + * string (or Expression with resultType string). */ tableName?: any; /** - * The table name of the Hive. Type: string (or Expression with resultType string). + * The table name of the Azure PostgreSQL database. Type: string (or Expression with resultType + * string). */ table?: any; /** - * The schema name of the Hive. Type: string (or Expression with resultType string). + * The schema name of the Azure PostgreSQL database. Type: string (or Expression with resultType + * string). */ - hiveObjectDatasetSchema?: any; + azurePostgreSqlTableDatasetSchema?: any; } /** - * HBase server dataset. + * Amazon Marketplace Web Service dataset. */ -export interface HBaseObjectDataset { +export interface AmazonMWSObjectDataset { /** * Polymorphic Discriminator */ - type: "HBaseObject"; + type: "AmazonMWSObject"; /** * Dataset description. */ @@ -9376,161 +9378,276 @@ export interface HBaseObjectDataset { } /** - * Greenplum Database dataset. + * Contains the possible cases for DatasetCompression. */ -export interface GreenplumTableDataset { +export type DatasetCompressionUnion = DatasetCompression | DatasetZipDeflateCompression | DatasetDeflateCompression | DatasetGZipCompression | DatasetBZip2Compression; + +/** + * The compression method used on a dataset. + */ +export interface DatasetCompression { /** * Polymorphic Discriminator */ - type: "GreenplumTable"; + type: "DatasetCompression"; + /** + * Describes unknown properties. The value of an unknown property can be of "any" type. + */ + [property: string]: any; +} + +/** + * The ZipDeflate compression method used on a dataset. + */ +export interface DatasetZipDeflateCompression { + /** + * Polymorphic Discriminator + */ + type: "ZipDeflate"; + /** + * The ZipDeflate compression level. + */ + level?: any; +} + +/** + * The Deflate compression method used on a dataset. + */ +export interface DatasetDeflateCompression { + /** + * Polymorphic Discriminator + */ + type: "Deflate"; + /** + * The Deflate compression level. + */ + level?: any; +} + +/** + * The GZip compression method used on a dataset. + */ +export interface DatasetGZipCompression { + /** + * Polymorphic Discriminator + */ + type: "GZip"; + /** + * The GZip compression level. + */ + level?: any; +} + +/** + * The BZip2 compression method used on a dataset. + */ +export interface DatasetBZip2Compression { + /** + * Polymorphic Discriminator + */ + type: "BZip2"; +} + +/** + * Contains the possible cases for DatasetStorageFormat. + */ +export type DatasetStorageFormatUnion = DatasetStorageFormat | ParquetFormat | OrcFormat | AvroFormat | JsonFormat | TextFormat; + +/** + * The format definition of a storage. + */ +export interface DatasetStorageFormat { /** - * Dataset description. + * Polymorphic Discriminator */ - description?: string; + type: "DatasetStorageFormat"; /** - * Columns that define the structure of the dataset. Type: array (or Expression with resultType - * array), itemType: DatasetDataElement. + * Serializer. Type: string (or Expression with resultType string). */ - structure?: any; + serializer?: any; /** - * Columns that define the physical type schema of the dataset. Type: array (or Expression with - * resultType array), itemType: DatasetSchemaDataElement. + * Deserializer. Type: string (or Expression with resultType string). */ - schema?: any; + deserializer?: any; /** - * Linked service reference. + * Describes unknown properties. The value of an unknown property can be of "any" type. */ - linkedServiceName: LinkedServiceReference; + [property: string]: any; +} + +/** + * The data stored in Parquet format. + */ +export interface ParquetFormat { /** - * Parameters for dataset. + * Polymorphic Discriminator */ - parameters?: { [propertyName: string]: ParameterSpecification }; + type: "ParquetFormat"; /** - * List of tags that can be used for describing the Dataset. + * Serializer. Type: string (or Expression with resultType string). */ - annotations?: any[]; + serializer?: any; /** - * The folder that this Dataset is in. If not specified, Dataset will appear at the root level. + * Deserializer. Type: string (or Expression with resultType string). */ - folder?: DatasetFolder; + deserializer?: any; +} + +/** + * The data stored in Optimized Row Columnar (ORC) format. + */ +export interface OrcFormat { /** - * This property will be retired. Please consider using schema + table properties instead. + * Polymorphic Discriminator */ - tableName?: any; + type: "OrcFormat"; /** - * The table name of Greenplum. Type: string (or Expression with resultType string). + * Serializer. Type: string (or Expression with resultType string). */ - table?: any; + serializer?: any; /** - * The schema name of Greenplum. Type: string (or Expression with resultType string). + * Deserializer. Type: string (or Expression with resultType string). */ - greenplumTableDatasetSchema?: any; + deserializer?: any; } /** - * Google BigQuery service dataset. + * The data stored in Avro format. */ -export interface GoogleBigQueryObjectDataset { +export interface AvroFormat { /** * Polymorphic Discriminator */ - type: "GoogleBigQueryObject"; + type: "AvroFormat"; /** - * Dataset description. + * Serializer. Type: string (or Expression with resultType string). */ - description?: string; + serializer?: any; /** - * Columns that define the structure of the dataset. Type: array (or Expression with resultType - * array), itemType: DatasetDataElement. + * Deserializer. Type: string (or Expression with resultType string). */ - structure?: any; + deserializer?: any; +} + +/** + * The data stored in JSON format. + */ +export interface JsonFormat { /** - * Columns that define the physical type schema of the dataset. Type: array (or Expression with - * resultType array), itemType: DatasetSchemaDataElement. + * Polymorphic Discriminator */ - schema?: any; + type: "JsonFormat"; /** - * Linked service reference. + * Serializer. Type: string (or Expression with resultType string). */ - linkedServiceName: LinkedServiceReference; + serializer?: any; /** - * Parameters for dataset. + * Deserializer. Type: string (or Expression with resultType string). */ - parameters?: { [propertyName: string]: ParameterSpecification }; + deserializer?: any; /** - * List of tags that can be used for describing the Dataset. + * File pattern of JSON. To be more specific, the way of separating a collection of JSON objects. + * The default value is 'setOfObjects'. It is case-sensitive. */ - annotations?: any[]; + filePattern?: any; /** - * The folder that this Dataset is in. If not specified, Dataset will appear at the root level. + * The character used to separate nesting levels. Default value is '.' (dot). Type: string (or + * Expression with resultType string). */ - folder?: DatasetFolder; + nestingSeparator?: any; /** - * This property will be retired. Please consider using database + table properties instead. + * The code page name of the preferred encoding. If not provided, the default value is 'utf-8', + * unless the byte order mark (BOM) denotes another Unicode encoding. The full list of supported + * values can be found in the 'Name' column of the table of encodings in the following reference: + * https://go.microsoft.com/fwlink/?linkid=861078. Type: string (or Expression with resultType + * string). */ - tableName?: any; + encodingName?: any; /** - * The table name of the Google BigQuery. Type: string (or Expression with resultType string). + * The JSONPath of the JSON array element to be flattened. Example: "$.ArrayPath". Type: string + * (or Expression with resultType string). */ - table?: any; + jsonNodeReference?: any; /** - * The database name of the Google BigQuery. Type: string (or Expression with resultType string). + * The JSONPath definition for each column mapping with a customized column name to extract data + * from JSON file. For fields under root object, start with "$"; for fields inside the array + * chosen by jsonNodeReference property, start from the array element. Example: {"Column1": + * "$.Column1Path", "Column2": "Column2PathInArray"}. Type: object (or Expression with resultType + * object). */ - dataset?: any; + jsonPathDefinition?: any; } /** - * Eloqua server dataset. + * The data stored in text format. */ -export interface EloquaObjectDataset { +export interface TextFormat { /** * Polymorphic Discriminator */ - type: "EloquaObject"; + type: "TextFormat"; /** - * Dataset description. + * Serializer. Type: string (or Expression with resultType string). */ - description?: string; + serializer?: any; /** - * Columns that define the structure of the dataset. Type: array (or Expression with resultType - * array), itemType: DatasetDataElement. + * Deserializer. Type: string (or Expression with resultType string). */ - structure?: any; + deserializer?: any; /** - * Columns that define the physical type schema of the dataset. Type: array (or Expression with - * resultType array), itemType: DatasetSchemaDataElement. + * The column delimiter. Type: string (or Expression with resultType string). */ - schema?: any; + columnDelimiter?: any; /** - * Linked service reference. + * The row delimiter. Type: string (or Expression with resultType string). */ - linkedServiceName: LinkedServiceReference; + rowDelimiter?: any; /** - * Parameters for dataset. + * The escape character. Type: string (or Expression with resultType string). */ - parameters?: { [propertyName: string]: ParameterSpecification }; + escapeChar?: any; /** - * List of tags that can be used for describing the Dataset. + * The quote character. Type: string (or Expression with resultType string). */ - annotations?: any[]; + quoteChar?: any; /** - * The folder that this Dataset is in. If not specified, Dataset will appear at the root level. + * The null value string. Type: string (or Expression with resultType string). */ - folder?: DatasetFolder; + nullValue?: any; /** - * The table name. Type: string (or Expression with resultType string). + * The code page name of the preferred encoding. If miss, the default value is ΓÇ£utf-8ΓÇ¥, + * unless BOM denotes another Unicode encoding. Refer to the ΓÇ£NameΓÇ¥ column of the table in + * the following link to set supported values: + * https://msdn.microsoft.com/library/system.text.encoding.aspx. Type: string (or Expression with + * resultType string). */ - tableName?: any; + encodingName?: any; + /** + * Treat empty column values in the text file as null. The default value is true. Type: boolean + * (or Expression with resultType boolean). + */ + treatEmptyAsNull?: any; + /** + * The number of lines/rows to be skipped when parsing text files. The default value is 0. Type: + * integer (or Expression with resultType integer). + */ + skipLineCount?: any; + /** + * When used as input, treat the first row of data as headers. When used as output,write the + * headers into the output as the first row of data. The default value is false. Type: boolean + * (or Expression with resultType boolean). + */ + firstRowAsHeader?: any; } /** - * Drill server dataset. + * A file in an HTTP web server. */ -export interface DrillTableDataset { +export interface HttpDataset { /** * Polymorphic Discriminator */ - type: "DrillTable"; + type: "HttpFile"; /** * Dataset description. */ @@ -9562,27 +9679,43 @@ export interface DrillTableDataset { */ folder?: DatasetFolder; /** - * This property will be retired. Please consider using schema + table properties instead. + * The relative URL based on the URL in the HttpLinkedService refers to an HTTP file Type: string + * (or Expression with resultType string). */ - tableName?: any; + relativeUrl?: any; /** - * The table name of the Drill. Type: string (or Expression with resultType string). + * The HTTP method for the HTTP request. Type: string (or Expression with resultType string). */ - table?: any; + requestMethod?: any; /** - * The schema name of the Drill. Type: string (or Expression with resultType string). + * The body for the HTTP request. Type: string (or Expression with resultType string). */ - drillTableDatasetSchema?: any; + requestBody?: any; + /** + * The headers for the HTTP Request. e.g. request-header-name-1:request-header-value-1 + * ... + * request-header-name-n:request-header-value-n Type: string (or Expression with resultType + * string). + */ + additionalHeaders?: any; + /** + * The format of files. + */ + format?: DatasetStorageFormatUnion; + /** + * The data compression method used on files. + */ + compression?: DatasetCompressionUnion; } /** - * Couchbase server dataset. + * The Azure Search Index. */ -export interface CouchbaseTableDataset { +export interface AzureSearchIndexDataset { /** * Polymorphic Discriminator */ - type: "CouchbaseTable"; + type: "AzureSearchIndex"; /** * Dataset description. */ @@ -9614,19 +9747,19 @@ export interface CouchbaseTableDataset { */ folder?: DatasetFolder; /** - * The table name. Type: string (or Expression with resultType string). + * The name of the Azure Search Index. Type: string (or Expression with resultType string). */ - tableName?: any; + indexName: any; } /** - * Concur Service dataset. + * The dataset points to a HTML table in the web page. */ -export interface ConcurObjectDataset { +export interface WebTableDataset { /** * Polymorphic Discriminator */ - type: "ConcurObject"; + type: "WebTable"; /** * Dataset description. */ @@ -9658,19 +9791,25 @@ export interface ConcurObjectDataset { */ folder?: DatasetFolder; /** - * The table name. Type: string (or Expression with resultType string). + * The zero-based index of the table in the web page. Type: integer (or Expression with + * resultType integer), minimum: 0. */ - tableName?: any; + index: any; + /** + * The relative URL to the web page from the linked service URL. Type: string (or Expression with + * resultType string). + */ + path?: any; } /** - * Azure PostgreSQL dataset. + * SAP Table Resource properties. */ -export interface AzurePostgreSqlTableDataset { +export interface SapTableResourceDataset { /** * Polymorphic Discriminator */ - type: "AzurePostgreSqlTable"; + type: "SapTableResource"; /** * Dataset description. */ @@ -9694,38 +9833,27 @@ export interface AzurePostgreSqlTableDataset { */ parameters?: { [propertyName: string]: ParameterSpecification }; /** - * List of tags that can be used for describing the Dataset. - */ - annotations?: any[]; - /** - * The folder that this Dataset is in. If not specified, Dataset will appear at the root level. - */ - folder?: DatasetFolder; - /** - * The table name of the Azure PostgreSQL database which includes both schema and table. Type: - * string (or Expression with resultType string). + * List of tags that can be used for describing the Dataset. */ - tableName?: any; + annotations?: any[]; /** - * The table name of the Azure PostgreSQL database. Type: string (or Expression with resultType - * string). + * The folder that this Dataset is in. If not specified, Dataset will appear at the root level. */ - table?: any; + folder?: DatasetFolder; /** - * The schema name of the Azure PostgreSQL database. Type: string (or Expression with resultType - * string). + * The name of the SAP Table. Type: string (or Expression with resultType string). */ - azurePostgreSqlTableDatasetSchema?: any; + tableName: any; } /** - * Amazon Marketplace Web Service dataset. + * A Rest service dataset. */ -export interface AmazonMWSObjectDataset { +export interface RestResourceDataset { /** * Polymorphic Discriminator */ - type: "AmazonMWSObject"; + type: "RestResource"; /** * Dataset description. */ @@ -9757,282 +9885,197 @@ export interface AmazonMWSObjectDataset { */ folder?: DatasetFolder; /** - * The table name. Type: string (or Expression with resultType string). + * The relative URL to the resource that the RESTful API provides. Type: string (or Expression + * with resultType string). */ - tableName?: any; -} - -/** - * Contains the possible cases for DatasetCompression. - */ -export type DatasetCompressionUnion = DatasetCompression | DatasetZipDeflateCompression | DatasetDeflateCompression | DatasetGZipCompression | DatasetBZip2Compression; - -/** - * The compression method used on a dataset. - */ -export interface DatasetCompression { + relativeUrl?: any; /** - * Polymorphic Discriminator + * The HTTP method used to call the RESTful API. The default is GET. Type: string (or Expression + * with resultType string). */ - type: "DatasetCompression"; + requestMethod?: any; /** - * Describes unknown properties. The value of an unknown property can be of "any" type. + * The HTTP request body to the RESTful API if requestMethod is POST. Type: string (or Expression + * with resultType string). */ - [property: string]: any; -} - -/** - * The ZipDeflate compression method used on a dataset. - */ -export interface DatasetZipDeflateCompression { + requestBody?: any; /** - * Polymorphic Discriminator + * The additional HTTP headers in the request to the RESTful API. Type: string (or Expression + * with resultType string). */ - type: "ZipDeflate"; + additionalHeaders?: any; /** - * The ZipDeflate compression level. + * The pagination rules to compose next page requests. Type: string (or Expression with + * resultType string). */ - level?: any; + paginationRules?: any; } /** - * The Deflate compression method used on a dataset. + * The on-premises SQL Server dataset. */ -export interface DatasetDeflateCompression { +export interface SqlServerTableDataset { /** * Polymorphic Discriminator */ - type: "Deflate"; - /** - * The Deflate compression level. - */ - level?: any; -} - -/** - * The GZip compression method used on a dataset. - */ -export interface DatasetGZipCompression { + type: "SqlServerTable"; /** - * Polymorphic Discriminator + * Dataset description. */ - type: "GZip"; + description?: string; /** - * The GZip compression level. + * Columns that define the structure of the dataset. Type: array (or Expression with resultType + * array), itemType: DatasetDataElement. */ - level?: any; -} - -/** - * The BZip2 compression method used on a dataset. - */ -export interface DatasetBZip2Compression { + structure?: any; /** - * Polymorphic Discriminator + * Columns that define the physical type schema of the dataset. Type: array (or Expression with + * resultType array), itemType: DatasetSchemaDataElement. */ - type: "BZip2"; -} - -/** - * Contains the possible cases for DatasetStorageFormat. - */ -export type DatasetStorageFormatUnion = DatasetStorageFormat | ParquetFormat | OrcFormat | AvroFormat | JsonFormat | TextFormat; - -/** - * The format definition of a storage. - */ -export interface DatasetStorageFormat { + schema?: any; /** - * Polymorphic Discriminator + * Linked service reference. */ - type: "DatasetStorageFormat"; + linkedServiceName: LinkedServiceReference; /** - * Serializer. Type: string (or Expression with resultType string). + * Parameters for dataset. */ - serializer?: any; + parameters?: { [propertyName: string]: ParameterSpecification }; /** - * Deserializer. Type: string (or Expression with resultType string). + * List of tags that can be used for describing the Dataset. */ - deserializer?: any; + annotations?: any[]; /** - * Describes unknown properties. The value of an unknown property can be of "any" type. + * The folder that this Dataset is in. If not specified, Dataset will appear at the root level. */ - [property: string]: any; -} - -/** - * The data stored in Parquet format. - */ -export interface ParquetFormat { + folder?: DatasetFolder; /** - * Polymorphic Discriminator + * This property will be retired. Please consider using schema + table properties instead. */ - type: "ParquetFormat"; + tableName?: any; /** - * Serializer. Type: string (or Expression with resultType string). + * The schema name of the SQL Server dataset. Type: string (or Expression with resultType + * string). */ - serializer?: any; + sqlServerTableDatasetSchema?: any; /** - * Deserializer. Type: string (or Expression with resultType string). + * The table name of the SQL Server dataset. Type: string (or Expression with resultType string). */ - deserializer?: any; + table?: any; } /** - * The data stored in Optimized Row Columnar (ORC) format. + * Sap Business Warehouse Open Hub Destination Table properties. */ -export interface OrcFormat { +export interface SapOpenHubTableDataset { /** * Polymorphic Discriminator */ - type: "OrcFormat"; - /** - * Serializer. Type: string (or Expression with resultType string). - */ - serializer?: any; - /** - * Deserializer. Type: string (or Expression with resultType string). - */ - deserializer?: any; -} - -/** - * The data stored in Avro format. - */ -export interface AvroFormat { + type: "SapOpenHubTable"; /** - * Polymorphic Discriminator + * Dataset description. */ - type: "AvroFormat"; + description?: string; /** - * Serializer. Type: string (or Expression with resultType string). + * Columns that define the structure of the dataset. Type: array (or Expression with resultType + * array), itemType: DatasetDataElement. */ - serializer?: any; + structure?: any; /** - * Deserializer. Type: string (or Expression with resultType string). + * Columns that define the physical type schema of the dataset. Type: array (or Expression with + * resultType array), itemType: DatasetSchemaDataElement. */ - deserializer?: any; -} - -/** - * The data stored in JSON format. - */ -export interface JsonFormat { + schema?: any; /** - * Polymorphic Discriminator + * Linked service reference. */ - type: "JsonFormat"; + linkedServiceName: LinkedServiceReference; /** - * Serializer. Type: string (or Expression with resultType string). + * Parameters for dataset. */ - serializer?: any; + parameters?: { [propertyName: string]: ParameterSpecification }; /** - * Deserializer. Type: string (or Expression with resultType string). + * List of tags that can be used for describing the Dataset. */ - deserializer?: any; + annotations?: any[]; /** - * File pattern of JSON. To be more specific, the way of separating a collection of JSON objects. - * The default value is 'setOfObjects'. It is case-sensitive. + * The folder that this Dataset is in. If not specified, Dataset will appear at the root level. */ - filePattern?: any; + folder?: DatasetFolder; /** - * The character used to separate nesting levels. Default value is '.' (dot). Type: string (or + * The name of the Open Hub Destination with destination type as Database Table. Type: string (or * Expression with resultType string). */ - nestingSeparator?: any; - /** - * The code page name of the preferred encoding. If not provided, the default value is 'utf-8', - * unless the byte order mark (BOM) denotes another Unicode encoding. The full list of supported - * values can be found in the 'Name' column of the table of encodings in the following reference: - * https://go.microsoft.com/fwlink/?linkid=861078. Type: string (or Expression with resultType - * string). - */ - encodingName?: any; + openHubDestinationName: any; /** - * The JSONPath of the JSON array element to be flattened. Example: "$.ArrayPath". Type: string - * (or Expression with resultType string). + * Whether to exclude the records of the last request. The default value is true. Type: boolean + * (or Expression with resultType boolean). */ - jsonNodeReference?: any; + excludeLastRequest?: any; /** - * The JSONPath definition for each column mapping with a customized column name to extract data - * from JSON file. For fields under root object, start with "$"; for fields inside the array - * chosen by jsonNodeReference property, start from the array element. Example: {"Column1": - * "$.Column1Path", "Column2": "Column2PathInArray"}. Type: object (or Expression with resultType - * object). + * The ID of request for delta loading. Once it is set, only data with requestId larger than the + * value of this property will be retrieved. The default value is 0. Type: integer (or Expression + * with resultType integer ). */ - jsonPathDefinition?: any; + baseRequestId?: any; } /** - * The data stored in text format. + * SAP HANA Table properties. */ -export interface TextFormat { +export interface SapHanaTableDataset { /** * Polymorphic Discriminator */ - type: "TextFormat"; - /** - * Serializer. Type: string (or Expression with resultType string). - */ - serializer?: any; - /** - * Deserializer. Type: string (or Expression with resultType string). - */ - deserializer?: any; + type: "SapHanaTable"; /** - * The column delimiter. Type: string (or Expression with resultType string). + * Dataset description. */ - columnDelimiter?: any; + description?: string; /** - * The row delimiter. Type: string (or Expression with resultType string). + * Columns that define the structure of the dataset. Type: array (or Expression with resultType + * array), itemType: DatasetDataElement. */ - rowDelimiter?: any; + structure?: any; /** - * The escape character. Type: string (or Expression with resultType string). + * Columns that define the physical type schema of the dataset. Type: array (or Expression with + * resultType array), itemType: DatasetSchemaDataElement. */ - escapeChar?: any; + schema?: any; /** - * The quote character. Type: string (or Expression with resultType string). + * Linked service reference. */ - quoteChar?: any; + linkedServiceName: LinkedServiceReference; /** - * The null value string. Type: string (or Expression with resultType string). + * Parameters for dataset. */ - nullValue?: any; + parameters?: { [propertyName: string]: ParameterSpecification }; /** - * The code page name of the preferred encoding. If miss, the default value is ΓÇ£utf-8ΓÇ¥, - * unless BOM denotes another Unicode encoding. Refer to the ΓÇ£NameΓÇ¥ column of the table in - * the following link to set supported values: - * https://msdn.microsoft.com/library/system.text.encoding.aspx. Type: string (or Expression with - * resultType string). + * List of tags that can be used for describing the Dataset. */ - encodingName?: any; + annotations?: any[]; /** - * Treat empty column values in the text file as null. The default value is true. Type: boolean - * (or Expression with resultType boolean). + * The folder that this Dataset is in. If not specified, Dataset will appear at the root level. */ - treatEmptyAsNull?: any; + folder?: DatasetFolder; /** - * The number of lines/rows to be skipped when parsing text files. The default value is 0. Type: - * integer (or Expression with resultType integer). + * The schema name of SAP HANA. Type: string (or Expression with resultType string). */ - skipLineCount?: any; + sapHanaTableDatasetSchema?: any; /** - * When used as input, treat the first row of data as headers. When used as output,write the - * headers into the output as the first row of data. The default value is false. Type: boolean - * (or Expression with resultType boolean). + * The table name of SAP HANA. Type: string (or Expression with resultType string). */ - firstRowAsHeader?: any; + table?: any; } /** - * A file in an HTTP web server. + * The path of the SAP ECC OData entity. */ -export interface HttpDataset { +export interface SapEccResourceDataset { /** * Polymorphic Discriminator */ - type: "HttpFile"; + type: "SapEccResource"; /** * Dataset description. */ @@ -10064,43 +10107,64 @@ export interface HttpDataset { */ folder?: DatasetFolder; /** - * The relative URL based on the URL in the HttpLinkedService refers to an HTTP file Type: string - * (or Expression with resultType string). + * The path of the SAP ECC OData entity. Type: string (or Expression with resultType string). */ - relativeUrl?: any; + path: any; +} + +/** + * The path of the SAP Cloud for Customer OData entity. + */ +export interface SapCloudForCustomerResourceDataset { /** - * The HTTP method for the HTTP request. Type: string (or Expression with resultType string). + * Polymorphic Discriminator */ - requestMethod?: any; + type: "SapCloudForCustomerResource"; /** - * The body for the HTTP request. Type: string (or Expression with resultType string). + * Dataset description. */ - requestBody?: any; + description?: string; /** - * The headers for the HTTP Request. e.g. request-header-name-1:request-header-value-1 - * ... - * request-header-name-n:request-header-value-n Type: string (or Expression with resultType - * string). + * Columns that define the structure of the dataset. Type: array (or Expression with resultType + * array), itemType: DatasetDataElement. */ - additionalHeaders?: any; + structure?: any; /** - * The format of files. + * Columns that define the physical type schema of the dataset. Type: array (or Expression with + * resultType array), itemType: DatasetSchemaDataElement. */ - format?: DatasetStorageFormatUnion; + schema?: any; /** - * The data compression method used on files. + * Linked service reference. */ - compression?: DatasetCompressionUnion; + linkedServiceName: LinkedServiceReference; + /** + * Parameters for dataset. + */ + parameters?: { [propertyName: string]: ParameterSpecification }; + /** + * List of tags that can be used for describing the Dataset. + */ + annotations?: any[]; + /** + * The folder that this Dataset is in. If not specified, Dataset will appear at the root level. + */ + folder?: DatasetFolder; + /** + * The path of the SAP Cloud for Customer OData entity. Type: string (or Expression with + * resultType string). + */ + path: any; } /** - * The Azure Search Index. + * The SAP BW cube dataset. */ -export interface AzureSearchIndexDataset { +export interface SapBwCubeDataset { /** * Polymorphic Discriminator */ - type: "AzureSearchIndex"; + type: "SapBwCube"; /** * Dataset description. */ @@ -10131,20 +10195,16 @@ export interface AzureSearchIndexDataset { * The folder that this Dataset is in. If not specified, Dataset will appear at the root level. */ folder?: DatasetFolder; - /** - * The name of the Azure Search Index. Type: string (or Expression with resultType string). - */ - indexName: any; } /** - * The dataset points to a HTML table in the web page. + * The Sybase table dataset. */ -export interface WebTableDataset { +export interface SybaseTableDataset { /** * Polymorphic Discriminator */ - type: "WebTable"; + type: "SybaseTable"; /** * Dataset description. */ @@ -10176,25 +10236,19 @@ export interface WebTableDataset { */ folder?: DatasetFolder; /** - * The zero-based index of the table in the web page. Type: integer (or Expression with - * resultType integer), minimum: 0. - */ - index: any; - /** - * The relative URL to the web page from the linked service URL. Type: string (or Expression with - * resultType string). + * The Sybase table name. Type: string (or Expression with resultType string). */ - path?: any; + tableName?: any; } /** - * SAP Table Resource properties. + * The Salesforce Service Cloud object dataset. */ -export interface SapTableResourceDataset { +export interface SalesforceServiceCloudObjectDataset { /** * Polymorphic Discriminator */ - type: "SapTableResource"; + type: "SalesforceServiceCloudObject"; /** * Dataset description. */ @@ -10226,19 +10280,20 @@ export interface SapTableResourceDataset { */ folder?: DatasetFolder; /** - * The name of the SAP Table. Type: string (or Expression with resultType string). + * The Salesforce Service Cloud object API name. Type: string (or Expression with resultType + * string). */ - tableName: any; + objectApiName?: any; } /** - * A Rest service dataset. + * The Salesforce object dataset. */ -export interface RestResourceDataset { +export interface SalesforceObjectDataset { /** * Polymorphic Discriminator */ - type: "RestResource"; + type: "SalesforceObject"; /** * Dataset description. */ @@ -10270,40 +10325,19 @@ export interface RestResourceDataset { */ folder?: DatasetFolder; /** - * The relative URL to the resource that the RESTful API provides. Type: string (or Expression - * with resultType string). - */ - relativeUrl?: any; - /** - * The HTTP method used to call the RESTful API. The default is GET. Type: string (or Expression - * with resultType string). - */ - requestMethod?: any; - /** - * The HTTP request body to the RESTful API if requestMethod is POST. Type: string (or Expression - * with resultType string). - */ - requestBody?: any; - /** - * The additional HTTP headers in the request to the RESTful API. Type: string (or Expression - * with resultType string). - */ - additionalHeaders?: any; - /** - * The pagination rules to compose next page requests. Type: string (or Expression with - * resultType string). + * The Salesforce object API name. Type: string (or Expression with resultType string). */ - paginationRules?: any; + objectApiName?: any; } /** - * The on-premises SQL Server dataset. + * The Microsoft Access table dataset. */ -export interface SqlServerTableDataset { +export interface MicrosoftAccessTableDataset { /** * Polymorphic Discriminator */ - type: "SqlServerTable"; + type: "MicrosoftAccessTable"; /** * Dataset description. */ @@ -10335,28 +10369,19 @@ export interface SqlServerTableDataset { */ folder?: DatasetFolder; /** - * This property will be retired. Please consider using schema + table properties instead. + * The Microsoft Access table name. Type: string (or Expression with resultType string). */ tableName?: any; - /** - * The schema name of the SQL Server dataset. Type: string (or Expression with resultType - * string). - */ - sqlServerTableDatasetSchema?: any; - /** - * The table name of the SQL Server dataset. Type: string (or Expression with resultType string). - */ - table?: any; } /** - * Sap Business Warehouse Open Hub Destination Table properties. + * The PostgreSQL table dataset. */ -export interface SapOpenHubTableDataset { +export interface PostgreSqlTableDataset { /** * Polymorphic Discriminator */ - type: "SapOpenHubTable"; + type: "PostgreSqlTable"; /** * Dataset description. */ @@ -10388,31 +10413,27 @@ export interface SapOpenHubTableDataset { */ folder?: DatasetFolder; /** - * The name of the Open Hub Destination with destination type as Database Table. Type: string (or - * Expression with resultType string). + * This property will be retired. Please consider using schema + table properties instead. */ - openHubDestinationName: any; + tableName?: any; /** - * Whether to exclude the records of the last request. The default value is true. Type: boolean - * (or Expression with resultType boolean). + * The PostgreSQL table name. Type: string (or Expression with resultType string). */ - excludeLastRequest?: any; + table?: any; /** - * The ID of request for delta loading. Once it is set, only data with requestId larger than the - * value of this property will be retrieved. The default value is 0. Type: integer (or Expression - * with resultType integer ). + * The PostgreSQL schema name. Type: string (or Expression with resultType string). */ - baseRequestId?: any; + postgreSqlTableDatasetSchema?: any; } /** - * SAP HANA Table properties. + * The MySQL table dataset. */ -export interface SapHanaTableDataset { +export interface MySqlTableDataset { /** * Polymorphic Discriminator */ - type: "SapHanaTable"; + type: "MySqlTable"; /** * Dataset description. */ @@ -10444,23 +10465,19 @@ export interface SapHanaTableDataset { */ folder?: DatasetFolder; /** - * The schema name of SAP HANA. Type: string (or Expression with resultType string). - */ - sapHanaTableDatasetSchema?: any; - /** - * The table name of SAP HANA. Type: string (or Expression with resultType string). + * The MySQL table name. Type: string (or Expression with resultType string). */ - table?: any; + tableName?: any; } /** - * The path of the SAP ECC OData entity. + * The ODBC table dataset. */ -export interface SapEccResourceDataset { +export interface OdbcTableDataset { /** * Polymorphic Discriminator */ - type: "SapEccResource"; + type: "OdbcTable"; /** * Dataset description. */ @@ -10492,19 +10509,19 @@ export interface SapEccResourceDataset { */ folder?: DatasetFolder; /** - * The path of the SAP ECC OData entity. Type: string (or Expression with resultType string). + * The ODBC table name. Type: string (or Expression with resultType string). */ - path: any; + tableName?: any; } /** - * The path of the SAP Cloud for Customer OData entity. + * The Informix table dataset. */ -export interface SapCloudForCustomerResourceDataset { +export interface InformixTableDataset { /** * Polymorphic Discriminator */ - type: "SapCloudForCustomerResource"; + type: "InformixTable"; /** * Dataset description. */ @@ -10536,20 +10553,19 @@ export interface SapCloudForCustomerResourceDataset { */ folder?: DatasetFolder; /** - * The path of the SAP Cloud for Customer OData entity. Type: string (or Expression with - * resultType string). + * The Informix table name. Type: string (or Expression with resultType string). */ - path: any; + tableName?: any; } /** - * The SAP BW cube dataset. + * The relational table dataset. */ -export interface SapBwCubeDataset { +export interface RelationalTableDataset { /** * Polymorphic Discriminator */ - type: "SapBwCube"; + type: "RelationalTable"; /** * Dataset description. */ @@ -10580,16 +10596,20 @@ export interface SapBwCubeDataset { * The folder that this Dataset is in. If not specified, Dataset will appear at the root level. */ folder?: DatasetFolder; + /** + * The relational table name. Type: string (or Expression with resultType string). + */ + tableName?: any; } /** - * The Sybase table dataset. + * The Db2 table dataset. */ -export interface SybaseTableDataset { +export interface Db2TableDataset { /** * Polymorphic Discriminator */ - type: "SybaseTable"; + type: "Db2Table"; /** * Dataset description. */ @@ -10617,23 +10637,31 @@ export interface SybaseTableDataset { */ annotations?: any[]; /** - * The folder that this Dataset is in. If not specified, Dataset will appear at the root level. + * The folder that this Dataset is in. If not specified, Dataset will appear at the root level. + */ + folder?: DatasetFolder; + /** + * This property will be retired. Please consider using schema + table properties instead. + */ + tableName?: any; + /** + * The Db2 schema name. Type: string (or Expression with resultType string). */ - folder?: DatasetFolder; + db2TableDatasetSchema?: any; /** - * The Sybase table name. Type: string (or Expression with resultType string). + * The Db2 table name. Type: string (or Expression with resultType string). */ - tableName?: any; + table?: any; } /** - * The Salesforce Service Cloud object dataset. + * The Amazon Redshift table dataset. */ -export interface SalesforceServiceCloudObjectDataset { +export interface AmazonRedshiftTableDataset { /** * Polymorphic Discriminator */ - type: "SalesforceServiceCloudObject"; + type: "AmazonRedshiftTable"; /** * Dataset description. */ @@ -10665,20 +10693,27 @@ export interface SalesforceServiceCloudObjectDataset { */ folder?: DatasetFolder; /** - * The Salesforce Service Cloud object API name. Type: string (or Expression with resultType - * string). + * This property will be retired. Please consider using schema + table properties instead. */ - objectApiName?: any; + tableName?: any; + /** + * The Amazon Redshift table name. Type: string (or Expression with resultType string). + */ + table?: any; + /** + * The Amazon Redshift schema name. Type: string (or Expression with resultType string). + */ + amazonRedshiftTableDatasetSchema?: any; } /** - * The Salesforce object dataset. + * The Azure MySQL database dataset. */ -export interface SalesforceObjectDataset { +export interface AzureMySqlTableDataset { /** * Polymorphic Discriminator */ - type: "SalesforceObject"; + type: "AzureMySqlTable"; /** * Dataset description. */ @@ -10710,19 +10745,23 @@ export interface SalesforceObjectDataset { */ folder?: DatasetFolder; /** - * The Salesforce object API name. Type: string (or Expression with resultType string). + * The Azure MySQL database table name. Type: string (or Expression with resultType string). */ - objectApiName?: any; + tableName?: any; + /** + * The name of Azure MySQL database table. Type: string (or Expression with resultType string). + */ + table?: any; } /** - * The Microsoft Access table dataset. + * The Teradata database dataset. */ -export interface MicrosoftAccessTableDataset { +export interface TeradataTableDataset { /** * Polymorphic Discriminator */ - type: "MicrosoftAccessTable"; + type: "TeradataTable"; /** * Dataset description. */ @@ -10754,19 +10793,23 @@ export interface MicrosoftAccessTableDataset { */ folder?: DatasetFolder; /** - * The Microsoft Access table name. Type: string (or Expression with resultType string). + * The database name of Teradata. Type: string (or Expression with resultType string). */ - tableName?: any; + database?: any; + /** + * The table name of Teradata. Type: string (or Expression with resultType string). + */ + table?: any; } /** - * The PostgreSQL table dataset. + * The on-premises Oracle database dataset. */ -export interface PostgreSqlTableDataset { +export interface OracleTableDataset { /** * Polymorphic Discriminator */ - type: "PostgreSqlTable"; + type: "OracleTable"; /** * Dataset description. */ @@ -10802,23 +10845,25 @@ export interface PostgreSqlTableDataset { */ tableName?: any; /** - * The PostgreSQL table name. Type: string (or Expression with resultType string). + * The schema name of the on-premises Oracle database. Type: string (or Expression with + * resultType string). */ - table?: any; + oracleTableDatasetSchema?: any; /** - * The PostgreSQL schema name. Type: string (or Expression with resultType string). + * The table name of the on-premises Oracle database. Type: string (or Expression with resultType + * string). */ - postgreSqlTableDatasetSchema?: any; + table?: any; } /** - * The MySQL table dataset. + * The Open Data Protocol (OData) resource dataset. */ -export interface MySqlTableDataset { +export interface ODataResourceDataset { /** * Polymorphic Discriminator */ - type: "MySqlTable"; + type: "ODataResource"; /** * Dataset description. */ @@ -10850,19 +10895,19 @@ export interface MySqlTableDataset { */ folder?: DatasetFolder; /** - * The MySQL table name. Type: string (or Expression with resultType string). + * The OData resource path. Type: string (or Expression with resultType string). */ - tableName?: any; + path?: any; } /** - * The ODBC table dataset. + * The CosmosDB (MongoDB API) database dataset. */ -export interface OdbcTableDataset { +export interface CosmosDbMongoDbApiCollectionDataset { /** * Polymorphic Discriminator */ - type: "OdbcTable"; + type: "CosmosDbMongoDbApiCollection"; /** * Dataset description. */ @@ -10894,19 +10939,20 @@ export interface OdbcTableDataset { */ folder?: DatasetFolder; /** - * The ODBC table name. Type: string (or Expression with resultType string). + * The collection name of the CosmosDB (MongoDB API) database. Type: string (or Expression with + * resultType string). */ - tableName?: any; + collection: any; } /** - * The Informix table dataset. + * The MongoDB database dataset. */ -export interface InformixTableDataset { +export interface MongoDbV2CollectionDataset { /** * Polymorphic Discriminator */ - type: "InformixTable"; + type: "MongoDbV2Collection"; /** * Dataset description. */ @@ -10938,19 +10984,20 @@ export interface InformixTableDataset { */ folder?: DatasetFolder; /** - * The Informix table name. Type: string (or Expression with resultType string). + * The collection name of the MongoDB database. Type: string (or Expression with resultType + * string). */ - tableName?: any; + collection: any; } /** - * The relational table dataset. + * The MongoDB database dataset. */ -export interface RelationalTableDataset { +export interface MongoDbCollectionDataset { /** * Polymorphic Discriminator */ - type: "RelationalTable"; + type: "MongoDbCollection"; /** * Dataset description. */ @@ -10982,19 +11029,19 @@ export interface RelationalTableDataset { */ folder?: DatasetFolder; /** - * The relational table name. Type: string (or Expression with resultType string). + * The table name of the MongoDB database. Type: string (or Expression with resultType string). */ - tableName?: any; + collectionName: any; } /** - * The Db2 table dataset. + * An on-premises file system dataset. */ -export interface Db2TableDataset { +export interface FileShareDataset { /** * Polymorphic Discriminator */ - type: "Db2Table"; + type: "FileShare"; /** * Dataset description. */ @@ -11026,27 +11073,44 @@ export interface Db2TableDataset { */ folder?: DatasetFolder; /** - * This property will be retired. Please consider using schema + table properties instead. + * The path of the on-premises file system. Type: string (or Expression with resultType string). */ - tableName?: any; + folderPath?: any; /** - * The Db2 schema name. Type: string (or Expression with resultType string). + * The name of the on-premises file system. Type: string (or Expression with resultType string). */ - db2TableDatasetSchema?: any; + fileName?: any; /** - * The Db2 table name. Type: string (or Expression with resultType string). + * The start of file's modified datetime. Type: string (or Expression with resultType string). */ - table?: any; + modifiedDatetimeStart?: any; + /** + * The end of file's modified datetime. Type: string (or Expression with resultType string). + */ + modifiedDatetimeEnd?: any; + /** + * The format of the files. + */ + format?: DatasetStorageFormatUnion; + /** + * Specify a filter to be used to select a subset of files in the folderPath rather than all + * files. Type: string (or Expression with resultType string). + */ + fileFilter?: any; + /** + * The data compression method used for the file system. + */ + compression?: DatasetCompressionUnion; } /** - * The Amazon Redshift table dataset. + * The Office365 account. */ -export interface AmazonRedshiftTableDataset { +export interface Office365Dataset { /** * Polymorphic Discriminator */ - type: "AmazonRedshiftTable"; + type: "Office365Table"; /** * Dataset description. */ @@ -11078,27 +11142,25 @@ export interface AmazonRedshiftTableDataset { */ folder?: DatasetFolder; /** - * This property will be retired. Please consider using schema + table properties instead. - */ - tableName?: any; - /** - * The Amazon Redshift table name. Type: string (or Expression with resultType string). + * Name of the dataset to extract from Office 365. Type: string (or Expression with resultType + * string). */ - table?: any; + tableName: any; /** - * The Amazon Redshift schema name. Type: string (or Expression with resultType string). + * A predicate expression that can be used to filter the specific rows to extract from Office + * 365. Type: string (or Expression with resultType string). */ - amazonRedshiftTableDatasetSchema?: any; + predicate?: any; } /** - * The Azure MySQL database dataset. + * The Azure Data Lake Storage Gen2 storage. */ -export interface AzureMySqlTableDataset { +export interface AzureBlobFSDataset { /** * Polymorphic Discriminator */ - type: "AzureMySqlTable"; + type: "AzureBlobFSFile"; /** * Dataset description. */ @@ -11130,23 +11192,33 @@ export interface AzureMySqlTableDataset { */ folder?: DatasetFolder; /** - * The Azure MySQL database table name. Type: string (or Expression with resultType string). + * The path of the Azure Data Lake Storage Gen2 storage. Type: string (or Expression with + * resultType string). */ - tableName?: any; + folderPath?: any; /** - * The name of Azure MySQL database table. Type: string (or Expression with resultType string). + * The name of the Azure Data Lake Storage Gen2. Type: string (or Expression with resultType + * string). */ - table?: any; + fileName?: any; + /** + * The format of the Azure Data Lake Storage Gen2 storage. + */ + format?: DatasetStorageFormatUnion; + /** + * The data compression method used for the blob storage. + */ + compression?: DatasetCompressionUnion; } /** - * The Teradata database dataset. + * Azure Data Lake Store dataset. */ -export interface TeradataTableDataset { +export interface AzureDataLakeStoreDataset { /** * Polymorphic Discriminator */ - type: "TeradataTable"; + type: "AzureDataLakeStoreFile"; /** * Dataset description. */ @@ -11178,23 +11250,33 @@ export interface TeradataTableDataset { */ folder?: DatasetFolder; /** - * The database name of Teradata. Type: string (or Expression with resultType string). + * Path to the folder in the Azure Data Lake Store. Type: string (or Expression with resultType + * string). */ - database?: any; + folderPath?: any; /** - * The table name of Teradata. Type: string (or Expression with resultType string). + * The name of the file in the Azure Data Lake Store. Type: string (or Expression with resultType + * string). */ - table?: any; + fileName?: any; + /** + * The format of the Data Lake Store. + */ + format?: DatasetStorageFormatUnion; + /** + * The data compression method used for the item(s) in the Azure Data Lake Store. + */ + compression?: DatasetCompressionUnion; } /** - * The on-premises Oracle database dataset. + * The Common Data Service for Apps entity dataset. */ -export interface OracleTableDataset { +export interface CommonDataServiceForAppsEntityDataset { /** * Polymorphic Discriminator */ - type: "OracleTable"; + type: "CommonDataServiceForAppsEntity"; /** * Dataset description. */ @@ -11226,29 +11308,19 @@ export interface OracleTableDataset { */ folder?: DatasetFolder; /** - * This property will be retired. Please consider using schema + table properties instead. - */ - tableName?: any; - /** - * The schema name of the on-premises Oracle database. Type: string (or Expression with - * resultType string). - */ - oracleTableDatasetSchema?: any; - /** - * The table name of the on-premises Oracle database. Type: string (or Expression with resultType - * string). + * The logical name of the entity. Type: string (or Expression with resultType string). */ - table?: any; + entityName?: any; } /** - * The Open Data Protocol (OData) resource dataset. + * The Dynamics CRM entity dataset. */ -export interface ODataResourceDataset { +export interface DynamicsCrmEntityDataset { /** * Polymorphic Discriminator */ - type: "ODataResource"; + type: "DynamicsCrmEntity"; /** * Dataset description. */ @@ -11280,19 +11352,19 @@ export interface ODataResourceDataset { */ folder?: DatasetFolder; /** - * The OData resource path. Type: string (or Expression with resultType string). + * The logical name of the entity. Type: string (or Expression with resultType string). */ - path?: any; + entityName?: any; } /** - * The CosmosDB (MongoDB API) database dataset. + * The Dynamics entity dataset. */ -export interface CosmosDbMongoDbApiCollectionDataset { +export interface DynamicsEntityDataset { /** * Polymorphic Discriminator */ - type: "CosmosDbMongoDbApiCollection"; + type: "DynamicsEntity"; /** * Dataset description. */ @@ -11323,21 +11395,20 @@ export interface CosmosDbMongoDbApiCollectionDataset { * The folder that this Dataset is in. If not specified, Dataset will appear at the root level. */ folder?: DatasetFolder; - /** - * The collection name of the CosmosDB (MongoDB API) database. Type: string (or Expression with - * resultType string). + /** + * The logical name of the entity. Type: string (or Expression with resultType string). */ - collection: any; + entityName?: any; } /** - * The MongoDB database dataset. + * Microsoft Azure Document Database Collection dataset. */ -export interface MongoDbV2CollectionDataset { +export interface DocumentDbCollectionDataset { /** * Polymorphic Discriminator */ - type: "MongoDbV2Collection"; + type: "DocumentDbCollection"; /** * Dataset description. */ @@ -11369,20 +11440,19 @@ export interface MongoDbV2CollectionDataset { */ folder?: DatasetFolder; /** - * The collection name of the MongoDB database. Type: string (or Expression with resultType - * string). + * Document Database collection name. Type: string (or Expression with resultType string). */ - collection: any; + collectionName: any; } /** - * The MongoDB database dataset. + * Microsoft Azure CosmosDB (SQL API) Collection dataset. */ -export interface MongoDbCollectionDataset { +export interface CosmosDbSqlApiCollectionDataset { /** * Polymorphic Discriminator */ - type: "MongoDbCollection"; + type: "CosmosDbSqlApiCollection"; /** * Dataset description. */ @@ -11414,19 +11484,19 @@ export interface MongoDbCollectionDataset { */ folder?: DatasetFolder; /** - * The table name of the MongoDB database. Type: string (or Expression with resultType string). + * CosmosDB (SQL API) collection name. Type: string (or Expression with resultType string). */ collectionName: any; } /** - * An on-premises file system dataset. + * The custom dataset. */ -export interface FileShareDataset { +export interface CustomDataset { /** * Polymorphic Discriminator */ - type: "FileShare"; + type: "CustomDataset"; /** * Dataset description. */ @@ -11458,44 +11528,19 @@ export interface FileShareDataset { */ folder?: DatasetFolder; /** - * The path of the on-premises file system. Type: string (or Expression with resultType string). - */ - folderPath?: any; - /** - * The name of the on-premises file system. Type: string (or Expression with resultType string). - */ - fileName?: any; - /** - * The start of file's modified datetime. Type: string (or Expression with resultType string). - */ - modifiedDatetimeStart?: any; - /** - * The end of file's modified datetime. Type: string (or Expression with resultType string). - */ - modifiedDatetimeEnd?: any; - /** - * The format of the files. - */ - format?: DatasetStorageFormatUnion; - /** - * Specify a filter to be used to select a subset of files in the folderPath rather than all - * files. Type: string (or Expression with resultType string). - */ - fileFilter?: any; - /** - * The data compression method used for the file system. + * Custom dataset properties. */ - compression?: DatasetCompressionUnion; + typeProperties?: any; } /** - * The Office365 account. + * The Cassandra database dataset. */ -export interface Office365Dataset { +export interface CassandraTableDataset { /** * Polymorphic Discriminator */ - type: "Office365Table"; + type: "CassandraTable"; /** * Dataset description. */ @@ -11527,25 +11572,23 @@ export interface Office365Dataset { */ folder?: DatasetFolder; /** - * Name of the dataset to extract from Office 365. Type: string (or Expression with resultType - * string). + * The table name of the Cassandra database. Type: string (or Expression with resultType string). */ - tableName: any; + tableName?: any; /** - * A predicate expression that can be used to filter the specific rows to extract from Office - * 365. Type: string (or Expression with resultType string). + * The keyspace of the Cassandra database. Type: string (or Expression with resultType string). */ - predicate?: any; + keyspace?: any; } /** - * The Azure Data Lake Storage Gen2 storage. + * The Azure SQL Data Warehouse dataset. */ -export interface AzureBlobFSDataset { +export interface AzureSqlDWTableDataset { /** * Polymorphic Discriminator */ - type: "AzureBlobFSFile"; + type: "AzureSqlDWTable"; /** * Dataset description. */ @@ -11577,33 +11620,29 @@ export interface AzureBlobFSDataset { */ folder?: DatasetFolder; /** - * The path of the Azure Data Lake Storage Gen2 storage. Type: string (or Expression with - * resultType string). + * This property will be retired. Please consider using schema + table properties instead. */ - folderPath?: any; + tableName?: any; /** - * The name of the Azure Data Lake Storage Gen2. Type: string (or Expression with resultType + * The schema name of the Azure SQL Data Warehouse. Type: string (or Expression with resultType * string). */ - fileName?: any; - /** - * The format of the Azure Data Lake Storage Gen2 storage. - */ - format?: DatasetStorageFormatUnion; + azureSqlDWTableDatasetSchema?: any; /** - * The data compression method used for the blob storage. + * The table name of the Azure SQL Data Warehouse. Type: string (or Expression with resultType + * string). */ - compression?: DatasetCompressionUnion; + table?: any; } /** - * Azure Data Lake Store dataset. + * The Azure SQL Managed Instance dataset. */ -export interface AzureDataLakeStoreDataset { +export interface AzureSqlMITableDataset { /** * Polymorphic Discriminator */ - type: "AzureDataLakeStoreFile"; + type: "AzureSqlMITable"; /** * Dataset description. */ @@ -11635,33 +11674,29 @@ export interface AzureDataLakeStoreDataset { */ folder?: DatasetFolder; /** - * Path to the folder in the Azure Data Lake Store. Type: string (or Expression with resultType - * string). + * This property will be retired. Please consider using schema + table properties instead. */ - folderPath?: any; + tableName?: any; /** - * The name of the file in the Azure Data Lake Store. Type: string (or Expression with resultType + * The schema name of the Azure SQL Managed Instance. Type: string (or Expression with resultType * string). */ - fileName?: any; - /** - * The format of the Data Lake Store. - */ - format?: DatasetStorageFormatUnion; + azureSqlMITableDatasetSchema?: any; /** - * The data compression method used for the item(s) in the Azure Data Lake Store. + * The table name of the Azure SQL Managed Instance dataset. Type: string (or Expression with + * resultType string). */ - compression?: DatasetCompressionUnion; + table?: any; } /** - * The Common Data Service for Apps entity dataset. + * The Azure SQL Server database dataset. */ -export interface CommonDataServiceForAppsEntityDataset { +export interface AzureSqlTableDataset { /** * Polymorphic Discriminator */ - type: "CommonDataServiceForAppsEntity"; + type: "AzureSqlTable"; /** * Dataset description. */ @@ -11693,19 +11728,28 @@ export interface CommonDataServiceForAppsEntityDataset { */ folder?: DatasetFolder; /** - * The logical name of the entity. Type: string (or Expression with resultType string). + * This property will be retired. Please consider using schema + table properties instead. */ - entityName?: any; + tableName?: any; + /** + * The schema name of the Azure SQL database. Type: string (or Expression with resultType + * string). + */ + azureSqlTableDatasetSchema?: any; + /** + * The table name of the Azure SQL database. Type: string (or Expression with resultType string). + */ + table?: any; } /** - * The Dynamics CRM entity dataset. + * The Azure Table storage dataset. */ -export interface DynamicsCrmEntityDataset { +export interface AzureTableDataset { /** * Polymorphic Discriminator */ - type: "DynamicsCrmEntity"; + type: "AzureTable"; /** * Dataset description. */ @@ -11737,19 +11781,20 @@ export interface DynamicsCrmEntityDataset { */ folder?: DatasetFolder; /** - * The logical name of the entity. Type: string (or Expression with resultType string). + * The table name of the Azure Table storage. Type: string (or Expression with resultType + * string). */ - entityName?: any; + tableName: any; } /** - * The Dynamics entity dataset. + * The Azure Blob storage. */ -export interface DynamicsEntityDataset { +export interface AzureBlobDataset { /** * Polymorphic Discriminator */ - type: "DynamicsEntity"; + type: "AzureBlob"; /** * Dataset description. */ @@ -11781,253 +11826,300 @@ export interface DynamicsEntityDataset { */ folder?: DatasetFolder; /** - * The logical name of the entity. Type: string (or Expression with resultType string). + * The path of the Azure Blob storage. Type: string (or Expression with resultType string). */ - entityName?: any; -} - -/** - * Microsoft Azure Document Database Collection dataset. - */ -export interface DocumentDbCollectionDataset { + folderPath?: any; /** - * Polymorphic Discriminator + * The root of blob path. Type: string (or Expression with resultType string). */ - type: "DocumentDbCollection"; + tableRootLocation?: any; /** - * Dataset description. + * The name of the Azure Blob. Type: string (or Expression with resultType string). */ - description?: string; + fileName?: any; /** - * Columns that define the structure of the dataset. Type: array (or Expression with resultType - * array), itemType: DatasetDataElement. + * The start of Azure Blob's modified datetime. Type: string (or Expression with resultType + * string). */ - structure?: any; + modifiedDatetimeStart?: any; /** - * Columns that define the physical type schema of the dataset. Type: array (or Expression with - * resultType array), itemType: DatasetSchemaDataElement. + * The end of Azure Blob's modified datetime. Type: string (or Expression with resultType + * string). */ - schema?: any; + modifiedDatetimeEnd?: any; /** - * Linked service reference. + * The format of the Azure Blob storage. */ - linkedServiceName: LinkedServiceReference; + format?: DatasetStorageFormatUnion; /** - * Parameters for dataset. + * The data compression method used for the blob storage. */ - parameters?: { [propertyName: string]: ParameterSpecification }; + compression?: DatasetCompressionUnion; +} + +/** + * Contains the possible cases for DatasetLocation. + */ +export type DatasetLocationUnion = DatasetLocation | HdfsLocation | HttpServerLocation | SftpLocation | FtpServerLocation | GoogleCloudStorageLocation | AzureFileStorageLocation | FileServerLocation | AmazonS3Location | AzureDataLakeStoreLocation | AzureBlobFSLocation | AzureBlobStorageLocation; + +/** + * Dataset location. + */ +export interface DatasetLocation { /** - * List of tags that can be used for describing the Dataset. + * Polymorphic Discriminator */ - annotations?: any[]; + type: "DatasetLocation"; /** - * The folder that this Dataset is in. If not specified, Dataset will appear at the root level. + * Specify the folder path of dataset. Type: string (or Expression with resultType string) */ - folder?: DatasetFolder; + folderPath?: any; /** - * Document Database collection name. Type: string (or Expression with resultType string). + * Specify the file name of dataset. Type: string (or Expression with resultType string). */ - collectionName: any; + fileName?: any; + /** + * Describes unknown properties. The value of an unknown property can be of "any" type. + */ + [property: string]: any; } /** - * Microsoft Azure CosmosDB (SQL API) Collection dataset. + * The location of HDFS. */ -export interface CosmosDbSqlApiCollectionDataset { +export interface HdfsLocation { /** * Polymorphic Discriminator */ - type: "CosmosDbSqlApiCollection"; + type: "HdfsLocation"; /** - * Dataset description. + * Specify the folder path of dataset. Type: string (or Expression with resultType string) */ - description?: string; + folderPath?: any; /** - * Columns that define the structure of the dataset. Type: array (or Expression with resultType - * array), itemType: DatasetDataElement. + * Specify the file name of dataset. Type: string (or Expression with resultType string). */ - structure?: any; + fileName?: any; +} + +/** + * The location of http server. + */ +export interface HttpServerLocation { /** - * Columns that define the physical type schema of the dataset. Type: array (or Expression with - * resultType array), itemType: DatasetSchemaDataElement. + * Polymorphic Discriminator */ - schema?: any; + type: "HttpServerLocation"; /** - * Linked service reference. + * Specify the folder path of dataset. Type: string (or Expression with resultType string) */ - linkedServiceName: LinkedServiceReference; + folderPath?: any; + /** + * Specify the file name of dataset. Type: string (or Expression with resultType string). + */ + fileName?: any; /** - * Parameters for dataset. + * Specify the relativeUrl of http server. Type: string (or Expression with resultType string) */ - parameters?: { [propertyName: string]: ParameterSpecification }; + relativeUrl?: any; +} + +/** + * The location of SFTP dataset. + */ +export interface SftpLocation { /** - * List of tags that can be used for describing the Dataset. + * Polymorphic Discriminator */ - annotations?: any[]; + type: "SftpLocation"; /** - * The folder that this Dataset is in. If not specified, Dataset will appear at the root level. + * Specify the folder path of dataset. Type: string (or Expression with resultType string) */ - folder?: DatasetFolder; + folderPath?: any; /** - * CosmosDB (SQL API) collection name. Type: string (or Expression with resultType string). + * Specify the file name of dataset. Type: string (or Expression with resultType string). */ - collectionName: any; + fileName?: any; } /** - * The custom dataset. + * The location of ftp server dataset. */ -export interface CustomDataset { +export interface FtpServerLocation { /** * Polymorphic Discriminator */ - type: "CustomDataset"; - /** - * Dataset description. - */ - description?: string; + type: "FtpServerLocation"; /** - * Columns that define the structure of the dataset. Type: array (or Expression with resultType - * array), itemType: DatasetDataElement. + * Specify the folder path of dataset. Type: string (or Expression with resultType string) */ - structure?: any; + folderPath?: any; /** - * Columns that define the physical type schema of the dataset. Type: array (or Expression with - * resultType array), itemType: DatasetSchemaDataElement. + * Specify the file name of dataset. Type: string (or Expression with resultType string). */ - schema?: any; + fileName?: any; +} + +/** + * The location of Google Cloud Storage dataset. + */ +export interface GoogleCloudStorageLocation { /** - * Linked service reference. + * Polymorphic Discriminator */ - linkedServiceName: LinkedServiceReference; + type: "GoogleCloudStorageLocation"; /** - * Parameters for dataset. + * Specify the folder path of dataset. Type: string (or Expression with resultType string) */ - parameters?: { [propertyName: string]: ParameterSpecification }; + folderPath?: any; /** - * List of tags that can be used for describing the Dataset. + * Specify the file name of dataset. Type: string (or Expression with resultType string). */ - annotations?: any[]; + fileName?: any; /** - * The folder that this Dataset is in. If not specified, Dataset will appear at the root level. + * Specify the bucketName of Google Cloud Storage. Type: string (or Expression with resultType + * string) */ - folder?: DatasetFolder; + bucketName?: any; /** - * Custom dataset properties. + * Specify the version of Google Cloud Storage. Type: string (or Expression with resultType + * string). */ - typeProperties?: any; + version?: any; } /** - * The Cassandra database dataset. + * The location of file server dataset. */ -export interface CassandraTableDataset { +export interface AzureFileStorageLocation { /** * Polymorphic Discriminator */ - type: "CassandraTable"; + type: "AzureFileStorageLocation"; /** - * Dataset description. + * Specify the folder path of dataset. Type: string (or Expression with resultType string) */ - description?: string; + folderPath?: any; /** - * Columns that define the structure of the dataset. Type: array (or Expression with resultType - * array), itemType: DatasetDataElement. + * Specify the file name of dataset. Type: string (or Expression with resultType string). */ - structure?: any; + fileName?: any; +} + +/** + * The location of file server dataset. + */ +export interface FileServerLocation { /** - * Columns that define the physical type schema of the dataset. Type: array (or Expression with - * resultType array), itemType: DatasetSchemaDataElement. + * Polymorphic Discriminator */ - schema?: any; + type: "FileServerLocation"; /** - * Linked service reference. + * Specify the folder path of dataset. Type: string (or Expression with resultType string) */ - linkedServiceName: LinkedServiceReference; + folderPath?: any; /** - * Parameters for dataset. + * Specify the file name of dataset. Type: string (or Expression with resultType string). */ - parameters?: { [propertyName: string]: ParameterSpecification }; + fileName?: any; +} + +/** + * The location of amazon S3 dataset. + */ +export interface AmazonS3Location { /** - * List of tags that can be used for describing the Dataset. + * Polymorphic Discriminator */ - annotations?: any[]; + type: "AmazonS3Location"; /** - * The folder that this Dataset is in. If not specified, Dataset will appear at the root level. + * Specify the folder path of dataset. Type: string (or Expression with resultType string) */ - folder?: DatasetFolder; + folderPath?: any; /** - * The table name of the Cassandra database. Type: string (or Expression with resultType string). + * Specify the file name of dataset. Type: string (or Expression with resultType string). */ - tableName?: any; + fileName?: any; /** - * The keyspace of the Cassandra database. Type: string (or Expression with resultType string). + * Specify the bucketName of amazon S3. Type: string (or Expression with resultType string) */ - keyspace?: any; + bucketName?: any; + /** + * Specify the version of amazon S3. Type: string (or Expression with resultType string). + */ + version?: any; } /** - * The Azure SQL Data Warehouse dataset. + * The location of azure data lake store dataset. */ -export interface AzureSqlDWTableDataset { +export interface AzureDataLakeStoreLocation { /** * Polymorphic Discriminator */ - type: "AzureSqlDWTable"; + type: "AzureDataLakeStoreLocation"; /** - * Dataset description. + * Specify the folder path of dataset. Type: string (or Expression with resultType string) */ - description?: string; + folderPath?: any; /** - * Columns that define the structure of the dataset. Type: array (or Expression with resultType - * array), itemType: DatasetDataElement. + * Specify the file name of dataset. Type: string (or Expression with resultType string). */ - structure?: any; + fileName?: any; +} + +/** + * The location of azure blobFS dataset. + */ +export interface AzureBlobFSLocation { /** - * Columns that define the physical type schema of the dataset. Type: array (or Expression with - * resultType array), itemType: DatasetSchemaDataElement. + * Polymorphic Discriminator */ - schema?: any; + type: "AzureBlobFSLocation"; /** - * Linked service reference. + * Specify the folder path of dataset. Type: string (or Expression with resultType string) */ - linkedServiceName: LinkedServiceReference; + folderPath?: any; /** - * Parameters for dataset. + * Specify the file name of dataset. Type: string (or Expression with resultType string). */ - parameters?: { [propertyName: string]: ParameterSpecification }; + fileName?: any; /** - * List of tags that can be used for describing the Dataset. + * Specify the fileSystem of azure blobFS. Type: string (or Expression with resultType string). */ - annotations?: any[]; + fileSystem?: any; +} + +/** + * The location of azure blob dataset. + */ +export interface AzureBlobStorageLocation { /** - * The folder that this Dataset is in. If not specified, Dataset will appear at the root level. + * Polymorphic Discriminator */ - folder?: DatasetFolder; + type: "AzureBlobStorageLocation"; /** - * This property will be retired. Please consider using schema + table properties instead. + * Specify the folder path of dataset. Type: string (or Expression with resultType string) */ - tableName?: any; + folderPath?: any; /** - * The schema name of the Azure SQL Data Warehouse. Type: string (or Expression with resultType - * string). + * Specify the file name of dataset. Type: string (or Expression with resultType string). */ - azureSqlDWTableDatasetSchema?: any; + fileName?: any; /** - * The table name of the Azure SQL Data Warehouse. Type: string (or Expression with resultType - * string). + * Specify the container of azure blob. Type: string (or Expression with resultType string). */ - table?: any; + container?: any; } /** - * The Azure SQL Managed Instance dataset. + * Binary dataset. */ -export interface AzureSqlMITableDataset { +export interface BinaryDataset { /** * Polymorphic Discriminator */ - type: "AzureSqlMITable"; + type: "Binary"; /** * Dataset description. */ @@ -12059,29 +12151,23 @@ export interface AzureSqlMITableDataset { */ folder?: DatasetFolder; /** - * This property will be retired. Please consider using schema + table properties instead. - */ - tableName?: any; - /** - * The schema name of the Azure SQL Managed Instance. Type: string (or Expression with resultType - * string). + * The location of the Binary storage. */ - azureSqlMITableDatasetSchema?: any; + location: DatasetLocationUnion; /** - * The table name of the Azure SQL Managed Instance dataset. Type: string (or Expression with - * resultType string). + * The data compression method used for the binary dataset. */ - table?: any; + compression?: DatasetCompressionUnion; } /** - * The Azure SQL Server database dataset. + * ORC dataset. */ -export interface AzureSqlTableDataset { +export interface OrcDataset { /** * Polymorphic Discriminator */ - type: "AzureSqlTable"; + type: "Orc"; /** * Dataset description. */ @@ -12113,28 +12199,23 @@ export interface AzureSqlTableDataset { */ folder?: DatasetFolder; /** - * This property will be retired. Please consider using schema + table properties instead. - */ - tableName?: any; - /** - * The schema name of the Azure SQL database. Type: string (or Expression with resultType - * string). + * The location of the ORC data storage. */ - azureSqlTableDatasetSchema?: any; + location: DatasetLocationUnion; /** - * The table name of the Azure SQL database. Type: string (or Expression with resultType string). + * Possible values include: 'none', 'zlib', 'snappy' */ - table?: any; + orcCompressionCodec?: OrcCompressionCodec; } /** - * The Azure Table storage dataset. + * Json dataset. */ -export interface AzureTableDataset { +export interface JsonDataset { /** * Polymorphic Discriminator */ - type: "AzureTable"; + type: "Json"; /** * Dataset description. */ @@ -12166,20 +12247,31 @@ export interface AzureTableDataset { */ folder?: DatasetFolder; /** - * The table name of the Azure Table storage. Type: string (or Expression with resultType - * string). + * The location of the json data storage. */ - tableName: any; + location: DatasetLocationUnion; + /** + * The code page name of the preferred encoding. If not specified, the default value is UTF-8, + * unless BOM denotes another Unicode encoding. Refer to the name column of the table in the + * following link to set supported values: + * https://msdn.microsoft.com/library/system.text.encoding.aspx. Type: string (or Expression with + * resultType string). + */ + encodingName?: any; + /** + * The data compression method used for the json dataset. + */ + compression?: DatasetCompressionUnion; } /** - * The Azure Blob storage. + * Delimited text dataset. */ -export interface AzureBlobDataset { +export interface DelimitedTextDataset { /** * Polymorphic Discriminator */ - type: "AzureBlob"; + type: "DelimitedText"; /** * Dataset description. */ @@ -12211,691 +12303,705 @@ export interface AzureBlobDataset { */ folder?: DatasetFolder; /** - * The path of the Azure Blob storage. Type: string (or Expression with resultType string). - */ - folderPath?: any; - /** - * The root of blob path. Type: string (or Expression with resultType string). - */ - tableRootLocation?: any; - /** - * The name of the Azure Blob. Type: string (or Expression with resultType string). + * The location of the delimited text storage. */ - fileName?: any; + location: DatasetLocationUnion; /** - * The start of Azure Blob's modified datetime. Type: string (or Expression with resultType - * string). + * The column delimiter. Type: string (or Expression with resultType string). */ - modifiedDatetimeStart?: any; + columnDelimiter?: any; /** - * The end of Azure Blob's modified datetime. Type: string (or Expression with resultType - * string). + * The row delimiter. Type: string (or Expression with resultType string). */ - modifiedDatetimeEnd?: any; + rowDelimiter?: any; /** - * The format of the Azure Blob storage. + * The code page name of the preferred encoding. If miss, the default value is UTF-8, unless BOM + * denotes another Unicode encoding. Refer to the name column of the table in the following link + * to set supported values: https://msdn.microsoft.com/library/system.text.encoding.aspx. Type: + * string (or Expression with resultType string). */ - format?: DatasetStorageFormatUnion; + encodingName?: any; + compressionCodec?: any; /** - * The data compression method used for the blob storage. + * The data compression method used for DelimitedText. */ - compression?: DatasetCompressionUnion; -} - -/** - * Contains the possible cases for DatasetLocation. - */ -export type DatasetLocationUnion = DatasetLocation | HdfsLocation | HttpServerLocation | SftpLocation | FtpServerLocation | GoogleCloudStorageLocation | AzureFileStorageLocation | FileServerLocation | AmazonS3Location | AzureDataLakeStoreLocation | AzureBlobFSLocation | AzureBlobStorageLocation; - -/** - * Dataset location. - */ -export interface DatasetLocation { + compressionLevel?: any; /** - * Polymorphic Discriminator + * The quote character. Type: string (or Expression with resultType string). */ - type: "DatasetLocation"; + quoteChar?: any; /** - * Specify the folder path of dataset. Type: string (or Expression with resultType string) + * The escape character. Type: string (or Expression with resultType string). */ - folderPath?: any; + escapeChar?: any; /** - * Specify the file name of dataset. Type: string (or Expression with resultType string). + * When used as input, treat the first row of data as headers. When used as output,write the + * headers into the output as the first row of data. The default value is false. Type: boolean + * (or Expression with resultType boolean). */ - fileName?: any; + firstRowAsHeader?: any; /** - * Describes unknown properties. The value of an unknown property can be of "any" type. + * The null value string. Type: string (or Expression with resultType string). */ - [property: string]: any; + nullValue?: any; } /** - * The location of HDFS. + * Parquet dataset. */ -export interface HdfsLocation { +export interface ParquetDataset { /** * Polymorphic Discriminator */ - type: "HdfsLocation"; - /** - * Specify the folder path of dataset. Type: string (or Expression with resultType string) - */ - folderPath?: any; + type: "Parquet"; /** - * Specify the file name of dataset. Type: string (or Expression with resultType string). + * Dataset description. */ - fileName?: any; -} - -/** - * The location of http server. - */ -export interface HttpServerLocation { + description?: string; /** - * Polymorphic Discriminator + * Columns that define the structure of the dataset. Type: array (or Expression with resultType + * array), itemType: DatasetDataElement. */ - type: "HttpServerLocation"; + structure?: any; /** - * Specify the folder path of dataset. Type: string (or Expression with resultType string) + * Columns that define the physical type schema of the dataset. Type: array (or Expression with + * resultType array), itemType: DatasetSchemaDataElement. */ - folderPath?: any; + schema?: any; /** - * Specify the file name of dataset. Type: string (or Expression with resultType string). + * Linked service reference. */ - fileName?: any; + linkedServiceName: LinkedServiceReference; /** - * Specify the relativeUrl of http server. Type: string (or Expression with resultType string) + * Parameters for dataset. */ - relativeUrl?: any; -} - -/** - * The location of SFTP dataset. - */ -export interface SftpLocation { + parameters?: { [propertyName: string]: ParameterSpecification }; /** - * Polymorphic Discriminator + * List of tags that can be used for describing the Dataset. */ - type: "SftpLocation"; + annotations?: any[]; /** - * Specify the folder path of dataset. Type: string (or Expression with resultType string) + * The folder that this Dataset is in. If not specified, Dataset will appear at the root level. */ - folderPath?: any; + folder?: DatasetFolder; /** - * Specify the file name of dataset. Type: string (or Expression with resultType string). + * The location of the parquet storage. */ - fileName?: any; + location: DatasetLocationUnion; + compressionCodec?: any; } /** - * The location of ftp server dataset. + * Avro dataset. */ -export interface FtpServerLocation { +export interface AvroDataset { /** * Polymorphic Discriminator */ - type: "FtpServerLocation"; + type: "Avro"; /** - * Specify the folder path of dataset. Type: string (or Expression with resultType string) + * Dataset description. */ - folderPath?: any; + description?: string; /** - * Specify the file name of dataset. Type: string (or Expression with resultType string). + * Columns that define the structure of the dataset. Type: array (or Expression with resultType + * array), itemType: DatasetDataElement. */ - fileName?: any; -} - -/** - * The location of Google Cloud Storage dataset. - */ -export interface GoogleCloudStorageLocation { + structure?: any; /** - * Polymorphic Discriminator + * Columns that define the physical type schema of the dataset. Type: array (or Expression with + * resultType array), itemType: DatasetSchemaDataElement. */ - type: "GoogleCloudStorageLocation"; + schema?: any; /** - * Specify the folder path of dataset. Type: string (or Expression with resultType string) + * Linked service reference. */ - folderPath?: any; + linkedServiceName: LinkedServiceReference; /** - * Specify the file name of dataset. Type: string (or Expression with resultType string). + * Parameters for dataset. */ - fileName?: any; + parameters?: { [propertyName: string]: ParameterSpecification }; /** - * Specify the bucketName of Google Cloud Storage. Type: string (or Expression with resultType - * string) + * List of tags that can be used for describing the Dataset. */ - bucketName?: any; + annotations?: any[]; /** - * Specify the version of Google Cloud Storage. Type: string (or Expression with resultType - * string). + * The folder that this Dataset is in. If not specified, Dataset will appear at the root level. */ - version?: any; + folder?: DatasetFolder; + /** + * The location of the avro storage. + */ + location: DatasetLocationUnion; + /** + * Possible values include: 'none', 'deflate', 'snappy', 'xz', 'bzip2' + */ + avroCompressionCodec?: AvroCompressionCodec; + avroCompressionLevel?: number; } /** - * The location of file server dataset. + * A single Amazon Simple Storage Service (S3) object or a set of S3 objects. */ -export interface AzureFileStorageLocation { +export interface AmazonS3Dataset { /** * Polymorphic Discriminator */ - type: "AzureFileStorageLocation"; + type: "AmazonS3Object"; /** - * Specify the folder path of dataset. Type: string (or Expression with resultType string) + * Dataset description. */ - folderPath?: any; + description?: string; /** - * Specify the file name of dataset. Type: string (or Expression with resultType string). + * Columns that define the structure of the dataset. Type: array (or Expression with resultType + * array), itemType: DatasetDataElement. */ - fileName?: any; -} - -/** - * The location of file server dataset. - */ -export interface FileServerLocation { + structure?: any; /** - * Polymorphic Discriminator + * Columns that define the physical type schema of the dataset. Type: array (or Expression with + * resultType array), itemType: DatasetSchemaDataElement. */ - type: "FileServerLocation"; + schema?: any; /** - * Specify the folder path of dataset. Type: string (or Expression with resultType string) + * Linked service reference. */ - folderPath?: any; + linkedServiceName: LinkedServiceReference; /** - * Specify the file name of dataset. Type: string (or Expression with resultType string). + * Parameters for dataset. */ - fileName?: any; -} - -/** - * The location of amazon S3 dataset. - */ -export interface AmazonS3Location { + parameters?: { [propertyName: string]: ParameterSpecification }; /** - * Polymorphic Discriminator + * List of tags that can be used for describing the Dataset. */ - type: "AmazonS3Location"; + annotations?: any[]; /** - * Specify the folder path of dataset. Type: string (or Expression with resultType string) + * The folder that this Dataset is in. If not specified, Dataset will appear at the root level. */ - folderPath?: any; + folder?: DatasetFolder; /** - * Specify the file name of dataset. Type: string (or Expression with resultType string). + * The name of the Amazon S3 bucket. Type: string (or Expression with resultType string). */ - fileName?: any; + bucketName: any; /** - * Specify the bucketName of amazon S3. Type: string (or Expression with resultType string) + * The key of the Amazon S3 object. Type: string (or Expression with resultType string). */ - bucketName?: any; + key?: any; /** - * Specify the version of amazon S3. Type: string (or Expression with resultType string). + * The prefix filter for the S3 object name. Type: string (or Expression with resultType string). + */ + prefix?: any; + /** + * The version for the S3 object. Type: string (or Expression with resultType string). */ version?: any; -} - -/** - * The location of azure data lake store dataset. - */ -export interface AzureDataLakeStoreLocation { /** - * Polymorphic Discriminator + * The start of S3 object's modified datetime. Type: string (or Expression with resultType + * string). */ - type: "AzureDataLakeStoreLocation"; + modifiedDatetimeStart?: any; /** - * Specify the folder path of dataset. Type: string (or Expression with resultType string) + * The end of S3 object's modified datetime. Type: string (or Expression with resultType string). + */ + modifiedDatetimeEnd?: any; + /** + * The format of files. */ - folderPath?: any; + format?: DatasetStorageFormatUnion; /** - * Specify the file name of dataset. Type: string (or Expression with resultType string). + * The data compression method used for the Amazon S3 object. */ - fileName?: any; + compression?: DatasetCompressionUnion; } /** - * The location of azure blobFS dataset. + * Trigger that allows the referenced pipeline to depend on other pipeline runs based on + * runDimension Name/Value pairs. Upstream pipelines should declare the same runDimension Name and + * their runs should have the values for those runDimensions. The referenced pipeline run would be + * triggered if the values for the runDimension match for all upstream pipeline runs. */ -export interface AzureBlobFSLocation { +export interface ChainingTrigger { /** * Polymorphic Discriminator */ - type: "AzureBlobFSLocation"; - /** - * Specify the folder path of dataset. Type: string (or Expression with resultType string) - */ - folderPath?: any; + type: "ChainingTrigger"; /** - * Specify the file name of dataset. Type: string (or Expression with resultType string). + * Trigger description. */ - fileName?: any; + description?: string; /** - * Specify the fileSystem of azure blobFS. Type: string (or Expression with resultType string). + * Indicates if trigger is running or not. Updated when Start/Stop APIs are called on the + * Trigger. Possible values include: 'Started', 'Stopped', 'Disabled' + * **NOTE: This property will not be serialized. It can only be populated by the server.** */ - fileSystem?: any; -} - -/** - * The location of azure blob dataset. - */ -export interface AzureBlobStorageLocation { + readonly runtimeState?: TriggerRuntimeState; /** - * Polymorphic Discriminator + * List of tags that can be used for describing the trigger. */ - type: "AzureBlobStorageLocation"; + annotations?: any[]; /** - * Specify the folder path of dataset. Type: string (or Expression with resultType string) + * Pipeline for which runs are created when all upstream pipelines complete successfully. */ - folderPath?: any; + pipelineProperty: TriggerPipelineReference; /** - * Specify the file name of dataset. Type: string (or Expression with resultType string). + * Upstream Pipelines. */ - fileName?: any; + dependsOn: PipelineReference[]; /** - * Specify the container of azure blob. Type: string (or Expression with resultType string). + * Run Dimension property that needs to be emitted by upstream pipelines. */ - container?: any; + runDimension: string; } /** - * Binary dataset. + * Trigger that schedules pipeline reruns for all fixed time interval windows from a requested + * start time to requested end time. */ -export interface BinaryDataset { +export interface RerunTumblingWindowTrigger { /** * Polymorphic Discriminator */ - type: "Binary"; + type: "RerunTumblingWindowTrigger"; /** - * Dataset description. + * Trigger description. */ description?: string; /** - * Columns that define the structure of the dataset. Type: array (or Expression with resultType - * array), itemType: DatasetDataElement. - */ - structure?: any; - /** - * Columns that define the physical type schema of the dataset. Type: array (or Expression with - * resultType array), itemType: DatasetSchemaDataElement. + * Indicates if trigger is running or not. Updated when Start/Stop APIs are called on the + * Trigger. Possible values include: 'Started', 'Stopped', 'Disabled' + * **NOTE: This property will not be serialized. It can only be populated by the server.** */ - schema?: any; + readonly runtimeState?: TriggerRuntimeState; /** - * Linked service reference. + * List of tags that can be used for describing the trigger. */ - linkedServiceName: LinkedServiceReference; + annotations?: any[]; /** - * Parameters for dataset. + * The parent trigger reference. */ - parameters?: { [propertyName: string]: ParameterSpecification }; + parentTrigger: any; /** - * List of tags that can be used for describing the Dataset. + * The start time for the time period for which restatement is initiated. Only UTC time is + * currently supported. */ - annotations?: any[]; + requestedStartTime: Date; /** - * The folder that this Dataset is in. If not specified, Dataset will appear at the root level. + * The end time for the time period for which restatement is initiated. Only UTC time is + * currently supported. */ - folder?: DatasetFolder; + requestedEndTime: Date; /** - * The location of the Binary storage. + * The max number of parallel time windows (ready for execution) for which a rerun is triggered. */ - location: DatasetLocationUnion; + rerunConcurrency: number; +} + +/** + * Contains the possible cases for DependencyReference. + */ +export type DependencyReferenceUnion = DependencyReference | SelfDependencyTumblingWindowTriggerReference | TriggerDependencyReferenceUnion; + +/** + * Referenced dependency. + */ +export interface DependencyReference { /** - * The data compression method used for the binary dataset. + * Polymorphic Discriminator */ - compression?: DatasetCompressionUnion; + type: "DependencyReference"; } /** - * ORC dataset. + * Self referenced tumbling window trigger dependency. */ -export interface OrcDataset { +export interface SelfDependencyTumblingWindowTriggerReference { /** * Polymorphic Discriminator */ - type: "Orc"; + type: "SelfDependencyTumblingWindowTriggerReference"; /** - * Dataset description. + * Timespan applied to the start time of a tumbling window when evaluating dependency. */ - description?: string; + offset: string; /** - * Columns that define the structure of the dataset. Type: array (or Expression with resultType - * array), itemType: DatasetDataElement. + * The size of the window when evaluating the dependency. If undefined the frequency of the + * tumbling window will be used. */ - structure?: any; + size?: string; +} + +/** + * Trigger reference type. + */ +export interface TriggerReference { /** - * Columns that define the physical type schema of the dataset. Type: array (or Expression with - * resultType array), itemType: DatasetSchemaDataElement. + * Reference trigger name. */ - schema?: any; + referenceName: string; +} + +/** + * Contains the possible cases for TriggerDependencyReference. + */ +export type TriggerDependencyReferenceUnion = TriggerDependencyReference | TumblingWindowTriggerDependencyReference; + +/** + * Trigger referenced dependency. + */ +export interface TriggerDependencyReference { /** - * Linked service reference. + * Polymorphic Discriminator */ - linkedServiceName: LinkedServiceReference; + type: "TriggerDependencyReference"; /** - * Parameters for dataset. + * Referenced trigger. */ - parameters?: { [propertyName: string]: ParameterSpecification }; + referenceTrigger: TriggerReference; +} + +/** + * Referenced tumbling window trigger dependency. + */ +export interface TumblingWindowTriggerDependencyReference { /** - * List of tags that can be used for describing the Dataset. + * Polymorphic Discriminator */ - annotations?: any[]; + type: "TumblingWindowTriggerDependencyReference"; /** - * The folder that this Dataset is in. If not specified, Dataset will appear at the root level. + * Referenced trigger. */ - folder?: DatasetFolder; + referenceTrigger: TriggerReference; /** - * The location of the ORC data storage. + * Timespan applied to the start time of a tumbling window when evaluating dependency. */ - location: DatasetLocationUnion; + offset?: string; /** - * Possible values include: 'none', 'zlib', 'snappy' + * The size of the window when evaluating the dependency. If undefined the frequency of the + * tumbling window will be used. */ - orcCompressionCodec?: OrcCompressionCodec; + size?: string; } /** - * Json dataset. + * Execution policy for an activity. */ -export interface JsonDataset { +export interface RetryPolicy { + /** + * Maximum ordinary retry attempts. Default is 0. Type: integer (or Expression with resultType + * integer), minimum: 0. + */ + count?: any; + /** + * Interval between retries in seconds. Default is 30. + */ + intervalInSeconds?: number; +} + +/** + * Trigger that schedules pipeline runs for all fixed time interval windows from a start time + * without gaps and also supports backfill scenarios (when start time is in the past). + */ +export interface TumblingWindowTrigger { /** * Polymorphic Discriminator */ - type: "Json"; + type: "TumblingWindowTrigger"; /** - * Dataset description. + * Trigger description. */ description?: string; /** - * Columns that define the structure of the dataset. Type: array (or Expression with resultType - * array), itemType: DatasetDataElement. + * Indicates if trigger is running or not. Updated when Start/Stop APIs are called on the + * Trigger. Possible values include: 'Started', 'Stopped', 'Disabled' + * **NOTE: This property will not be serialized. It can only be populated by the server.** */ - structure?: any; + readonly runtimeState?: TriggerRuntimeState; /** - * Columns that define the physical type schema of the dataset. Type: array (or Expression with - * resultType array), itemType: DatasetSchemaDataElement. + * List of tags that can be used for describing the trigger. */ - schema?: any; + annotations?: any[]; /** - * Linked service reference. + * Pipeline for which runs are created when an event is fired for trigger window that is ready. */ - linkedServiceName: LinkedServiceReference; + pipelineProperty: TriggerPipelineReference; /** - * Parameters for dataset. + * The frequency of the time windows. Possible values include: 'Minute', 'Hour' */ - parameters?: { [propertyName: string]: ParameterSpecification }; + frequency: TumblingWindowFrequency; /** - * List of tags that can be used for describing the Dataset. + * The interval of the time windows. The minimum interval allowed is 15 Minutes. */ - annotations?: any[]; + interval: number; /** - * The folder that this Dataset is in. If not specified, Dataset will appear at the root level. + * The start time for the time period for the trigger during which events are fired for windows + * that are ready. Only UTC time is currently supported. */ - folder?: DatasetFolder; + startTime: Date; /** - * The location of the json data storage. + * The end time for the time period for the trigger during which events are fired for windows + * that are ready. Only UTC time is currently supported. */ - location: DatasetLocationUnion; + endTime?: Date; /** - * The code page name of the preferred encoding. If not specified, the default value is UTF-8, - * unless BOM denotes another Unicode encoding. Refer to the name column of the table in the - * following link to set supported values: - * https://msdn.microsoft.com/library/system.text.encoding.aspx. Type: string (or Expression with - * resultType string). + * Specifies how long the trigger waits past due time before triggering new run. It doesn't alter + * window start and end time. The default is 0. Type: string (or Expression with resultType + * string), pattern: ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). */ - encodingName?: any; + delay?: any; /** - * The data compression method used for the json dataset. + * The max number of parallel time windows (ready for execution) for which a new run is + * triggered. + */ + maxConcurrency: number; + /** + * Retry policy that will be applied for failed pipeline runs. + */ + retryPolicy?: RetryPolicy; + /** + * Triggers that this trigger depends on. Only tumbling window triggers are supported. */ - compression?: DatasetCompressionUnion; + dependsOn?: DependencyReferenceUnion[]; } /** - * Delimited text dataset. + * Contains the possible cases for MultiplePipelineTrigger. */ -export interface DelimitedTextDataset { +export type MultiplePipelineTriggerUnion = MultiplePipelineTrigger | BlobEventsTrigger | BlobTrigger | ScheduleTrigger; + +/** + * Base class for all triggers that support one to many model for trigger to pipeline. + */ +export interface MultiplePipelineTrigger { /** * Polymorphic Discriminator */ - type: "DelimitedText"; + type: "MultiplePipelineTrigger"; /** - * Dataset description. + * Trigger description. */ description?: string; /** - * Columns that define the structure of the dataset. Type: array (or Expression with resultType - * array), itemType: DatasetDataElement. - */ - structure?: any; - /** - * Columns that define the physical type schema of the dataset. Type: array (or Expression with - * resultType array), itemType: DatasetSchemaDataElement. - */ - schema?: any; - /** - * Linked service reference. + * Indicates if trigger is running or not. Updated when Start/Stop APIs are called on the + * Trigger. Possible values include: 'Started', 'Stopped', 'Disabled' + * **NOTE: This property will not be serialized. It can only be populated by the server.** */ - linkedServiceName: LinkedServiceReference; + readonly runtimeState?: TriggerRuntimeState; /** - * Parameters for dataset. + * List of tags that can be used for describing the trigger. */ - parameters?: { [propertyName: string]: ParameterSpecification }; + annotations?: any[]; /** - * List of tags that can be used for describing the Dataset. + * Pipelines that need to be started. */ - annotations?: any[]; + pipelines?: TriggerPipelineReference[]; +} + +/** + * Trigger that runs every time a Blob event occurs. + */ +export interface BlobEventsTrigger { /** - * The folder that this Dataset is in. If not specified, Dataset will appear at the root level. + * Polymorphic Discriminator */ - folder?: DatasetFolder; + type: "BlobEventsTrigger"; /** - * The location of the delimited text storage. + * Trigger description. */ - location: DatasetLocationUnion; + description?: string; /** - * The column delimiter. Type: string (or Expression with resultType string). + * Indicates if trigger is running or not. Updated when Start/Stop APIs are called on the + * Trigger. Possible values include: 'Started', 'Stopped', 'Disabled' + * **NOTE: This property will not be serialized. It can only be populated by the server.** */ - columnDelimiter?: any; + readonly runtimeState?: TriggerRuntimeState; /** - * The row delimiter. Type: string (or Expression with resultType string). + * List of tags that can be used for describing the trigger. */ - rowDelimiter?: any; + annotations?: any[]; /** - * The code page name of the preferred encoding. If miss, the default value is UTF-8, unless BOM - * denotes another Unicode encoding. Refer to the name column of the table in the following link - * to set supported values: https://msdn.microsoft.com/library/system.text.encoding.aspx. Type: - * string (or Expression with resultType string). + * Pipelines that need to be started. */ - encodingName?: any; - compressionCodec?: any; + pipelines?: TriggerPipelineReference[]; /** - * The data compression method used for DelimitedText. + * The blob path must begin with the pattern provided for trigger to fire. For example, + * '/records/blobs/december/' will only fire the trigger for blobs in the december folder under + * the records container. At least one of these must be provided: blobPathBeginsWith, + * blobPathEndsWith. */ - compressionLevel?: any; + blobPathBeginsWith?: string; /** - * The quote character. Type: string (or Expression with resultType string). + * The blob path must end with the pattern provided for trigger to fire. For example, + * 'december/boxes.csv' will only fire the trigger for blobs named boxes in a december folder. At + * least one of these must be provided: blobPathBeginsWith, blobPathEndsWith. */ - quoteChar?: any; + blobPathEndsWith?: string; /** - * The escape character. Type: string (or Expression with resultType string). + * If set to true, blobs with zero bytes will be ignored. */ - escapeChar?: any; + ignoreEmptyBlobs?: boolean; /** - * When used as input, treat the first row of data as headers. When used as output,write the - * headers into the output as the first row of data. The default value is false. Type: boolean - * (or Expression with resultType boolean). + * The type of events that cause this trigger to fire. */ - firstRowAsHeader?: any; + events: BlobEventTypes[]; /** - * The null value string. Type: string (or Expression with resultType string). + * The ARM resource ID of the Storage Account. */ - nullValue?: any; + scope: string; } /** - * Parquet dataset. + * Trigger that runs every time the selected Blob container changes. */ -export interface ParquetDataset { +export interface BlobTrigger { /** * Polymorphic Discriminator */ - type: "Parquet"; + type: "BlobTrigger"; /** - * Dataset description. + * Trigger description. */ description?: string; /** - * Columns that define the structure of the dataset. Type: array (or Expression with resultType - * array), itemType: DatasetDataElement. - */ - structure?: any; - /** - * Columns that define the physical type schema of the dataset. Type: array (or Expression with - * resultType array), itemType: DatasetSchemaDataElement. + * Indicates if trigger is running or not. Updated when Start/Stop APIs are called on the + * Trigger. Possible values include: 'Started', 'Stopped', 'Disabled' + * **NOTE: This property will not be serialized. It can only be populated by the server.** */ - schema?: any; + readonly runtimeState?: TriggerRuntimeState; /** - * Linked service reference. + * List of tags that can be used for describing the trigger. */ - linkedServiceName: LinkedServiceReference; + annotations?: any[]; /** - * Parameters for dataset. + * Pipelines that need to be started. */ - parameters?: { [propertyName: string]: ParameterSpecification }; + pipelines?: TriggerPipelineReference[]; /** - * List of tags that can be used for describing the Dataset. + * The path of the container/folder that will trigger the pipeline. */ - annotations?: any[]; + folderPath: string; /** - * The folder that this Dataset is in. If not specified, Dataset will appear at the root level. + * The max number of parallel files to handle when it is triggered. */ - folder?: DatasetFolder; + maxConcurrency: number; /** - * The location of the parquet storage. + * The Azure Storage linked service reference. */ - location: DatasetLocationUnion; - compressionCodec?: any; + linkedService: LinkedServiceReference; } /** - * Avro dataset. + * The recurrence schedule occurrence. */ -export interface AvroDataset { - /** - * Polymorphic Discriminator - */ - type: "Avro"; +export interface RecurrenceScheduleOccurrence { /** - * Dataset description. + * The day of the week. Possible values include: 'Sunday', 'Monday', 'Tuesday', 'Wednesday', + * 'Thursday', 'Friday', 'Saturday' */ - description?: string; + day?: DayOfWeek; /** - * Columns that define the structure of the dataset. Type: array (or Expression with resultType - * array), itemType: DatasetDataElement. + * The occurrence. */ - structure?: any; + occurrence?: number; /** - * Columns that define the physical type schema of the dataset. Type: array (or Expression with - * resultType array), itemType: DatasetSchemaDataElement. + * Describes unknown properties. The value of an unknown property can be of "any" type. */ - schema?: any; + [property: string]: any; +} + +/** + * The recurrence schedule. + */ +export interface RecurrenceSchedule { /** - * Linked service reference. + * The minutes. */ - linkedServiceName: LinkedServiceReference; + minutes?: number[]; /** - * Parameters for dataset. + * The hours. */ - parameters?: { [propertyName: string]: ParameterSpecification }; + hours?: number[]; /** - * List of tags that can be used for describing the Dataset. + * The days of the week. */ - annotations?: any[]; + weekDays?: DaysOfWeek[]; /** - * The folder that this Dataset is in. If not specified, Dataset will appear at the root level. + * The month days. */ - folder?: DatasetFolder; + monthDays?: number[]; /** - * The location of the avro storage. + * The monthly occurrences. */ - location: DatasetLocationUnion; + monthlyOccurrences?: RecurrenceScheduleOccurrence[]; /** - * Possible values include: 'none', 'deflate', 'snappy', 'xz', 'bzip2' + * Describes unknown properties. The value of an unknown property can be of "any" type. */ - avroCompressionCodec?: AvroCompressionCodec; - avroCompressionLevel?: number; + [property: string]: any; } /** - * A single Amazon Simple Storage Service (S3) object or a set of S3 objects. + * The workflow trigger recurrence. */ -export interface AmazonS3Dataset { - /** - * Polymorphic Discriminator - */ - type: "AmazonS3Object"; - /** - * Dataset description. - */ - description?: string; - /** - * Columns that define the structure of the dataset. Type: array (or Expression with resultType - * array), itemType: DatasetDataElement. - */ - structure?: any; +export interface ScheduleTriggerRecurrence { /** - * Columns that define the physical type schema of the dataset. Type: array (or Expression with - * resultType array), itemType: DatasetSchemaDataElement. + * The frequency. Possible values include: 'NotSpecified', 'Minute', 'Hour', 'Day', 'Week', + * 'Month', 'Year' */ - schema?: any; + frequency?: RecurrenceFrequency; /** - * Linked service reference. + * The interval. */ - linkedServiceName: LinkedServiceReference; + interval?: number; /** - * Parameters for dataset. + * The start time. */ - parameters?: { [propertyName: string]: ParameterSpecification }; + startTime?: Date; /** - * List of tags that can be used for describing the Dataset. + * The end time. */ - annotations?: any[]; + endTime?: Date; /** - * The folder that this Dataset is in. If not specified, Dataset will appear at the root level. + * The time zone. */ - folder?: DatasetFolder; + timeZone?: string; /** - * The name of the Amazon S3 bucket. Type: string (or Expression with resultType string). + * The recurrence schedule. */ - bucketName: any; + schedule?: RecurrenceSchedule; /** - * The key of the Amazon S3 object. Type: string (or Expression with resultType string). + * Describes unknown properties. The value of an unknown property can be of "any" type. */ - key?: any; + [property: string]: any; +} + +/** + * Trigger that creates pipeline runs periodically, on schedule. + */ +export interface ScheduleTrigger { /** - * The prefix filter for the S3 object name. Type: string (or Expression with resultType string). + * Polymorphic Discriminator */ - prefix?: any; + type: "ScheduleTrigger"; /** - * The version for the S3 object. Type: string (or Expression with resultType string). + * Trigger description. */ - version?: any; + description?: string; /** - * The start of S3 object's modified datetime. Type: string (or Expression with resultType - * string). + * Indicates if trigger is running or not. Updated when Start/Stop APIs are called on the + * Trigger. Possible values include: 'Started', 'Stopped', 'Disabled' + * **NOTE: This property will not be serialized. It can only be populated by the server.** */ - modifiedDatetimeStart?: any; + readonly runtimeState?: TriggerRuntimeState; /** - * The end of S3 object's modified datetime. Type: string (or Expression with resultType string). + * List of tags that can be used for describing the trigger. */ - modifiedDatetimeEnd?: any; + annotations?: any[]; /** - * The format of files. + * Pipelines that need to be started. */ - format?: DatasetStorageFormatUnion; + pipelines?: TriggerPipelineReference[]; /** - * The data compression method used for the Amazon S3 object. + * Recurrence schedule configuration. */ - compression?: DatasetCompressionUnion; + recurrence: ScheduleTriggerRecurrence; } /** @@ -13620,10 +13726,51 @@ export interface WebActivity { connectVia?: IntegrationRuntimeReference; } +/** + * Contains the possible cases for ExportSettings. + */ +export type ExportSettingsUnion = ExportSettings | SnowflakeExportCopyCommand; + +/** + * Export command settings. + */ +export interface ExportSettings { + /** + * Polymorphic Discriminator + */ + type: "ExportSettings"; + /** + * Describes unknown properties. The value of an unknown property can be of "any" type. + */ + [property: string]: any; +} + +/** + * Snowflake export command settings. + */ +export interface SnowflakeExportCopyCommand { + /** + * Polymorphic Discriminator + */ + type: "SnowflakeExportCopyCommand"; + /** + * Additional copy options directly passed to snowflake Copy Command. Type: key value pairs + * (value should be string type) (or Expression with resultType object). Example: + * "additionalCopyOptions": { "DATE_FORMAT": "MM/DD/YYYY", "TIME_FORMAT": "'HH24:MI:SS.FF'" } + */ + additionalCopyOptions?: { [propertyName: string]: any }; + /** + * Additional format options directly passed to snowflake Copy Command. Type: key value pairs + * (value should be string type) (or Expression with resultType object). Example: + * "additionalFormatOptions": { "OVERWRITE": "TRUE", "MAX_FILE_SIZE": "'FALSE'" } + */ + additionalFormatOptions?: { [propertyName: string]: any }; +} + /** * Contains the possible cases for CopySource. */ -export type CopySourceUnion = CopySource | HttpSource | AzureBlobFSSource | AzureDataLakeStoreSource | Office365Source | CosmosDbMongoDbApiSource | MongoDbV2Source | MongoDbSource | WebSource | OracleSource | AzureDataExplorerSource | HdfsSource | FileSystemSource | RestSource | SalesforceServiceCloudSource | ODataSource | MicrosoftAccessSource | RelationalSource | CommonDataServiceForAppsSource | DynamicsCrmSource | DynamicsSource | CosmosDbSqlApiSource | DocumentDbCollectionSource | BlobSource | TabularSourceUnion | BinarySource | OrcSource | JsonSource | DelimitedTextSource | ParquetSource | AvroSource; +export type CopySourceUnion = CopySource | SnowflakeSource | HttpSource | AzureBlobFSSource | AzureDataLakeStoreSource | Office365Source | CosmosDbMongoDbApiSource | MongoDbV2Source | MongoDbSource | WebSource | OracleSource | AzureDataExplorerSource | HdfsSource | FileSystemSource | RestSource | SalesforceServiceCloudSource | ODataSource | MicrosoftAccessSource | RelationalSource | CommonDataServiceForAppsSource | DynamicsCrmSource | DynamicsSource | CosmosDbSqlApiSource | DocumentDbCollectionSource | BlobSource | TabularSourceUnion | BinarySource | OrcSource | JsonSource | DelimitedTextSource | ParquetSource | AvroSource; /** * A copy activity source. @@ -13653,6 +13800,38 @@ export interface CopySource { [property: string]: any; } +/** + * A copy activity snowflake source. + */ +export interface SnowflakeSource { + /** + * Polymorphic Discriminator + */ + type: "SnowflakeSource"; + /** + * Source retry count. Type: integer (or Expression with resultType integer). + */ + sourceRetryCount?: any; + /** + * Source retry wait. Type: string (or Expression with resultType string), pattern: + * ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + */ + sourceRetryWait?: any; + /** + * The maximum concurrent connection count for the source data store. Type: integer (or + * Expression with resultType integer). + */ + maxConcurrentConnections?: any; + /** + * Snowflake Sql query. Type: string (or Expression with resultType string). + */ + query?: any; + /** + * Snowflake export settings. + */ + exportSettings?: SnowflakeExportCopyCommand; +} + /** * A copy activity source for an HTTP file. */ @@ -13804,6 +13983,20 @@ export interface Office365Source { outputColumns?: any; } +/** + * Specify the column name and value of additional columns. + */ +export interface AdditionalColumns { + /** + * Additional column name. Type: string (or Expression with resultType string). + */ + name?: any; + /** + * Additional column value. Type: string (or Expression with resultType string). + */ + value?: any; +} + /** * Cursor methods for Mongodb query */ @@ -13880,6 +14073,11 @@ export interface CosmosDbMongoDbApiSource { * ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). */ queryTimeout?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; } /** @@ -13926,6 +14124,11 @@ export interface MongoDbV2Source { * ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). */ queryTimeout?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; } /** @@ -13955,6 +14158,11 @@ export interface MongoDbSource { * resultType string). */ query?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; } /** @@ -13979,6 +14187,11 @@ export interface WebSource { * Expression with resultType integer). */ maxConcurrentConnections?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; } /** @@ -14046,6 +14259,11 @@ export interface OracleSource { * The settings that will be leveraged for Oracle source partitioning. */ partitionSettings?: OraclePartitionSettings; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; } /** @@ -14085,6 +14303,11 @@ export interface AzureDataExplorerSource { * ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])).. */ queryTimeout?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; } /** @@ -14168,6 +14391,11 @@ export interface FileSystemSource { * (or Expression with resultType boolean). */ recursive?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; } /** @@ -14222,6 +14450,11 @@ export interface RestSource { * The time to await before sending next page request. */ requestInterval?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; } /** @@ -14255,6 +14488,11 @@ export interface SalesforceServiceCloudSource { * 'QueryAll' */ readBehavior?: SalesforceSourceReadBehavior; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; } /** @@ -14283,6 +14521,11 @@ export interface ODataSource { * OData query. For example, "$top=1". Type: string (or Expression with resultType string). */ query?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; } /** @@ -14311,6 +14554,11 @@ export interface MicrosoftAccessSource { * Database query. Type: string (or Expression with resultType string). */ query?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; } /** @@ -14339,6 +14587,11 @@ export interface RelationalSource { * Database query. Type: string (or Expression with resultType string). */ query?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; } /** @@ -14368,6 +14621,11 @@ export interface CommonDataServiceForAppsSource { * Apps (online & on-premises). Type: string (or Expression with resultType string). */ query?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; } /** @@ -14397,6 +14655,11 @@ export interface DynamicsCrmSource { * on-premises). Type: string (or Expression with resultType string). */ query?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; } /** @@ -14426,6 +14689,11 @@ export interface DynamicsSource { * on-premises). Type: string (or Expression with resultType string). */ query?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; } /** @@ -14462,6 +14730,11 @@ export interface CosmosDbSqlApiSource { * Preferred regions. Type: array of strings (or Expression with resultType array of strings). */ preferredRegions?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; } /** @@ -14499,6 +14772,11 @@ export interface DocumentDbCollectionSource { * ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). */ queryTimeout?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; } /** @@ -14590,6 +14868,11 @@ export interface TabularSource { * ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). */ queryTimeout?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; } /** @@ -14619,6 +14902,11 @@ export interface AmazonRedshiftSource { * ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). */ queryTimeout?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; /** * Database query. Type: string (or Expression with resultType string). */ @@ -14658,6 +14946,11 @@ export interface GoogleAdWordsSource { * ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). */ queryTimeout?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; /** * A query to retrieve data from source. Type: string (or Expression with resultType string). */ @@ -14691,6 +14984,11 @@ export interface OracleServiceCloudSource { * ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). */ queryTimeout?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; /** * A query to retrieve data from source. Type: string (or Expression with resultType string). */ @@ -14724,6 +15022,11 @@ export interface DynamicsAXSource { * ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). */ queryTimeout?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; /** * A query to retrieve data from source. Type: string (or Expression with resultType string). */ @@ -14757,6 +15060,11 @@ export interface ResponsysSource { * ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). */ queryTimeout?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; /** * A query to retrieve data from source. Type: string (or Expression with resultType string). */ @@ -14790,6 +15098,11 @@ export interface SalesforceMarketingCloudSource { * ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). */ queryTimeout?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; /** * A query to retrieve data from source. Type: string (or Expression with resultType string). */ @@ -14823,6 +15136,11 @@ export interface VerticaSource { * ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). */ queryTimeout?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; /** * A query to retrieve data from source. Type: string (or Expression with resultType string). */ @@ -14877,6 +15195,11 @@ export interface NetezzaSource { * ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). */ queryTimeout?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; /** * A query to retrieve data from source. Type: string (or Expression with resultType string). */ @@ -14919,6 +15242,11 @@ export interface ZohoSource { * ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). */ queryTimeout?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; /** * A query to retrieve data from source. Type: string (or Expression with resultType string). */ @@ -14952,6 +15280,11 @@ export interface XeroSource { * ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). */ queryTimeout?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; /** * A query to retrieve data from source. Type: string (or Expression with resultType string). */ @@ -14985,6 +15318,11 @@ export interface SquareSource { * ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). */ queryTimeout?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; /** * A query to retrieve data from source. Type: string (or Expression with resultType string). */ @@ -15018,6 +15356,11 @@ export interface SparkSource { * ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). */ queryTimeout?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; /** * A query to retrieve data from source. Type: string (or Expression with resultType string). */ @@ -15051,6 +15394,11 @@ export interface ShopifySource { * ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). */ queryTimeout?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; /** * A query to retrieve data from source. Type: string (or Expression with resultType string). */ @@ -15084,6 +15432,11 @@ export interface ServiceNowSource { * ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). */ queryTimeout?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; /** * A query to retrieve data from source. Type: string (or Expression with resultType string). */ @@ -15117,6 +15470,11 @@ export interface QuickBooksSource { * ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). */ queryTimeout?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; /** * A query to retrieve data from source. Type: string (or Expression with resultType string). */ @@ -15150,6 +15508,11 @@ export interface PrestoSource { * ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). */ queryTimeout?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; /** * A query to retrieve data from source. Type: string (or Expression with resultType string). */ @@ -15183,6 +15546,11 @@ export interface PhoenixSource { * ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). */ queryTimeout?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; /** * A query to retrieve data from source. Type: string (or Expression with resultType string). */ @@ -15216,6 +15584,11 @@ export interface PaypalSource { * ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). */ queryTimeout?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; /** * A query to retrieve data from source. Type: string (or Expression with resultType string). */ @@ -15249,6 +15622,11 @@ export interface MarketoSource { * ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). */ queryTimeout?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; /** * A query to retrieve data from source. Type: string (or Expression with resultType string). */ @@ -15282,6 +15660,11 @@ export interface AzureMariaDBSource { * ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). */ queryTimeout?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; /** * A query to retrieve data from source. Type: string (or Expression with resultType string). */ @@ -15315,6 +15698,11 @@ export interface MariaDBSource { * ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). */ queryTimeout?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; /** * A query to retrieve data from source. Type: string (or Expression with resultType string). */ @@ -15348,6 +15736,11 @@ export interface MagentoSource { * ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). */ queryTimeout?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; /** * A query to retrieve data from source. Type: string (or Expression with resultType string). */ @@ -15381,6 +15774,11 @@ export interface JiraSource { * ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). */ queryTimeout?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; /** * A query to retrieve data from source. Type: string (or Expression with resultType string). */ @@ -15414,6 +15812,11 @@ export interface ImpalaSource { * ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). */ queryTimeout?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; /** * A query to retrieve data from source. Type: string (or Expression with resultType string). */ @@ -15447,6 +15850,11 @@ export interface HubspotSource { * ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). */ queryTimeout?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; /** * A query to retrieve data from source. Type: string (or Expression with resultType string). */ @@ -15480,6 +15888,11 @@ export interface HiveSource { * ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). */ queryTimeout?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; /** * A query to retrieve data from source. Type: string (or Expression with resultType string). */ @@ -15513,6 +15926,11 @@ export interface HBaseSource { * ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). */ queryTimeout?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; /** * A query to retrieve data from source. Type: string (or Expression with resultType string). */ @@ -15546,6 +15964,11 @@ export interface GreenplumSource { * ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). */ queryTimeout?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; /** * A query to retrieve data from source. Type: string (or Expression with resultType string). */ @@ -15579,6 +16002,11 @@ export interface GoogleBigQuerySource { * ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). */ queryTimeout?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; /** * A query to retrieve data from source. Type: string (or Expression with resultType string). */ @@ -15612,6 +16040,11 @@ export interface EloquaSource { * ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). */ queryTimeout?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; /** * A query to retrieve data from source. Type: string (or Expression with resultType string). */ @@ -15645,6 +16078,11 @@ export interface DrillSource { * ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). */ queryTimeout?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; /** * A query to retrieve data from source. Type: string (or Expression with resultType string). */ @@ -15678,6 +16116,11 @@ export interface CouchbaseSource { * ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). */ queryTimeout?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; /** * A query to retrieve data from source. Type: string (or Expression with resultType string). */ @@ -15711,6 +16154,11 @@ export interface ConcurSource { * ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). */ queryTimeout?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; /** * A query to retrieve data from source. Type: string (or Expression with resultType string). */ @@ -15744,6 +16192,11 @@ export interface AzurePostgreSqlSource { * ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). */ queryTimeout?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; /** * A query to retrieve data from source. Type: string (or Expression with resultType string). */ @@ -15777,6 +16230,11 @@ export interface AmazonMWSSource { * ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). */ queryTimeout?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; /** * A query to retrieve data from source. Type: string (or Expression with resultType string). */ @@ -15810,6 +16268,11 @@ export interface CassandraSource { * ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). */ queryTimeout?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; /** * Database query. Should be a SQL-92 query expression or Cassandra Query Language (CQL) command. * Type: string (or Expression with resultType string). @@ -15874,6 +16337,11 @@ export interface TeradataSource { * ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). */ queryTimeout?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; /** * Teradata query. Type: string (or Expression with resultType string). */ @@ -15916,6 +16384,11 @@ export interface AzureMySqlSource { * ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). */ queryTimeout?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; /** * Database query. Type: string (or Expression with resultType string). */ @@ -15949,6 +16422,11 @@ export interface SqlDWSource { * ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). */ queryTimeout?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; /** * SQL Data Warehouse reader query. Type: string (or Expression with resultType string). */ @@ -16008,6 +16486,11 @@ export interface SqlMISource { * ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). */ queryTimeout?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; /** * SQL reader query. Type: string (or Expression with resultType string). */ @@ -16055,6 +16538,11 @@ export interface AzureSqlSource { * ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). */ queryTimeout?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; /** * SQL reader query. Type: string (or Expression with resultType string). */ @@ -16102,6 +16590,11 @@ export interface SqlServerSource { * ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). */ queryTimeout?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; /** * SQL reader query. Type: string (or Expression with resultType string). */ @@ -16149,6 +16642,11 @@ export interface SqlSource { * ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). */ queryTimeout?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; /** * SQL reader query. Type: string (or Expression with resultType string). */ @@ -16163,6 +16661,12 @@ export interface SqlSource { * type: "int"}}". */ storedProcedureParameters?: { [propertyName: string]: StoredProcedureParameter }; + /** + * Specifies the transaction locking behavior for the SQL source. Allowed values: + * ReadCommitted/ReadUncommitted/RepeatableRead/Serializable/Snapshot. The default value is + * ReadCommitted. Type: string (or Expression with resultType string). + */ + isolationLevel?: any; } /** @@ -16218,6 +16722,11 @@ export interface SapTableSource { * ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). */ queryTimeout?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; /** * The number of rows to be retrieved. Type: integer(or Expression with resultType integer). */ @@ -16286,6 +16795,11 @@ export interface SapOpenHubSource { * ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). */ queryTimeout?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; /** * Whether to exclude the records of the last request. The default value is true. Type: boolean * (or Expression with resultType boolean). @@ -16337,6 +16851,11 @@ export interface SapHanaSource { * ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). */ queryTimeout?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; /** * SAP HANA Sql query. Type: string (or Expression with resultType string). */ @@ -16384,6 +16903,11 @@ export interface SapEccSource { * ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). */ queryTimeout?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; /** * SAP ECC OData query. For example, "$top=1". Type: string (or Expression with resultType * string). @@ -16418,6 +16942,11 @@ export interface SapCloudForCustomerSource { * ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). */ queryTimeout?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; /** * SAP Cloud for Customer OData query. For example, "$top=1". Type: string (or Expression with * resultType string). @@ -16452,6 +16981,11 @@ export interface SalesforceSource { * ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). */ queryTimeout?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; /** * Database query. Type: string (or Expression with resultType string). */ @@ -16490,6 +17024,11 @@ export interface SapBwSource { * ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). */ queryTimeout?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; /** * MDX query. Type: string (or Expression with resultType string). */ @@ -16523,6 +17062,11 @@ export interface SybaseSource { * ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). */ queryTimeout?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; /** * Database query. Type: string (or Expression with resultType string). */ @@ -16556,6 +17100,11 @@ export interface PostgreSqlSource { * ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). */ queryTimeout?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; /** * Database query. Type: string (or Expression with resultType string). */ @@ -16589,6 +17138,11 @@ export interface MySqlSource { * ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). */ queryTimeout?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; /** * Database query. Type: string (or Expression with resultType string). */ @@ -16622,6 +17176,11 @@ export interface OdbcSource { * ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). */ queryTimeout?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; /** * Database query. Type: string (or Expression with resultType string). */ @@ -16655,6 +17214,11 @@ export interface Db2Source { * ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). */ queryTimeout?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; /** * Database query. Type: string (or Expression with resultType string). */ @@ -16688,6 +17252,11 @@ export interface InformixSource { * ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). */ queryTimeout?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; /** * Database query. Type: string (or Expression with resultType string). */ @@ -16721,6 +17290,11 @@ export interface AzureTableSource { * ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). */ queryTimeout?: any; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; /** * Azure Table source query. Type: string (or Expression with resultType string). */ @@ -16782,6 +17356,11 @@ export interface HdfsReadSettings { * HDFS wildcardFileName. Type: string (or Expression with resultType string). */ wildcardFileName?: any; + /** + * Point to a text file that lists each file (relative path to the path configured in the + * dataset) that you want to copy. Type: string (or Expression with resultType string). + */ + fileListPath?: any; /** * Indicates whether to enable partition discovery. */ @@ -16860,6 +17439,11 @@ export interface SftpReadSettings { * Sftp wildcardFileName. Type: string (or Expression with resultType string). */ wildcardFileName?: any; + /** + * Point to a text file that lists each file (relative path to the path configured in the + * dataset) that you want to copy. Type: string (or Expression with resultType string). + */ + fileListPath?: any; /** * The start of file's modified datetime. Type: string (or Expression with resultType string). */ @@ -16896,6 +17480,11 @@ export interface FtpReadSettings { * Ftp wildcardFileName. Type: string (or Expression with resultType string). */ wildcardFileName?: any; + /** + * Point to a text file that lists each file (relative path to the path configured in the + * dataset) that you want to copy. Type: string (or Expression with resultType string). + */ + fileListPath?: any; /** * Specify whether to use binary transfer mode for FTP stores. */ @@ -16933,6 +17522,11 @@ export interface GoogleCloudStorageReadSettings { * resultType string). */ prefix?: any; + /** + * Point to a text file that lists each file (relative path to the path configured in the + * dataset) that you want to copy. Type: string (or Expression with resultType string). + */ + fileListPath?: any; /** * Indicates whether to enable partition discovery. */ @@ -16973,6 +17567,11 @@ export interface AzureFileStorageReadSettings { * Azure File Storage wildcardFileName. Type: string (or Expression with resultType string). */ wildcardFileName?: any; + /** + * Point to a text file that lists each file (relative path to the path configured in the + * dataset) that you want to copy. Type: string (or Expression with resultType string). + */ + fileListPath?: any; /** * Indicates whether to enable partition discovery. */ @@ -17013,6 +17612,11 @@ export interface FileServerReadSettings { * FileServer wildcardFileName. Type: string (or Expression with resultType string). */ wildcardFileName?: any; + /** + * Point to a text file that lists each file (relative path to the path configured in the + * dataset) that you want to copy. Type: string (or Expression with resultType string). + */ + fileListPath?: any; /** * Indicates whether to enable partition discovery. */ @@ -17057,6 +17661,11 @@ export interface AmazonS3ReadSettings { * The prefix filter for the S3 object name. Type: string (or Expression with resultType string). */ prefix?: any; + /** + * Point to a text file that lists each file (relative path to the path configured in the + * dataset) that you want to copy. Type: string (or Expression with resultType string). + */ + fileListPath?: any; /** * Indicates whether to enable partition discovery. */ @@ -17097,6 +17706,11 @@ export interface AzureDataLakeStoreReadSettings { * ADLS wildcardFileName. Type: string (or Expression with resultType string). */ wildcardFileName?: any; + /** + * Point to a text file that lists each file (relative path to the path configured in the + * dataset) that you want to copy. Type: string (or Expression with resultType string). + */ + fileListPath?: any; /** * Indicates whether to enable partition discovery. */ @@ -17137,6 +17751,11 @@ export interface AzureBlobFSReadSettings { * Azure blobFS wildcardFileName. Type: string (or Expression with resultType string). */ wildcardFileName?: any; + /** + * Point to a text file that lists each file (relative path to the path configured in the + * dataset) that you want to copy. Type: string (or Expression with resultType string). + */ + fileListPath?: any; /** * Indicates whether to enable partition discovery. */ @@ -17182,6 +17801,11 @@ export interface AzureBlobStorageReadSettings { * string). */ prefix?: any; + /** + * Point to a text file that lists each file (relative path to the path configured in the + * dataset) that you want to copy. Type: string (or Expression with resultType string). + */ + fileListPath?: any; /** * Indicates whether to enable partition discovery. */ @@ -17250,6 +17874,11 @@ export interface OrcSource { * ORC store settings. */ storeSettings?: StoreReadSettingsUnion; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; } /** @@ -17278,6 +17907,11 @@ export interface JsonSource { * Json store settings. */ storeSettings?: StoreReadSettingsUnion; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; } /** @@ -17343,7 +17977,12 @@ export interface DelimitedTextSource { /** * DelimitedText format settings. */ - formatSettings?: DelimitedTextReadSettings; + formatSettings?: DelimitedTextReadSettings; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; } /** @@ -17372,6 +18011,11 @@ export interface ParquetSource { * Parquet store settings. */ storeSettings?: StoreReadSettingsUnion; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; } /** @@ -17400,6 +18044,11 @@ export interface AvroSource { * Avro store settings. */ storeSettings?: StoreReadSettingsUnion; + /** + * Specifies the additional columns to be added to source data. Type: array of objects (or + * Expression with resultType array of objects). + */ + additionalColumns?: AdditionalColumns[]; } /** @@ -18261,6 +18910,22 @@ export interface HDInsightHiveActivity { queryTimeout?: number; } +/** + * Skip error file. + */ +export interface SkipErrorFile { + /** + * Skip if file is deleted by other client during copy. Default is true. Type: boolean (or + * Expression with resultType boolean). + */ + fileMissing?: any; + /** + * Skip if source/sink file changed by other concurrent write. Default is false. Type: boolean + * (or Expression with resultType boolean). + */ + dataInconsistency?: any; +} + /** * Redirect incompatible row settings */ @@ -18309,7 +18974,7 @@ export interface StagingSettings { /** * Contains the possible cases for CopySink. */ -export type CopySinkUnion = CopySink | CosmosDbMongoDbApiSink | SalesforceServiceCloudSink | SalesforceSink | AzureDataExplorerSink | CommonDataServiceForAppsSink | DynamicsCrmSink | DynamicsSink | MicrosoftAccessSink | InformixSink | OdbcSink | AzureSearchIndexSink | AzureBlobFSSink | AzureDataLakeStoreSink | OracleSink | SqlDWSink | SqlMISink | AzureSqlSink | SqlServerSink | SqlSink | CosmosDbSqlApiSink | DocumentDbCollectionSink | FileSystemSink | BlobSink | BinarySink | ParquetSink | AvroSink | AzureTableSink | AzureQueueSink | SapCloudForCustomerSink | AzureMySqlSink | AzurePostgreSqlSink | OrcSink | JsonSink | DelimitedTextSink; +export type CopySinkUnion = CopySink | CosmosDbMongoDbApiSink | SalesforceServiceCloudSink | SalesforceSink | AzureDataExplorerSink | CommonDataServiceForAppsSink | DynamicsCrmSink | DynamicsSink | MicrosoftAccessSink | InformixSink | OdbcSink | AzureSearchIndexSink | AzureBlobFSSink | AzureDataLakeStoreSink | OracleSink | SnowflakeSink | SqlDWSink | SqlMISink | AzureSqlSink | SqlServerSink | SqlSink | CosmosDbSqlApiSink | DocumentDbCollectionSink | FileSystemSink | BlobSink | BinarySink | ParquetSink | AvroSink | AzureTableSink | AzureQueueSink | SapCloudForCustomerSink | AzureMySqlSink | AzurePostgreSqlSink | OrcSink | JsonSink | DelimitedTextSink; /** * A copy activity sink. @@ -18934,6 +19599,88 @@ export interface OracleSink { preCopyScript?: any; } +/** + * Contains the possible cases for ImportSettings. + */ +export type ImportSettingsUnion = ImportSettings | SnowflakeImportCopyCommand; + +/** + * Import command settings. + */ +export interface ImportSettings { + /** + * Polymorphic Discriminator + */ + type: "ImportSettings"; + /** + * Describes unknown properties. The value of an unknown property can be of "any" type. + */ + [property: string]: any; +} + +/** + * Snowflake import command settings. + */ +export interface SnowflakeImportCopyCommand { + /** + * Polymorphic Discriminator + */ + type: "SnowflakeImportCopyCommand"; + /** + * Additional copy options directly passed to snowflake Copy Command. Type: key value pairs + * (value should be string type) (or Expression with resultType object). Example: + * "additionalCopyOptions": { "DATE_FORMAT": "MM/DD/YYYY", "TIME_FORMAT": "'HH24:MI:SS.FF'" } + */ + additionalCopyOptions?: { [propertyName: string]: any }; + /** + * Additional format options directly passed to snowflake Copy Command. Type: key value pairs + * (value should be string type) (or Expression with resultType object). Example: + * "additionalFormatOptions": { "FORCE": "TRUE", "LOAD_UNCERTAIN_FILES": "'FALSE'" } + */ + additionalFormatOptions?: { [propertyName: string]: any }; +} + +/** + * A copy activity snowflake sink. + */ +export interface SnowflakeSink { + /** + * Polymorphic Discriminator + */ + type: "SnowflakeSink"; + /** + * Write batch size. Type: integer (or Expression with resultType integer), minimum: 0. + */ + writeBatchSize?: any; + /** + * Write batch timeout. Type: string (or Expression with resultType string), pattern: + * ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + */ + writeBatchTimeout?: any; + /** + * Sink retry count. Type: integer (or Expression with resultType integer). + */ + sinkRetryCount?: any; + /** + * Sink retry wait. Type: string (or Expression with resultType string), pattern: + * ((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + */ + sinkRetryWait?: any; + /** + * The maximum concurrent connection count for the sink data store. Type: integer (or Expression + * with resultType integer). + */ + maxConcurrentConnections?: any; + /** + * SQL pre-copy script. Type: string (or Expression with resultType string). + */ + preCopyScript?: any; + /** + * Snowflake import settings. + */ + importSettings?: SnowflakeImportCopyCommand; +} + /** * Default value. */ @@ -19523,6 +20270,12 @@ export interface AzureDataLakeStoreWriteSettings { * The type of copy behavior for copy sink. */ copyBehavior?: any; + /** + * Specifies the expiry time of the written files. The time is applied to the UTC time zone in + * the format of "2018-12-01T05:00:00Z". Default value is NULL. Type: integer (or Expression with + * resultType integer). + */ + expiryDateTime?: any; } /** @@ -19595,6 +20348,11 @@ export interface SftpWriteSettings { * hour). Type: string (or Expression with resultType string). */ operationTimeout?: any; + /** + * Upload to temporary file(s) and rename. Disable this option if your SFTP server doesn't + * support rename operation. Type: boolean (or Expression with resultType boolean). + */ + useTempFileRename?: any; } /** @@ -20172,6 +20930,10 @@ export interface CopyActivity { * Redirect incompatible row settings when EnableSkipIncompatibleRow is true. */ redirectIncompatibleRowSettings?: RedirectIncompatibleRowSettings; + /** + * Log storage settings customer need to provide when enabling session log. + */ + logStorageSettings?: LogStorageSettings; /** * Preserve Rules. */ @@ -20180,6 +20942,15 @@ export interface CopyActivity { * Preserve rules. */ preserve?: any[]; + /** + * Whether to enable Data Consistency validation. Type: boolean (or Expression with resultType + * boolean). + */ + validateDataConsistency?: any; + /** + * Specify the fault tolerance for data consistency. + */ + skipErrorFile?: SkipErrorFile; /** * List of inputs for the activity. */ @@ -21975,6 +22746,11 @@ export interface PipelinesCreateRunOptionalParams extends msRest.RequestOptionsB * will run. */ startActivityName?: string; + /** + * In recovery mode, if set to true, the rerun will start from failed activities. The property + * will be used only if startActivityName is not specified. + */ + startFromFailure?: boolean; /** * Parameters of the pipeline run. These parameters will be used only if the runId is not * specified. @@ -22147,20 +22923,6 @@ export interface TriggerListResponse extends Array { nextLink?: string; } -/** - * @interface - * A list of rerun triggers. - * @extends Array - */ -export interface RerunTriggerListResponse extends Array { - /** - * The continuation token for getting the next page of results, if any remaining results exist, - * null otherwise. - * **NOTE: This property will not be serialized. It can only be populated by the server.** - */ - readonly nextLink?: string; -} - /** * @interface * A list of data flow resources. @@ -22294,48 +23056,6 @@ export type TriggerRunStatus = 'Succeeded' | 'Failed' | 'Inprogress'; */ export type DataFlowDebugCommandType = 'executePreviewQuery' | 'executeStatisticsQuery' | 'executeExpressionQuery'; -/** - * Defines values for TumblingWindowFrequency. - * Possible values include: 'Minute', 'Hour' - * @readonly - * @enum {string} - */ -export type TumblingWindowFrequency = 'Minute' | 'Hour'; - -/** - * Defines values for BlobEventTypes. - * Possible values include: 'Microsoft.Storage.BlobCreated', 'Microsoft.Storage.BlobDeleted' - * @readonly - * @enum {string} - */ -export type BlobEventTypes = 'Microsoft.Storage.BlobCreated' | 'Microsoft.Storage.BlobDeleted'; - -/** - * Defines values for DayOfWeek. - * Possible values include: 'Sunday', 'Monday', 'Tuesday', 'Wednesday', 'Thursday', 'Friday', - * 'Saturday' - * @readonly - * @enum {string} - */ -export type DayOfWeek = 'Sunday' | 'Monday' | 'Tuesday' | 'Wednesday' | 'Thursday' | 'Friday' | 'Saturday'; - -/** - * Defines values for DaysOfWeek. - * Possible values include: 'Sunday', 'Monday', 'Tuesday', 'Wednesday', 'Thursday', 'Friday', - * 'Saturday' - * @readonly - * @enum {string} - */ -export type DaysOfWeek = 'Sunday' | 'Monday' | 'Tuesday' | 'Wednesday' | 'Thursday' | 'Friday' | 'Saturday'; - -/** - * Defines values for RecurrenceFrequency. - * Possible values include: 'NotSpecified', 'Minute', 'Hour', 'Day', 'Week', 'Month', 'Year' - * @readonly - * @enum {string} - */ -export type RecurrenceFrequency = 'NotSpecified' | 'Minute' | 'Hour' | 'Day' | 'Week' | 'Month' | 'Year'; - /** * Defines values for GoogleAdWordsAuthenticationType. * Possible values include: 'ServiceAuthentication', 'UserAuthentication' @@ -22563,6 +23283,48 @@ export type OrcCompressionCodec = 'none' | 'zlib' | 'snappy'; */ export type AvroCompressionCodec = 'none' | 'deflate' | 'snappy' | 'xz' | 'bzip2'; +/** + * Defines values for TumblingWindowFrequency. + * Possible values include: 'Minute', 'Hour' + * @readonly + * @enum {string} + */ +export type TumblingWindowFrequency = 'Minute' | 'Hour'; + +/** + * Defines values for BlobEventTypes. + * Possible values include: 'Microsoft.Storage.BlobCreated', 'Microsoft.Storage.BlobDeleted' + * @readonly + * @enum {string} + */ +export type BlobEventTypes = 'Microsoft.Storage.BlobCreated' | 'Microsoft.Storage.BlobDeleted'; + +/** + * Defines values for DayOfWeek. + * Possible values include: 'Sunday', 'Monday', 'Tuesday', 'Wednesday', 'Thursday', 'Friday', + * 'Saturday' + * @readonly + * @enum {string} + */ +export type DayOfWeek = 'Sunday' | 'Monday' | 'Tuesday' | 'Wednesday' | 'Thursday' | 'Friday' | 'Saturday'; + +/** + * Defines values for DaysOfWeek. + * Possible values include: 'Sunday', 'Monday', 'Tuesday', 'Wednesday', 'Thursday', 'Friday', + * 'Saturday' + * @readonly + * @enum {string} + */ +export type DaysOfWeek = 'Sunday' | 'Monday' | 'Tuesday' | 'Wednesday' | 'Thursday' | 'Friday' | 'Saturday'; + +/** + * Defines values for RecurrenceFrequency. + * Possible values include: 'NotSpecified', 'Minute', 'Hour', 'Day', 'Week', 'Month', 'Year' + * @readonly + * @enum {string} + */ +export type RecurrenceFrequency = 'NotSpecified' | 'Minute' | 'Hour' | 'Day' | 'Week' | 'Month' | 'Year'; + /** * Defines values for DataFlowComputeType. * Possible values include: 'General', 'MemoryOptimized', 'ComputeOptimized' @@ -23806,6 +24568,26 @@ export type TriggersListByFactoryResponse = TriggerListResponse & { }; }; +/** + * Contains response data for the queryByFactory operation. + */ +export type TriggersQueryByFactoryResponse = TriggerQueryResponse & { + /** + * The underlying HTTP response. + */ + _response: msRest.HttpResponse & { + /** + * The response body as text (string format) + */ + bodyAsText: string; + + /** + * The response body as parsed JSON or XML + */ + parsedBody: TriggerQueryResponse; + }; +}; + /** * Contains response data for the createOrUpdate operation. */ @@ -23986,66 +24768,6 @@ export type TriggerRunsQueryByFactoryResponse = TriggerRunsQueryResponse & { }; }; -/** - * Contains response data for the create operation. - */ -export type RerunTriggersCreateResponse = TriggerResource & { - /** - * The underlying HTTP response. - */ - _response: msRest.HttpResponse & { - /** - * The response body as text (string format) - */ - bodyAsText: string; - - /** - * The response body as parsed JSON or XML - */ - parsedBody: TriggerResource; - }; -}; - -/** - * Contains response data for the listByTrigger operation. - */ -export type RerunTriggersListByTriggerResponse = RerunTriggerListResponse & { - /** - * The underlying HTTP response. - */ - _response: msRest.HttpResponse & { - /** - * The response body as text (string format) - */ - bodyAsText: string; - - /** - * The response body as parsed JSON or XML - */ - parsedBody: RerunTriggerListResponse; - }; -}; - -/** - * Contains response data for the listByTriggerNext operation. - */ -export type RerunTriggersListByTriggerNextResponse = RerunTriggerListResponse & { - /** - * The underlying HTTP response. - */ - _response: msRest.HttpResponse & { - /** - * The response body as text (string format) - */ - bodyAsText: string; - - /** - * The response body as parsed JSON or XML - */ - parsedBody: RerunTriggerListResponse; - }; -}; - /** * Contains response data for the createOrUpdate operation. */ diff --git a/sdk/datafactory/arm-datafactory/src/models/integrationRuntimesMappers.ts b/sdk/datafactory/arm-datafactory/src/models/integrationRuntimesMappers.ts index 467b87ecd36d..76dc0eb02340 100644 --- a/sdk/datafactory/arm-datafactory/src/models/integrationRuntimesMappers.ts +++ b/sdk/datafactory/arm-datafactory/src/models/integrationRuntimesMappers.ts @@ -11,6 +11,7 @@ export { Activity, ActivityDependency, ActivityPolicy, + AdditionalColumns, AmazonMWSLinkedService, AmazonMWSObjectDataset, AmazonMWSSource, @@ -200,6 +201,7 @@ export { ExecutePipelineActivity, ExecuteSSISPackageActivity, ExecutionActivity, + ExportSettings, Expression, Factory, FactoryGitHubConfiguration, @@ -262,6 +264,7 @@ export { ImpalaLinkedService, ImpalaObjectDataset, ImpalaSource, + ImportSettings, InformixLinkedService, InformixSink, InformixSource, @@ -388,7 +391,6 @@ export { RedshiftUnloadSettings, RelationalSource, RelationalTableDataset, - RerunTriggerResource, RerunTumblingWindowTrigger, Resource, ResponsysLinkedService, @@ -450,6 +452,13 @@ export { ShopifyLinkedService, ShopifyObjectDataset, ShopifySource, + SkipErrorFile, + SnowflakeDataset, + SnowflakeExportCopyCommand, + SnowflakeImportCopyCommand, + SnowflakeLinkedService, + SnowflakeSink, + SnowflakeSource, SparkLinkedService, SparkObjectDataset, SparkSource, diff --git a/sdk/datafactory/arm-datafactory/src/models/linkedServicesMappers.ts b/sdk/datafactory/arm-datafactory/src/models/linkedServicesMappers.ts index f076eff667ce..bf82f7d72075 100644 --- a/sdk/datafactory/arm-datafactory/src/models/linkedServicesMappers.ts +++ b/sdk/datafactory/arm-datafactory/src/models/linkedServicesMappers.ts @@ -11,6 +11,7 @@ export { Activity, ActivityDependency, ActivityPolicy, + AdditionalColumns, AmazonMWSLinkedService, AmazonMWSObjectDataset, AmazonMWSSource, @@ -199,6 +200,7 @@ export { ExecutePipelineActivity, ExecuteSSISPackageActivity, ExecutionActivity, + ExportSettings, Expression, Factory, FactoryGitHubConfiguration, @@ -261,6 +263,7 @@ export { ImpalaLinkedService, ImpalaObjectDataset, ImpalaSource, + ImportSettings, InformixLinkedService, InformixSink, InformixSource, @@ -374,7 +377,6 @@ export { RedshiftUnloadSettings, RelationalSource, RelationalTableDataset, - RerunTriggerResource, RerunTumblingWindowTrigger, Resource, ResponsysLinkedService, @@ -434,6 +436,13 @@ export { ShopifyLinkedService, ShopifyObjectDataset, ShopifySource, + SkipErrorFile, + SnowflakeDataset, + SnowflakeExportCopyCommand, + SnowflakeImportCopyCommand, + SnowflakeLinkedService, + SnowflakeSink, + SnowflakeSource, SparkLinkedService, SparkObjectDataset, SparkSource, diff --git a/sdk/datafactory/arm-datafactory/src/models/mappers.ts b/sdk/datafactory/arm-datafactory/src/models/mappers.ts index e59bd06d97f9..ec2b7962bdd8 100644 --- a/sdk/datafactory/arm-datafactory/src/models/mappers.ts +++ b/sdk/datafactory/arm-datafactory/src/models/mappers.ts @@ -1289,6 +1289,35 @@ export const TriggerResource: msRest.CompositeMapper = { } }; +export const TriggerQueryResponse: msRest.CompositeMapper = { + serializedName: "TriggerQueryResponse", + type: { + name: "Composite", + className: "TriggerQueryResponse", + modelProperties: { + value: { + required: true, + serializedName: "value", + type: { + name: "Sequence", + element: { + type: { + name: "Composite", + className: "TriggerResource" + } + } + } + }, + continuationToken: { + serializedName: "continuationToken", + type: { + name: "String" + } + } + } + } +}; + export const CreateRunResponse: msRest.CompositeMapper = { serializedName: "CreateRunResponse", type: { @@ -1688,6 +1717,28 @@ export const LinkedServiceDebugResource: msRest.CompositeMapper = { } }; +export const TriggerFilterParameters: msRest.CompositeMapper = { + serializedName: "TriggerFilterParameters", + type: { + name: "Composite", + className: "TriggerFilterParameters", + modelProperties: { + continuationToken: { + serializedName: "continuationToken", + type: { + name: "String" + } + }, + parentTriggerName: { + serializedName: "parentTriggerName", + type: { + name: "String" + } + } + } + } +}; + export const RunQueryFilter: msRest.CompositeMapper = { serializedName: "RunQueryFilter", type: { @@ -2263,106 +2314,6 @@ export const TriggerRunsQueryResponse: msRest.CompositeMapper = { } }; -export const RerunTumblingWindowTriggerActionParameters: msRest.CompositeMapper = { - serializedName: "RerunTumblingWindowTriggerActionParameters", - type: { - name: "Composite", - className: "RerunTumblingWindowTriggerActionParameters", - modelProperties: { - startTime: { - required: true, - serializedName: "startTime", - type: { - name: "DateTime" - } - }, - endTime: { - required: true, - serializedName: "endTime", - type: { - name: "DateTime" - } - }, - maxConcurrency: { - required: true, - serializedName: "maxConcurrency", - constraints: { - InclusiveMaximum: 50, - InclusiveMinimum: 1 - }, - type: { - name: "Number" - } - } - } - } -}; - -export const RerunTumblingWindowTrigger: msRest.CompositeMapper = { - serializedName: "RerunTumblingWindowTrigger", - type: { - name: "Composite", - polymorphicDiscriminator: Trigger.type.polymorphicDiscriminator, - uberParent: "Trigger", - className: "RerunTumblingWindowTrigger", - modelProperties: { - ...Trigger.type.modelProperties, - parentTrigger: { - serializedName: "typeProperties.parentTrigger", - type: { - name: "Object" - } - }, - requestedStartTime: { - required: true, - serializedName: "typeProperties.requestedStartTime", - type: { - name: "DateTime" - } - }, - requestedEndTime: { - required: true, - serializedName: "typeProperties.requestedEndTime", - type: { - name: "DateTime" - } - }, - maxConcurrency: { - required: true, - serializedName: "typeProperties.maxConcurrency", - constraints: { - InclusiveMaximum: 50, - InclusiveMinimum: 1 - }, - type: { - name: "Number" - } - } - }, - additionalProperties: Trigger.type.additionalProperties - } -}; - -export const RerunTriggerResource: msRest.CompositeMapper = { - serializedName: "RerunTriggerResource", - type: { - name: "Composite", - className: "RerunTriggerResource", - modelProperties: { - ...SubResource.type.modelProperties, - properties: { - required: true, - serializedName: "properties", - type: { - name: "Composite", - className: "RerunTumblingWindowTrigger", - additionalProperties: Trigger.type.additionalProperties - } - } - } - } -}; - export const OperationDisplay: msRest.CompositeMapper = { serializedName: "Operation_display", type: { @@ -3404,838 +3355,257 @@ export const MappingDataFlow: msRest.CompositeMapper = { } }; -export const ChainingTrigger: msRest.CompositeMapper = { - serializedName: "ChainingTrigger", +export const SnowflakeLinkedService: msRest.CompositeMapper = { + serializedName: "Snowflake", type: { name: "Composite", - polymorphicDiscriminator: Trigger.type.polymorphicDiscriminator, - uberParent: "Trigger", - className: "ChainingTrigger", + polymorphicDiscriminator: LinkedService.type.polymorphicDiscriminator, + uberParent: "LinkedService", + className: "SnowflakeLinkedService", modelProperties: { - ...Trigger.type.modelProperties, - pipelineProperty: { + ...LinkedService.type.modelProperties, + connectionString: { required: true, - serializedName: "pipeline", - defaultValue: {}, + serializedName: "typeProperties.connectionString", type: { - name: "Composite", - className: "TriggerPipelineReference" + name: "Object" } }, - dependsOn: { - required: true, - serializedName: "typeProperties.dependsOn", + password: { + serializedName: "typeProperties.password", type: { - name: "Sequence", - element: { - type: { - name: "Composite", - className: "PipelineReference" - } - } + name: "Composite", + className: "AzureKeyVaultSecretReference" } }, - runDimension: { - required: true, - serializedName: "typeProperties.runDimension", + encryptedCredential: { + serializedName: "typeProperties.encryptedCredential", type: { - name: "String" + name: "Object" } } }, - additionalProperties: Trigger.type.additionalProperties + additionalProperties: LinkedService.type.additionalProperties } }; -export const DependencyReference: msRest.CompositeMapper = { - serializedName: "DependencyReference", +export const AzureFunctionLinkedService: msRest.CompositeMapper = { + serializedName: "AzureFunction", type: { name: "Composite", - polymorphicDiscriminator: { - serializedName: "type", - clientName: "type" - }, - uberParent: "DependencyReference", - className: "DependencyReference", + polymorphicDiscriminator: LinkedService.type.polymorphicDiscriminator, + uberParent: "LinkedService", + className: "AzureFunctionLinkedService", modelProperties: { - type: { + ...LinkedService.type.modelProperties, + functionAppUrl: { required: true, - serializedName: "type", + serializedName: "typeProperties.functionAppUrl", type: { - name: "String" + name: "Object" } - } - } - } -}; - -export const SelfDependencyTumblingWindowTriggerReference: msRest.CompositeMapper = { - serializedName: "SelfDependencyTumblingWindowTriggerReference", - type: { - name: "Composite", - polymorphicDiscriminator: DependencyReference.type.polymorphicDiscriminator, - uberParent: "DependencyReference", - className: "SelfDependencyTumblingWindowTriggerReference", - modelProperties: { - ...DependencyReference.type.modelProperties, - offset: { - required: true, - serializedName: "offset", - constraints: { - MaxLength: 15, - MinLength: 8, - Pattern: /-((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9]))/ - }, + }, + functionKey: { + serializedName: "typeProperties.functionKey", type: { - name: "String" + name: "Composite", + className: "SecretBase" } }, - size: { - serializedName: "size", - constraints: { - MaxLength: 15, - MinLength: 8, - Pattern: /((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9]))/ - }, + encryptedCredential: { + serializedName: "typeProperties.encryptedCredential", type: { - name: "String" + name: "Object" } } - } + }, + additionalProperties: LinkedService.type.additionalProperties } }; -export const TriggerReference: msRest.CompositeMapper = { - serializedName: "TriggerReference", +export const AzureDataExplorerLinkedService: msRest.CompositeMapper = { + serializedName: "AzureDataExplorer", type: { name: "Composite", - className: "TriggerReference", + polymorphicDiscriminator: LinkedService.type.polymorphicDiscriminator, + uberParent: "LinkedService", + className: "AzureDataExplorerLinkedService", modelProperties: { - type: { + ...LinkedService.type.modelProperties, + endpoint: { required: true, - isConstant: true, - serializedName: "type", - defaultValue: 'TriggerReference', + serializedName: "typeProperties.endpoint", type: { - name: "String" + name: "Object" } }, - referenceName: { + servicePrincipalId: { required: true, - serializedName: "referenceName", + serializedName: "typeProperties.servicePrincipalId", type: { - name: "String" + name: "Object" } - } - } - } -}; - -export const TriggerDependencyReference: msRest.CompositeMapper = { - serializedName: "TriggerDependencyReference", - type: { - name: "Composite", - polymorphicDiscriminator: DependencyReference.type.polymorphicDiscriminator, - uberParent: "DependencyReference", - className: "TriggerDependencyReference", - modelProperties: { - ...DependencyReference.type.modelProperties, - referenceTrigger: { + }, + servicePrincipalKey: { required: true, - serializedName: "referenceTrigger", - defaultValue: {}, + serializedName: "typeProperties.servicePrincipalKey", type: { name: "Composite", - className: "TriggerReference" + className: "SecretBase" } - } - } - } -}; - -export const TumblingWindowTriggerDependencyReference: msRest.CompositeMapper = { - serializedName: "TumblingWindowTriggerDependencyReference", - type: { - name: "Composite", - polymorphicDiscriminator: DependencyReference.type.polymorphicDiscriminator, - uberParent: "DependencyReference", - className: "TumblingWindowTriggerDependencyReference", - modelProperties: { - ...TriggerDependencyReference.type.modelProperties, - offset: { - serializedName: "offset", - constraints: { - MaxLength: 15, - MinLength: 8, - Pattern: /-?((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9]))/ - }, + }, + database: { + required: true, + serializedName: "typeProperties.database", type: { - name: "String" + name: "Object" } }, - size: { - serializedName: "size", - constraints: { - MaxLength: 15, - MinLength: 8, - Pattern: /((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9]))/ - }, + tenant: { + required: true, + serializedName: "typeProperties.tenant", type: { - name: "String" + name: "Object" } } - } + }, + additionalProperties: LinkedService.type.additionalProperties } }; -export const RetryPolicy: msRest.CompositeMapper = { - serializedName: "RetryPolicy", +export const SapTableLinkedService: msRest.CompositeMapper = { + serializedName: "SapTable", type: { name: "Composite", - className: "RetryPolicy", + polymorphicDiscriminator: LinkedService.type.polymorphicDiscriminator, + uberParent: "LinkedService", + className: "SapTableLinkedService", modelProperties: { - count: { - serializedName: "count", + ...LinkedService.type.modelProperties, + server: { + serializedName: "typeProperties.server", type: { name: "Object" } }, - intervalInSeconds: { - serializedName: "intervalInSeconds", - constraints: { - InclusiveMaximum: 86400, - InclusiveMinimum: 30 - }, + systemNumber: { + serializedName: "typeProperties.systemNumber", type: { - name: "Number" + name: "Object" } - } - } - } -}; - -export const TumblingWindowTrigger: msRest.CompositeMapper = { - serializedName: "TumblingWindowTrigger", - type: { - name: "Composite", - polymorphicDiscriminator: Trigger.type.polymorphicDiscriminator, - uberParent: "Trigger", - className: "TumblingWindowTrigger", - modelProperties: { - ...Trigger.type.modelProperties, - pipelineProperty: { - required: true, - serializedName: "pipeline", - defaultValue: {}, + }, + clientId: { + serializedName: "typeProperties.clientId", + type: { + name: "Object" + } + }, + language: { + serializedName: "typeProperties.language", + type: { + name: "Object" + } + }, + systemId: { + serializedName: "typeProperties.systemId", + type: { + name: "Object" + } + }, + userName: { + serializedName: "typeProperties.userName", + type: { + name: "Object" + } + }, + password: { + serializedName: "typeProperties.password", type: { name: "Composite", - className: "TriggerPipelineReference" + className: "SecretBase" } }, - frequency: { - required: true, - serializedName: "typeProperties.frequency", + messageServer: { + serializedName: "typeProperties.messageServer", type: { - name: "String" + name: "Object" } }, - interval: { - required: true, - serializedName: "typeProperties.interval", + messageServerService: { + serializedName: "typeProperties.messageServerService", type: { - name: "Number" + name: "Object" } }, - startTime: { - required: true, - serializedName: "typeProperties.startTime", + sncMode: { + serializedName: "typeProperties.sncMode", type: { - name: "DateTime" + name: "Object" } }, - endTime: { - serializedName: "typeProperties.endTime", + sncMyName: { + serializedName: "typeProperties.sncMyName", type: { - name: "DateTime" + name: "Object" } }, - delay: { - serializedName: "typeProperties.delay", + sncPartnerName: { + serializedName: "typeProperties.sncPartnerName", type: { name: "Object" } }, - maxConcurrency: { - required: true, - serializedName: "typeProperties.maxConcurrency", - constraints: { - InclusiveMaximum: 50, - InclusiveMinimum: 1 - }, + sncLibraryPath: { + serializedName: "typeProperties.sncLibraryPath", type: { - name: "Number" + name: "Object" } }, - retryPolicy: { - serializedName: "typeProperties.retryPolicy", + sncQop: { + serializedName: "typeProperties.sncQop", type: { - name: "Composite", - className: "RetryPolicy" + name: "Object" } }, - dependsOn: { - serializedName: "typeProperties.dependsOn", + logonGroup: { + serializedName: "typeProperties.logonGroup", type: { - name: "Sequence", - element: { - type: { - name: "Composite", - className: "DependencyReference" - } - } + name: "Object" } - } - }, - additionalProperties: Trigger.type.additionalProperties - } -}; - -export const MultiplePipelineTrigger: msRest.CompositeMapper = { - serializedName: "MultiplePipelineTrigger", - type: { - name: "Composite", - polymorphicDiscriminator: Trigger.type.polymorphicDiscriminator, - uberParent: "Trigger", - className: "MultiplePipelineTrigger", - modelProperties: { - ...Trigger.type.modelProperties, - pipelines: { - serializedName: "pipelines", + }, + encryptedCredential: { + serializedName: "typeProperties.encryptedCredential", type: { - name: "Sequence", - element: { - type: { - name: "Composite", - className: "TriggerPipelineReference" - } - } + name: "Object" } } }, - additionalProperties: Trigger.type.additionalProperties + additionalProperties: LinkedService.type.additionalProperties } }; -export const BlobEventsTrigger: msRest.CompositeMapper = { - serializedName: "BlobEventsTrigger", +export const GoogleAdWordsLinkedService: msRest.CompositeMapper = { + serializedName: "GoogleAdWords", type: { name: "Composite", - polymorphicDiscriminator: Trigger.type.polymorphicDiscriminator, - uberParent: "Trigger", - className: "BlobEventsTrigger", + polymorphicDiscriminator: LinkedService.type.polymorphicDiscriminator, + uberParent: "LinkedService", + className: "GoogleAdWordsLinkedService", modelProperties: { - ...MultiplePipelineTrigger.type.modelProperties, - blobPathBeginsWith: { - serializedName: "typeProperties.blobPathBeginsWith", + ...LinkedService.type.modelProperties, + clientCustomerID: { + required: true, + serializedName: "typeProperties.clientCustomerID", type: { - name: "String" + name: "Object" } }, - blobPathEndsWith: { - serializedName: "typeProperties.blobPathEndsWith", + developerToken: { + required: true, + serializedName: "typeProperties.developerToken", type: { - name: "String" - } - }, - ignoreEmptyBlobs: { - serializedName: "typeProperties.ignoreEmptyBlobs", - type: { - name: "Boolean" - } - }, - events: { - required: true, - serializedName: "typeProperties.events", - type: { - name: "Sequence", - element: { - type: { - name: "String" - } - } - } - }, - scope: { - required: true, - serializedName: "typeProperties.scope", - type: { - name: "String" - } - } - }, - additionalProperties: Trigger.type.additionalProperties - } -}; - -export const BlobTrigger: msRest.CompositeMapper = { - serializedName: "BlobTrigger", - type: { - name: "Composite", - polymorphicDiscriminator: Trigger.type.polymorphicDiscriminator, - uberParent: "Trigger", - className: "BlobTrigger", - modelProperties: { - ...MultiplePipelineTrigger.type.modelProperties, - folderPath: { - required: true, - serializedName: "typeProperties.folderPath", - type: { - name: "String" - } - }, - maxConcurrency: { - required: true, - serializedName: "typeProperties.maxConcurrency", - type: { - name: "Number" - } - }, - linkedService: { - required: true, - serializedName: "typeProperties.linkedService", - defaultValue: {}, - type: { - name: "Composite", - className: "LinkedServiceReference" - } - } - }, - additionalProperties: Trigger.type.additionalProperties - } -}; - -export const RecurrenceScheduleOccurrence: msRest.CompositeMapper = { - serializedName: "RecurrenceScheduleOccurrence", - type: { - name: "Composite", - className: "RecurrenceScheduleOccurrence", - modelProperties: { - day: { - serializedName: "day", - type: { - name: "Enum", - allowedValues: [ - "Sunday", - "Monday", - "Tuesday", - "Wednesday", - "Thursday", - "Friday", - "Saturday" - ] - } - }, - occurrence: { - serializedName: "occurrence", - type: { - name: "Number" - } - } - }, - additionalProperties: { - type: { - name: "Object" - } - } - } -}; - -export const RecurrenceSchedule: msRest.CompositeMapper = { - serializedName: "RecurrenceSchedule", - type: { - name: "Composite", - className: "RecurrenceSchedule", - modelProperties: { - minutes: { - serializedName: "minutes", - type: { - name: "Sequence", - element: { - type: { - name: "Number" - } - } - } - }, - hours: { - serializedName: "hours", - type: { - name: "Sequence", - element: { - type: { - name: "Number" - } - } - } - }, - weekDays: { - serializedName: "weekDays", - type: { - name: "Sequence", - element: { - type: { - name: "Enum", - allowedValues: [ - "Sunday", - "Monday", - "Tuesday", - "Wednesday", - "Thursday", - "Friday", - "Saturday" - ] - } - } - } - }, - monthDays: { - serializedName: "monthDays", - type: { - name: "Sequence", - element: { - type: { - name: "Number" - } - } - } - }, - monthlyOccurrences: { - serializedName: "monthlyOccurrences", - type: { - name: "Sequence", - element: { - type: { - name: "Composite", - className: "RecurrenceScheduleOccurrence", - additionalProperties: { - type: { - name: "Object" - } - } - } - } - } - } - }, - additionalProperties: { - type: { - name: "Object" - } - } - } -}; - -export const ScheduleTriggerRecurrence: msRest.CompositeMapper = { - serializedName: "ScheduleTriggerRecurrence", - type: { - name: "Composite", - className: "ScheduleTriggerRecurrence", - modelProperties: { - frequency: { - serializedName: "frequency", - type: { - name: "String" - } - }, - interval: { - serializedName: "interval", - type: { - name: "Number" - } - }, - startTime: { - serializedName: "startTime", - type: { - name: "DateTime" - } - }, - endTime: { - serializedName: "endTime", - type: { - name: "DateTime" - } - }, - timeZone: { - serializedName: "timeZone", - type: { - name: "String" - } - }, - schedule: { - serializedName: "schedule", - type: { - name: "Composite", - className: "RecurrenceSchedule", - additionalProperties: { - type: { - name: "Object" - } - } - } - } - }, - additionalProperties: { - type: { - name: "Object" - } - } - } -}; - -export const ScheduleTrigger: msRest.CompositeMapper = { - serializedName: "ScheduleTrigger", - type: { - name: "Composite", - polymorphicDiscriminator: Trigger.type.polymorphicDiscriminator, - uberParent: "Trigger", - className: "ScheduleTrigger", - modelProperties: { - ...MultiplePipelineTrigger.type.modelProperties, - recurrence: { - required: true, - serializedName: "typeProperties.recurrence", - type: { - name: "Composite", - className: "ScheduleTriggerRecurrence", - additionalProperties: { - type: { - name: "Object" - } - } - } - } - }, - additionalProperties: Trigger.type.additionalProperties - } -}; - -export const AzureFunctionLinkedService: msRest.CompositeMapper = { - serializedName: "AzureFunction", - type: { - name: "Composite", - polymorphicDiscriminator: LinkedService.type.polymorphicDiscriminator, - uberParent: "LinkedService", - className: "AzureFunctionLinkedService", - modelProperties: { - ...LinkedService.type.modelProperties, - functionAppUrl: { - required: true, - serializedName: "typeProperties.functionAppUrl", - type: { - name: "Object" - } - }, - functionKey: { - serializedName: "typeProperties.functionKey", - type: { - name: "Composite", - className: "SecretBase" - } - }, - encryptedCredential: { - serializedName: "typeProperties.encryptedCredential", - type: { - name: "Object" - } - } - }, - additionalProperties: LinkedService.type.additionalProperties - } -}; - -export const AzureDataExplorerLinkedService: msRest.CompositeMapper = { - serializedName: "AzureDataExplorer", - type: { - name: "Composite", - polymorphicDiscriminator: LinkedService.type.polymorphicDiscriminator, - uberParent: "LinkedService", - className: "AzureDataExplorerLinkedService", - modelProperties: { - ...LinkedService.type.modelProperties, - endpoint: { - required: true, - serializedName: "typeProperties.endpoint", - type: { - name: "Object" - } - }, - servicePrincipalId: { - required: true, - serializedName: "typeProperties.servicePrincipalId", - type: { - name: "Object" - } - }, - servicePrincipalKey: { - required: true, - serializedName: "typeProperties.servicePrincipalKey", - type: { - name: "Composite", - className: "SecretBase" - } - }, - database: { - required: true, - serializedName: "typeProperties.database", - type: { - name: "Object" - } - }, - tenant: { - required: true, - serializedName: "typeProperties.tenant", - type: { - name: "Object" - } - } - }, - additionalProperties: LinkedService.type.additionalProperties - } -}; - -export const SapTableLinkedService: msRest.CompositeMapper = { - serializedName: "SapTable", - type: { - name: "Composite", - polymorphicDiscriminator: LinkedService.type.polymorphicDiscriminator, - uberParent: "LinkedService", - className: "SapTableLinkedService", - modelProperties: { - ...LinkedService.type.modelProperties, - server: { - serializedName: "typeProperties.server", - type: { - name: "Object" - } - }, - systemNumber: { - serializedName: "typeProperties.systemNumber", - type: { - name: "Object" - } - }, - clientId: { - serializedName: "typeProperties.clientId", - type: { - name: "Object" - } - }, - language: { - serializedName: "typeProperties.language", - type: { - name: "Object" - } - }, - systemId: { - serializedName: "typeProperties.systemId", - type: { - name: "Object" - } - }, - userName: { - serializedName: "typeProperties.userName", - type: { - name: "Object" - } - }, - password: { - serializedName: "typeProperties.password", - type: { - name: "Composite", - className: "SecretBase" - } - }, - messageServer: { - serializedName: "typeProperties.messageServer", - type: { - name: "Object" - } - }, - messageServerService: { - serializedName: "typeProperties.messageServerService", - type: { - name: "Object" - } - }, - sncMode: { - serializedName: "typeProperties.sncMode", - type: { - name: "Object" - } - }, - sncMyName: { - serializedName: "typeProperties.sncMyName", - type: { - name: "Object" - } - }, - sncPartnerName: { - serializedName: "typeProperties.sncPartnerName", - type: { - name: "Object" - } - }, - sncLibraryPath: { - serializedName: "typeProperties.sncLibraryPath", - type: { - name: "Object" - } - }, - sncQop: { - serializedName: "typeProperties.sncQop", - type: { - name: "Object" - } - }, - logonGroup: { - serializedName: "typeProperties.logonGroup", - type: { - name: "Object" - } - }, - encryptedCredential: { - serializedName: "typeProperties.encryptedCredential", - type: { - name: "Object" - } - } - }, - additionalProperties: LinkedService.type.additionalProperties - } -}; - -export const GoogleAdWordsLinkedService: msRest.CompositeMapper = { - serializedName: "GoogleAdWords", - type: { - name: "Composite", - polymorphicDiscriminator: LinkedService.type.polymorphicDiscriminator, - uberParent: "LinkedService", - className: "GoogleAdWordsLinkedService", - modelProperties: { - ...LinkedService.type.modelProperties, - clientCustomerID: { - required: true, - serializedName: "typeProperties.clientCustomerID", - type: { - name: "Object" - } - }, - developerToken: { - required: true, - serializedName: "typeProperties.developerToken", - type: { - name: "Composite", - className: "SecretBase" + name: "Composite", + className: "SecretBase" } }, authenticationType: { @@ -7510,6 +6880,12 @@ export const SalesforceServiceCloudLinkedService: msRest.CompositeMapper = { className: "SecretBase" } }, + apiVersion: { + serializedName: "typeProperties.apiVersion", + type: { + name: "Object" + } + }, extendedProperties: { serializedName: "typeProperties.extendedProperties", type: { @@ -7562,6 +6938,12 @@ export const SalesforceLinkedService: msRest.CompositeMapper = { className: "SecretBase" } }, + apiVersion: { + serializedName: "typeProperties.apiVersion", + type: { + name: "Object" + } + }, encryptedCredential: { serializedName: "typeProperties.encryptedCredential", type: { @@ -8520,15 +7902,19 @@ export const Db2LinkedService: msRest.CompositeMapper = { className: "Db2LinkedService", modelProperties: { ...LinkedService.type.modelProperties, + connectionString: { + serializedName: "typeProperties.connectionString", + type: { + name: "Object" + } + }, server: { - required: true, serializedName: "typeProperties.server", type: { name: "Object" } }, database: { - required: true, serializedName: "typeProperties.database", type: { name: "Object" @@ -9696,18 +9082,44 @@ export const AzureStorageLinkedService: msRest.CompositeMapper = { sasToken: { serializedName: "typeProperties.sasToken", type: { - name: "Composite", - className: "AzureKeyVaultSecretReference" + name: "Composite", + className: "AzureKeyVaultSecretReference" + } + }, + encryptedCredential: { + serializedName: "typeProperties.encryptedCredential", + type: { + name: "String" + } + } + }, + additionalProperties: LinkedService.type.additionalProperties + } +}; + +export const SnowflakeDataset: msRest.CompositeMapper = { + serializedName: "SnowflakeTable", + type: { + name: "Composite", + polymorphicDiscriminator: Dataset.type.polymorphicDiscriminator, + uberParent: "Dataset", + className: "SnowflakeDataset", + modelProperties: { + ...Dataset.type.modelProperties, + snowflakeDatasetSchema: { + serializedName: "typeProperties.schema", + type: { + name: "Object" } }, - encryptedCredential: { - serializedName: "typeProperties.encryptedCredential", + table: { + serializedName: "typeProperties.table", type: { - name: "String" + name: "Object" } } }, - additionalProperties: LinkedService.type.additionalProperties + additionalProperties: Dataset.type.additionalProperties } }; @@ -11160,13 +10572,269 @@ export const SapBwCubeDataset: msRest.CompositeMapper = { } }; -export const SybaseTableDataset: msRest.CompositeMapper = { - serializedName: "SybaseTable", +export const SybaseTableDataset: msRest.CompositeMapper = { + serializedName: "SybaseTable", + type: { + name: "Composite", + polymorphicDiscriminator: Dataset.type.polymorphicDiscriminator, + uberParent: "Dataset", + className: "SybaseTableDataset", + modelProperties: { + ...Dataset.type.modelProperties, + tableName: { + serializedName: "typeProperties.tableName", + type: { + name: "Object" + } + } + }, + additionalProperties: Dataset.type.additionalProperties + } +}; + +export const SalesforceServiceCloudObjectDataset: msRest.CompositeMapper = { + serializedName: "SalesforceServiceCloudObject", + type: { + name: "Composite", + polymorphicDiscriminator: Dataset.type.polymorphicDiscriminator, + uberParent: "Dataset", + className: "SalesforceServiceCloudObjectDataset", + modelProperties: { + ...Dataset.type.modelProperties, + objectApiName: { + serializedName: "typeProperties.objectApiName", + type: { + name: "Object" + } + } + }, + additionalProperties: Dataset.type.additionalProperties + } +}; + +export const SalesforceObjectDataset: msRest.CompositeMapper = { + serializedName: "SalesforceObject", + type: { + name: "Composite", + polymorphicDiscriminator: Dataset.type.polymorphicDiscriminator, + uberParent: "Dataset", + className: "SalesforceObjectDataset", + modelProperties: { + ...Dataset.type.modelProperties, + objectApiName: { + serializedName: "typeProperties.objectApiName", + type: { + name: "Object" + } + } + }, + additionalProperties: Dataset.type.additionalProperties + } +}; + +export const MicrosoftAccessTableDataset: msRest.CompositeMapper = { + serializedName: "MicrosoftAccessTable", + type: { + name: "Composite", + polymorphicDiscriminator: Dataset.type.polymorphicDiscriminator, + uberParent: "Dataset", + className: "MicrosoftAccessTableDataset", + modelProperties: { + ...Dataset.type.modelProperties, + tableName: { + serializedName: "typeProperties.tableName", + type: { + name: "Object" + } + } + }, + additionalProperties: Dataset.type.additionalProperties + } +}; + +export const PostgreSqlTableDataset: msRest.CompositeMapper = { + serializedName: "PostgreSqlTable", + type: { + name: "Composite", + polymorphicDiscriminator: Dataset.type.polymorphicDiscriminator, + uberParent: "Dataset", + className: "PostgreSqlTableDataset", + modelProperties: { + ...Dataset.type.modelProperties, + tableName: { + serializedName: "typeProperties.tableName", + type: { + name: "Object" + } + }, + table: { + serializedName: "typeProperties.table", + type: { + name: "Object" + } + }, + postgreSqlTableDatasetSchema: { + serializedName: "typeProperties.schema", + type: { + name: "Object" + } + } + }, + additionalProperties: Dataset.type.additionalProperties + } +}; + +export const MySqlTableDataset: msRest.CompositeMapper = { + serializedName: "MySqlTable", + type: { + name: "Composite", + polymorphicDiscriminator: Dataset.type.polymorphicDiscriminator, + uberParent: "Dataset", + className: "MySqlTableDataset", + modelProperties: { + ...Dataset.type.modelProperties, + tableName: { + serializedName: "typeProperties.tableName", + type: { + name: "Object" + } + } + }, + additionalProperties: Dataset.type.additionalProperties + } +}; + +export const OdbcTableDataset: msRest.CompositeMapper = { + serializedName: "OdbcTable", + type: { + name: "Composite", + polymorphicDiscriminator: Dataset.type.polymorphicDiscriminator, + uberParent: "Dataset", + className: "OdbcTableDataset", + modelProperties: { + ...Dataset.type.modelProperties, + tableName: { + serializedName: "typeProperties.tableName", + type: { + name: "Object" + } + } + }, + additionalProperties: Dataset.type.additionalProperties + } +}; + +export const InformixTableDataset: msRest.CompositeMapper = { + serializedName: "InformixTable", + type: { + name: "Composite", + polymorphicDiscriminator: Dataset.type.polymorphicDiscriminator, + uberParent: "Dataset", + className: "InformixTableDataset", + modelProperties: { + ...Dataset.type.modelProperties, + tableName: { + serializedName: "typeProperties.tableName", + type: { + name: "Object" + } + } + }, + additionalProperties: Dataset.type.additionalProperties + } +}; + +export const RelationalTableDataset: msRest.CompositeMapper = { + serializedName: "RelationalTable", + type: { + name: "Composite", + polymorphicDiscriminator: Dataset.type.polymorphicDiscriminator, + uberParent: "Dataset", + className: "RelationalTableDataset", + modelProperties: { + ...Dataset.type.modelProperties, + tableName: { + serializedName: "typeProperties.tableName", + type: { + name: "Object" + } + } + }, + additionalProperties: Dataset.type.additionalProperties + } +}; + +export const Db2TableDataset: msRest.CompositeMapper = { + serializedName: "Db2Table", + type: { + name: "Composite", + polymorphicDiscriminator: Dataset.type.polymorphicDiscriminator, + uberParent: "Dataset", + className: "Db2TableDataset", + modelProperties: { + ...Dataset.type.modelProperties, + tableName: { + serializedName: "typeProperties.tableName", + type: { + name: "Object" + } + }, + db2TableDatasetSchema: { + serializedName: "typeProperties.schema", + type: { + name: "Object" + } + }, + table: { + serializedName: "typeProperties.table", + type: { + name: "Object" + } + } + }, + additionalProperties: Dataset.type.additionalProperties + } +}; + +export const AmazonRedshiftTableDataset: msRest.CompositeMapper = { + serializedName: "AmazonRedshiftTable", + type: { + name: "Composite", + polymorphicDiscriminator: Dataset.type.polymorphicDiscriminator, + uberParent: "Dataset", + className: "AmazonRedshiftTableDataset", + modelProperties: { + ...Dataset.type.modelProperties, + tableName: { + serializedName: "typeProperties.tableName", + type: { + name: "Object" + } + }, + table: { + serializedName: "typeProperties.table", + type: { + name: "Object" + } + }, + amazonRedshiftTableDatasetSchema: { + serializedName: "typeProperties.schema", + type: { + name: "Object" + } + } + }, + additionalProperties: Dataset.type.additionalProperties + } +}; + +export const AzureMySqlTableDataset: msRest.CompositeMapper = { + serializedName: "AzureMySqlTable", type: { name: "Composite", polymorphicDiscriminator: Dataset.type.polymorphicDiscriminator, uberParent: "Dataset", - className: "SybaseTableDataset", + className: "AzureMySqlTableDataset", modelProperties: { ...Dataset.type.modelProperties, tableName: { @@ -11174,23 +10842,35 @@ export const SybaseTableDataset: msRest.CompositeMapper = { type: { name: "Object" } + }, + table: { + serializedName: "typeProperties.table", + type: { + name: "Object" + } } }, additionalProperties: Dataset.type.additionalProperties } }; -export const SalesforceServiceCloudObjectDataset: msRest.CompositeMapper = { - serializedName: "SalesforceServiceCloudObject", +export const TeradataTableDataset: msRest.CompositeMapper = { + serializedName: "TeradataTable", type: { name: "Composite", polymorphicDiscriminator: Dataset.type.polymorphicDiscriminator, uberParent: "Dataset", - className: "SalesforceServiceCloudObjectDataset", + className: "TeradataTableDataset", modelProperties: { ...Dataset.type.modelProperties, - objectApiName: { - serializedName: "typeProperties.objectApiName", + database: { + serializedName: "typeProperties.database", + type: { + name: "Object" + } + }, + table: { + serializedName: "typeProperties.table", type: { name: "Object" } @@ -11200,17 +10880,29 @@ export const SalesforceServiceCloudObjectDataset: msRest.CompositeMapper = { } }; -export const SalesforceObjectDataset: msRest.CompositeMapper = { - serializedName: "SalesforceObject", +export const OracleTableDataset: msRest.CompositeMapper = { + serializedName: "OracleTable", type: { name: "Composite", polymorphicDiscriminator: Dataset.type.polymorphicDiscriminator, uberParent: "Dataset", - className: "SalesforceObjectDataset", + className: "OracleTableDataset", modelProperties: { ...Dataset.type.modelProperties, - objectApiName: { - serializedName: "typeProperties.objectApiName", + tableName: { + serializedName: "typeProperties.tableName", + type: { + name: "Object" + } + }, + oracleTableDatasetSchema: { + serializedName: "typeProperties.schema", + type: { + name: "Object" + } + }, + table: { + serializedName: "typeProperties.table", type: { name: "Object" } @@ -11220,17 +10912,17 @@ export const SalesforceObjectDataset: msRest.CompositeMapper = { } }; -export const MicrosoftAccessTableDataset: msRest.CompositeMapper = { - serializedName: "MicrosoftAccessTable", +export const ODataResourceDataset: msRest.CompositeMapper = { + serializedName: "ODataResource", type: { name: "Composite", polymorphicDiscriminator: Dataset.type.polymorphicDiscriminator, uberParent: "Dataset", - className: "MicrosoftAccessTableDataset", + className: "ODataResourceDataset", modelProperties: { ...Dataset.type.modelProperties, - tableName: { - serializedName: "typeProperties.tableName", + path: { + serializedName: "typeProperties.path", type: { name: "Object" } @@ -11240,29 +10932,18 @@ export const MicrosoftAccessTableDataset: msRest.CompositeMapper = { } }; -export const PostgreSqlTableDataset: msRest.CompositeMapper = { - serializedName: "PostgreSqlTable", +export const CosmosDbMongoDbApiCollectionDataset: msRest.CompositeMapper = { + serializedName: "CosmosDbMongoDbApiCollection", type: { name: "Composite", polymorphicDiscriminator: Dataset.type.polymorphicDiscriminator, uberParent: "Dataset", - className: "PostgreSqlTableDataset", + className: "CosmosDbMongoDbApiCollectionDataset", modelProperties: { ...Dataset.type.modelProperties, - tableName: { - serializedName: "typeProperties.tableName", - type: { - name: "Object" - } - }, - table: { - serializedName: "typeProperties.table", - type: { - name: "Object" - } - }, - postgreSqlTableDatasetSchema: { - serializedName: "typeProperties.schema", + collection: { + required: true, + serializedName: "typeProperties.collection", type: { name: "Object" } @@ -11272,17 +10953,18 @@ export const PostgreSqlTableDataset: msRest.CompositeMapper = { } }; -export const MySqlTableDataset: msRest.CompositeMapper = { - serializedName: "MySqlTable", +export const MongoDbV2CollectionDataset: msRest.CompositeMapper = { + serializedName: "MongoDbV2Collection", type: { name: "Composite", polymorphicDiscriminator: Dataset.type.polymorphicDiscriminator, uberParent: "Dataset", - className: "MySqlTableDataset", + className: "MongoDbV2CollectionDataset", modelProperties: { ...Dataset.type.modelProperties, - tableName: { - serializedName: "typeProperties.tableName", + collection: { + required: true, + serializedName: "typeProperties.collection", type: { name: "Object" } @@ -11292,17 +10974,18 @@ export const MySqlTableDataset: msRest.CompositeMapper = { } }; -export const OdbcTableDataset: msRest.CompositeMapper = { - serializedName: "OdbcTable", +export const MongoDbCollectionDataset: msRest.CompositeMapper = { + serializedName: "MongoDbCollection", type: { name: "Composite", polymorphicDiscriminator: Dataset.type.polymorphicDiscriminator, uberParent: "Dataset", - className: "OdbcTableDataset", + className: "MongoDbCollectionDataset", modelProperties: { ...Dataset.type.modelProperties, - tableName: { - serializedName: "typeProperties.tableName", + collectionName: { + required: true, + serializedName: "typeProperties.collectionName", type: { name: "Object" } @@ -11312,71 +10995,144 @@ export const OdbcTableDataset: msRest.CompositeMapper = { } }; -export const InformixTableDataset: msRest.CompositeMapper = { - serializedName: "InformixTable", +export const FileShareDataset: msRest.CompositeMapper = { + serializedName: "FileShare", type: { name: "Composite", polymorphicDiscriminator: Dataset.type.polymorphicDiscriminator, uberParent: "Dataset", - className: "InformixTableDataset", + className: "FileShareDataset", modelProperties: { ...Dataset.type.modelProperties, - tableName: { - serializedName: "typeProperties.tableName", + folderPath: { + serializedName: "typeProperties.folderPath", + type: { + name: "Object" + } + }, + fileName: { + serializedName: "typeProperties.fileName", + type: { + name: "Object" + } + }, + modifiedDatetimeStart: { + serializedName: "typeProperties.modifiedDatetimeStart", + type: { + name: "Object" + } + }, + modifiedDatetimeEnd: { + serializedName: "typeProperties.modifiedDatetimeEnd", + type: { + name: "Object" + } + }, + format: { + serializedName: "typeProperties.format", + type: { + name: "Composite", + className: "DatasetStorageFormat", + additionalProperties: { + type: { + name: "Object" + } + } + } + }, + fileFilter: { + serializedName: "typeProperties.fileFilter", type: { name: "Object" } + }, + compression: { + serializedName: "typeProperties.compression", + type: { + name: "Composite", + className: "DatasetCompression", + additionalProperties: { + type: { + name: "Object" + } + } + } } }, additionalProperties: Dataset.type.additionalProperties } }; -export const RelationalTableDataset: msRest.CompositeMapper = { - serializedName: "RelationalTable", +export const Office365Dataset: msRest.CompositeMapper = { + serializedName: "Office365Table", type: { name: "Composite", polymorphicDiscriminator: Dataset.type.polymorphicDiscriminator, uberParent: "Dataset", - className: "RelationalTableDataset", + className: "Office365Dataset", modelProperties: { ...Dataset.type.modelProperties, tableName: { + required: true, serializedName: "typeProperties.tableName", type: { name: "Object" } + }, + predicate: { + serializedName: "typeProperties.predicate", + type: { + name: "Object" + } } }, additionalProperties: Dataset.type.additionalProperties } }; -export const Db2TableDataset: msRest.CompositeMapper = { - serializedName: "Db2Table", +export const AzureBlobFSDataset: msRest.CompositeMapper = { + serializedName: "AzureBlobFSFile", type: { name: "Composite", polymorphicDiscriminator: Dataset.type.polymorphicDiscriminator, uberParent: "Dataset", - className: "Db2TableDataset", + className: "AzureBlobFSDataset", modelProperties: { ...Dataset.type.modelProperties, - tableName: { - serializedName: "typeProperties.tableName", + folderPath: { + serializedName: "typeProperties.folderPath", type: { name: "Object" } }, - db2TableDatasetSchema: { - serializedName: "typeProperties.schema", + fileName: { + serializedName: "typeProperties.fileName", type: { name: "Object" } }, - table: { - serializedName: "typeProperties.table", + format: { + serializedName: "typeProperties.format", type: { - name: "Object" + name: "Composite", + className: "DatasetStorageFormat", + additionalProperties: { + type: { + name: "Object" + } + } + } + }, + compression: { + serializedName: "typeProperties.compression", + type: { + name: "Composite", + className: "DatasetCompression", + additionalProperties: { + type: { + name: "Object" + } + } } } }, @@ -11384,31 +11140,49 @@ export const Db2TableDataset: msRest.CompositeMapper = { } }; -export const AmazonRedshiftTableDataset: msRest.CompositeMapper = { - serializedName: "AmazonRedshiftTable", +export const AzureDataLakeStoreDataset: msRest.CompositeMapper = { + serializedName: "AzureDataLakeStoreFile", type: { name: "Composite", polymorphicDiscriminator: Dataset.type.polymorphicDiscriminator, uberParent: "Dataset", - className: "AmazonRedshiftTableDataset", + className: "AzureDataLakeStoreDataset", modelProperties: { ...Dataset.type.modelProperties, - tableName: { - serializedName: "typeProperties.tableName", + folderPath: { + serializedName: "typeProperties.folderPath", type: { name: "Object" } }, - table: { - serializedName: "typeProperties.table", + fileName: { + serializedName: "typeProperties.fileName", type: { name: "Object" } }, - amazonRedshiftTableDatasetSchema: { - serializedName: "typeProperties.schema", + format: { + serializedName: "typeProperties.format", type: { - name: "Object" + name: "Composite", + className: "DatasetStorageFormat", + additionalProperties: { + type: { + name: "Object" + } + } + } + }, + compression: { + serializedName: "typeProperties.compression", + type: { + name: "Composite", + className: "DatasetCompression", + additionalProperties: { + type: { + name: "Object" + } + } } } }, @@ -11416,23 +11190,17 @@ export const AmazonRedshiftTableDataset: msRest.CompositeMapper = { } }; -export const AzureMySqlTableDataset: msRest.CompositeMapper = { - serializedName: "AzureMySqlTable", +export const CommonDataServiceForAppsEntityDataset: msRest.CompositeMapper = { + serializedName: "CommonDataServiceForAppsEntity", type: { name: "Composite", polymorphicDiscriminator: Dataset.type.polymorphicDiscriminator, uberParent: "Dataset", - className: "AzureMySqlTableDataset", + className: "CommonDataServiceForAppsEntityDataset", modelProperties: { ...Dataset.type.modelProperties, - tableName: { - serializedName: "typeProperties.tableName", - type: { - name: "Object" - } - }, - table: { - serializedName: "typeProperties.table", + entityName: { + serializedName: "typeProperties.entityName", type: { name: "Object" } @@ -11442,23 +11210,17 @@ export const AzureMySqlTableDataset: msRest.CompositeMapper = { } }; -export const TeradataTableDataset: msRest.CompositeMapper = { - serializedName: "TeradataTable", +export const DynamicsCrmEntityDataset: msRest.CompositeMapper = { + serializedName: "DynamicsCrmEntity", type: { name: "Composite", polymorphicDiscriminator: Dataset.type.polymorphicDiscriminator, uberParent: "Dataset", - className: "TeradataTableDataset", + className: "DynamicsCrmEntityDataset", modelProperties: { ...Dataset.type.modelProperties, - database: { - serializedName: "typeProperties.database", - type: { - name: "Object" - } - }, - table: { - serializedName: "typeProperties.table", + entityName: { + serializedName: "typeProperties.entityName", type: { name: "Object" } @@ -11468,29 +11230,17 @@ export const TeradataTableDataset: msRest.CompositeMapper = { } }; -export const OracleTableDataset: msRest.CompositeMapper = { - serializedName: "OracleTable", +export const DynamicsEntityDataset: msRest.CompositeMapper = { + serializedName: "DynamicsEntity", type: { name: "Composite", polymorphicDiscriminator: Dataset.type.polymorphicDiscriminator, uberParent: "Dataset", - className: "OracleTableDataset", + className: "DynamicsEntityDataset", modelProperties: { ...Dataset.type.modelProperties, - tableName: { - serializedName: "typeProperties.tableName", - type: { - name: "Object" - } - }, - oracleTableDatasetSchema: { - serializedName: "typeProperties.schema", - type: { - name: "Object" - } - }, - table: { - serializedName: "typeProperties.table", + entityName: { + serializedName: "typeProperties.entityName", type: { name: "Object" } @@ -11500,17 +11250,18 @@ export const OracleTableDataset: msRest.CompositeMapper = { } }; -export const ODataResourceDataset: msRest.CompositeMapper = { - serializedName: "ODataResource", +export const DocumentDbCollectionDataset: msRest.CompositeMapper = { + serializedName: "DocumentDbCollection", type: { name: "Composite", polymorphicDiscriminator: Dataset.type.polymorphicDiscriminator, uberParent: "Dataset", - className: "ODataResourceDataset", + className: "DocumentDbCollectionDataset", modelProperties: { ...Dataset.type.modelProperties, - path: { - serializedName: "typeProperties.path", + collectionName: { + required: true, + serializedName: "typeProperties.collectionName", type: { name: "Object" } @@ -11520,18 +11271,18 @@ export const ODataResourceDataset: msRest.CompositeMapper = { } }; -export const CosmosDbMongoDbApiCollectionDataset: msRest.CompositeMapper = { - serializedName: "CosmosDbMongoDbApiCollection", +export const CosmosDbSqlApiCollectionDataset: msRest.CompositeMapper = { + serializedName: "CosmosDbSqlApiCollection", type: { name: "Composite", polymorphicDiscriminator: Dataset.type.polymorphicDiscriminator, uberParent: "Dataset", - className: "CosmosDbMongoDbApiCollectionDataset", + className: "CosmosDbSqlApiCollectionDataset", modelProperties: { ...Dataset.type.modelProperties, - collection: { + collectionName: { required: true, - serializedName: "typeProperties.collection", + serializedName: "typeProperties.collectionName", type: { name: "Object" } @@ -11541,18 +11292,17 @@ export const CosmosDbMongoDbApiCollectionDataset: msRest.CompositeMapper = { } }; -export const MongoDbV2CollectionDataset: msRest.CompositeMapper = { - serializedName: "MongoDbV2Collection", +export const CustomDataset: msRest.CompositeMapper = { + serializedName: "CustomDataset", type: { name: "Composite", polymorphicDiscriminator: Dataset.type.polymorphicDiscriminator, uberParent: "Dataset", - className: "MongoDbV2CollectionDataset", + className: "CustomDataset", modelProperties: { ...Dataset.type.modelProperties, - collection: { - required: true, - serializedName: "typeProperties.collection", + typeProperties: { + serializedName: "typeProperties", type: { name: "Object" } @@ -11562,18 +11312,23 @@ export const MongoDbV2CollectionDataset: msRest.CompositeMapper = { } }; -export const MongoDbCollectionDataset: msRest.CompositeMapper = { - serializedName: "MongoDbCollection", +export const CassandraTableDataset: msRest.CompositeMapper = { + serializedName: "CassandraTable", type: { name: "Composite", polymorphicDiscriminator: Dataset.type.polymorphicDiscriminator, uberParent: "Dataset", - className: "MongoDbCollectionDataset", + className: "CassandraTableDataset", modelProperties: { ...Dataset.type.modelProperties, - collectionName: { - required: true, - serializedName: "typeProperties.collectionName", + tableName: { + serializedName: "typeProperties.tableName", + type: { + name: "Object" + } + }, + keyspace: { + serializedName: "typeProperties.keyspace", type: { name: "Object" } @@ -11583,67 +11338,63 @@ export const MongoDbCollectionDataset: msRest.CompositeMapper = { } }; -export const FileShareDataset: msRest.CompositeMapper = { - serializedName: "FileShare", +export const AzureSqlDWTableDataset: msRest.CompositeMapper = { + serializedName: "AzureSqlDWTable", type: { name: "Composite", polymorphicDiscriminator: Dataset.type.polymorphicDiscriminator, uberParent: "Dataset", - className: "FileShareDataset", + className: "AzureSqlDWTableDataset", modelProperties: { ...Dataset.type.modelProperties, - folderPath: { - serializedName: "typeProperties.folderPath", + tableName: { + serializedName: "typeProperties.tableName", type: { name: "Object" } }, - fileName: { - serializedName: "typeProperties.fileName", + azureSqlDWTableDatasetSchema: { + serializedName: "typeProperties.schema", type: { name: "Object" } }, - modifiedDatetimeStart: { - serializedName: "typeProperties.modifiedDatetimeStart", + table: { + serializedName: "typeProperties.table", type: { name: "Object" } - }, - modifiedDatetimeEnd: { - serializedName: "typeProperties.modifiedDatetimeEnd", + } + }, + additionalProperties: Dataset.type.additionalProperties + } +}; + +export const AzureSqlMITableDataset: msRest.CompositeMapper = { + serializedName: "AzureSqlMITable", + type: { + name: "Composite", + polymorphicDiscriminator: Dataset.type.polymorphicDiscriminator, + uberParent: "Dataset", + className: "AzureSqlMITableDataset", + modelProperties: { + ...Dataset.type.modelProperties, + tableName: { + serializedName: "typeProperties.tableName", type: { name: "Object" } }, - format: { - serializedName: "typeProperties.format", - type: { - name: "Composite", - className: "DatasetStorageFormat", - additionalProperties: { - type: { - name: "Object" - } - } - } - }, - fileFilter: { - serializedName: "typeProperties.fileFilter", + azureSqlMITableDatasetSchema: { + serializedName: "typeProperties.schema", type: { name: "Object" } }, - compression: { - serializedName: "typeProperties.compression", + table: { + serializedName: "typeProperties.table", type: { - name: "Composite", - className: "DatasetCompression", - additionalProperties: { - type: { - name: "Object" - } - } + name: "Object" } } }, @@ -11651,24 +11402,29 @@ export const FileShareDataset: msRest.CompositeMapper = { } }; -export const Office365Dataset: msRest.CompositeMapper = { - serializedName: "Office365Table", +export const AzureSqlTableDataset: msRest.CompositeMapper = { + serializedName: "AzureSqlTable", type: { name: "Composite", polymorphicDiscriminator: Dataset.type.polymorphicDiscriminator, uberParent: "Dataset", - className: "Office365Dataset", + className: "AzureSqlTableDataset", modelProperties: { ...Dataset.type.modelProperties, tableName: { - required: true, serializedName: "typeProperties.tableName", type: { name: "Object" } }, - predicate: { - serializedName: "typeProperties.predicate", + azureSqlTableDatasetSchema: { + serializedName: "typeProperties.schema", + type: { + name: "Object" + } + }, + table: { + serializedName: "typeProperties.table", type: { name: "Object" } @@ -11678,63 +11434,34 @@ export const Office365Dataset: msRest.CompositeMapper = { } }; -export const AzureBlobFSDataset: msRest.CompositeMapper = { - serializedName: "AzureBlobFSFile", +export const AzureTableDataset: msRest.CompositeMapper = { + serializedName: "AzureTable", type: { name: "Composite", polymorphicDiscriminator: Dataset.type.polymorphicDiscriminator, uberParent: "Dataset", - className: "AzureBlobFSDataset", + className: "AzureTableDataset", modelProperties: { ...Dataset.type.modelProperties, - folderPath: { - serializedName: "typeProperties.folderPath", - type: { - name: "Object" - } - }, - fileName: { - serializedName: "typeProperties.fileName", + tableName: { + required: true, + serializedName: "typeProperties.tableName", type: { name: "Object" } - }, - format: { - serializedName: "typeProperties.format", - type: { - name: "Composite", - className: "DatasetStorageFormat", - additionalProperties: { - type: { - name: "Object" - } - } - } - }, - compression: { - serializedName: "typeProperties.compression", - type: { - name: "Composite", - className: "DatasetCompression", - additionalProperties: { - type: { - name: "Object" - } - } - } } }, additionalProperties: Dataset.type.additionalProperties } }; -export const AzureDataLakeStoreDataset: msRest.CompositeMapper = { - serializedName: "AzureDataLakeStoreFile", +export const AzureBlobDataset: msRest.CompositeMapper = { + serializedName: "AzureBlob", type: { name: "Composite", polymorphicDiscriminator: Dataset.type.polymorphicDiscriminator, uberParent: "Dataset", - className: "AzureDataLakeStoreDataset", + className: "AzureBlobDataset", modelProperties: { ...Dataset.type.modelProperties, folderPath: { @@ -11743,8 +11470,26 @@ export const AzureDataLakeStoreDataset: msRest.CompositeMapper = { name: "Object" } }, - fileName: { - serializedName: "typeProperties.fileName", + tableRootLocation: { + serializedName: "typeProperties.tableRootLocation", + type: { + name: "Object" + } + }, + fileName: { + serializedName: "typeProperties.fileName", + type: { + name: "Object" + } + }, + modifiedDatetimeStart: { + serializedName: "typeProperties.modifiedDatetimeStart", + type: { + name: "Object" + } + }, + modifiedDatetimeEnd: { + serializedName: "typeProperties.modifiedDatetimeEnd", type: { name: "Object" } @@ -11778,121 +11523,273 @@ export const AzureDataLakeStoreDataset: msRest.CompositeMapper = { } }; -export const CommonDataServiceForAppsEntityDataset: msRest.CompositeMapper = { - serializedName: "CommonDataServiceForAppsEntity", +export const DatasetLocation: msRest.CompositeMapper = { + serializedName: "DatasetLocation", type: { name: "Composite", - polymorphicDiscriminator: Dataset.type.polymorphicDiscriminator, - uberParent: "Dataset", - className: "CommonDataServiceForAppsEntityDataset", + polymorphicDiscriminator: { + serializedName: "type", + clientName: "type" + }, + uberParent: "DatasetLocation", + className: "DatasetLocation", modelProperties: { - ...Dataset.type.modelProperties, - entityName: { - serializedName: "typeProperties.entityName", + folderPath: { + serializedName: "folderPath", type: { name: "Object" } + }, + fileName: { + serializedName: "fileName", + type: { + name: "Object" + } + }, + type: { + required: true, + serializedName: "type", + type: { + name: "String" + } } }, - additionalProperties: Dataset.type.additionalProperties + additionalProperties: { + type: { + name: "Object" + } + } } }; -export const DynamicsCrmEntityDataset: msRest.CompositeMapper = { - serializedName: "DynamicsCrmEntity", +export const HdfsLocation: msRest.CompositeMapper = { + serializedName: "HdfsLocation", type: { name: "Composite", - polymorphicDiscriminator: Dataset.type.polymorphicDiscriminator, - uberParent: "Dataset", - className: "DynamicsCrmEntityDataset", + polymorphicDiscriminator: DatasetLocation.type.polymorphicDiscriminator, + uberParent: "DatasetLocation", + className: "HdfsLocation", modelProperties: { - ...Dataset.type.modelProperties, - entityName: { - serializedName: "typeProperties.entityName", + ...DatasetLocation.type.modelProperties + }, + additionalProperties: DatasetLocation.type.additionalProperties + } +}; + +export const HttpServerLocation: msRest.CompositeMapper = { + serializedName: "HttpServerLocation", + type: { + name: "Composite", + polymorphicDiscriminator: DatasetLocation.type.polymorphicDiscriminator, + uberParent: "DatasetLocation", + className: "HttpServerLocation", + modelProperties: { + ...DatasetLocation.type.modelProperties, + relativeUrl: { + serializedName: "relativeUrl", type: { name: "Object" } } }, - additionalProperties: Dataset.type.additionalProperties + additionalProperties: DatasetLocation.type.additionalProperties } }; -export const DynamicsEntityDataset: msRest.CompositeMapper = { - serializedName: "DynamicsEntity", +export const SftpLocation: msRest.CompositeMapper = { + serializedName: "SftpLocation", type: { name: "Composite", - polymorphicDiscriminator: Dataset.type.polymorphicDiscriminator, - uberParent: "Dataset", - className: "DynamicsEntityDataset", + polymorphicDiscriminator: DatasetLocation.type.polymorphicDiscriminator, + uberParent: "DatasetLocation", + className: "SftpLocation", modelProperties: { - ...Dataset.type.modelProperties, - entityName: { - serializedName: "typeProperties.entityName", + ...DatasetLocation.type.modelProperties + }, + additionalProperties: DatasetLocation.type.additionalProperties + } +}; + +export const FtpServerLocation: msRest.CompositeMapper = { + serializedName: "FtpServerLocation", + type: { + name: "Composite", + polymorphicDiscriminator: DatasetLocation.type.polymorphicDiscriminator, + uberParent: "DatasetLocation", + className: "FtpServerLocation", + modelProperties: { + ...DatasetLocation.type.modelProperties + }, + additionalProperties: DatasetLocation.type.additionalProperties + } +}; + +export const GoogleCloudStorageLocation: msRest.CompositeMapper = { + serializedName: "GoogleCloudStorageLocation", + type: { + name: "Composite", + polymorphicDiscriminator: DatasetLocation.type.polymorphicDiscriminator, + uberParent: "DatasetLocation", + className: "GoogleCloudStorageLocation", + modelProperties: { + ...DatasetLocation.type.modelProperties, + bucketName: { + serializedName: "bucketName", + type: { + name: "Object" + } + }, + version: { + serializedName: "version", type: { name: "Object" } } }, - additionalProperties: Dataset.type.additionalProperties + additionalProperties: DatasetLocation.type.additionalProperties } }; -export const DocumentDbCollectionDataset: msRest.CompositeMapper = { - serializedName: "DocumentDbCollection", +export const AzureFileStorageLocation: msRest.CompositeMapper = { + serializedName: "AzureFileStorageLocation", type: { name: "Composite", - polymorphicDiscriminator: Dataset.type.polymorphicDiscriminator, - uberParent: "Dataset", - className: "DocumentDbCollectionDataset", + polymorphicDiscriminator: DatasetLocation.type.polymorphicDiscriminator, + uberParent: "DatasetLocation", + className: "AzureFileStorageLocation", modelProperties: { - ...Dataset.type.modelProperties, - collectionName: { - required: true, - serializedName: "typeProperties.collectionName", + ...DatasetLocation.type.modelProperties + }, + additionalProperties: DatasetLocation.type.additionalProperties + } +}; + +export const FileServerLocation: msRest.CompositeMapper = { + serializedName: "FileServerLocation", + type: { + name: "Composite", + polymorphicDiscriminator: DatasetLocation.type.polymorphicDiscriminator, + uberParent: "DatasetLocation", + className: "FileServerLocation", + modelProperties: { + ...DatasetLocation.type.modelProperties + }, + additionalProperties: DatasetLocation.type.additionalProperties + } +}; + +export const AmazonS3Location: msRest.CompositeMapper = { + serializedName: "AmazonS3Location", + type: { + name: "Composite", + polymorphicDiscriminator: DatasetLocation.type.polymorphicDiscriminator, + uberParent: "DatasetLocation", + className: "AmazonS3Location", + modelProperties: { + ...DatasetLocation.type.modelProperties, + bucketName: { + serializedName: "bucketName", + type: { + name: "Object" + } + }, + version: { + serializedName: "version", type: { name: "Object" } } }, - additionalProperties: Dataset.type.additionalProperties + additionalProperties: DatasetLocation.type.additionalProperties } }; -export const CosmosDbSqlApiCollectionDataset: msRest.CompositeMapper = { - serializedName: "CosmosDbSqlApiCollection", +export const AzureDataLakeStoreLocation: msRest.CompositeMapper = { + serializedName: "AzureDataLakeStoreLocation", type: { name: "Composite", - polymorphicDiscriminator: Dataset.type.polymorphicDiscriminator, - uberParent: "Dataset", - className: "CosmosDbSqlApiCollectionDataset", + polymorphicDiscriminator: DatasetLocation.type.polymorphicDiscriminator, + uberParent: "DatasetLocation", + className: "AzureDataLakeStoreLocation", modelProperties: { - ...Dataset.type.modelProperties, - collectionName: { - required: true, - serializedName: "typeProperties.collectionName", + ...DatasetLocation.type.modelProperties + }, + additionalProperties: DatasetLocation.type.additionalProperties + } +}; + +export const AzureBlobFSLocation: msRest.CompositeMapper = { + serializedName: "AzureBlobFSLocation", + type: { + name: "Composite", + polymorphicDiscriminator: DatasetLocation.type.polymorphicDiscriminator, + uberParent: "DatasetLocation", + className: "AzureBlobFSLocation", + modelProperties: { + ...DatasetLocation.type.modelProperties, + fileSystem: { + serializedName: "fileSystem", type: { name: "Object" } } }, - additionalProperties: Dataset.type.additionalProperties + additionalProperties: DatasetLocation.type.additionalProperties } }; -export const CustomDataset: msRest.CompositeMapper = { - serializedName: "CustomDataset", +export const AzureBlobStorageLocation: msRest.CompositeMapper = { + serializedName: "AzureBlobStorageLocation", + type: { + name: "Composite", + polymorphicDiscriminator: DatasetLocation.type.polymorphicDiscriminator, + uberParent: "DatasetLocation", + className: "AzureBlobStorageLocation", + modelProperties: { + ...DatasetLocation.type.modelProperties, + container: { + serializedName: "container", + type: { + name: "Object" + } + } + }, + additionalProperties: DatasetLocation.type.additionalProperties + } +}; + +export const BinaryDataset: msRest.CompositeMapper = { + serializedName: "Binary", type: { name: "Composite", polymorphicDiscriminator: Dataset.type.polymorphicDiscriminator, uberParent: "Dataset", - className: "CustomDataset", + className: "BinaryDataset", modelProperties: { ...Dataset.type.modelProperties, - typeProperties: { - serializedName: "typeProperties", + location: { + required: true, + serializedName: "typeProperties.location", + type: { + name: "Composite", + className: "DatasetLocation", + additionalProperties: { + type: { + name: "Object" + } + } + } + }, + compression: { + serializedName: "typeProperties.compression", type: { - name: "Object" + name: "Composite", + className: "DatasetCompression", + additionalProperties: { + type: { + name: "Object" + } + } } } }, @@ -11900,25 +11797,32 @@ export const CustomDataset: msRest.CompositeMapper = { } }; -export const CassandraTableDataset: msRest.CompositeMapper = { - serializedName: "CassandraTable", +export const OrcDataset: msRest.CompositeMapper = { + serializedName: "Orc", type: { name: "Composite", polymorphicDiscriminator: Dataset.type.polymorphicDiscriminator, uberParent: "Dataset", - className: "CassandraTableDataset", + className: "OrcDataset", modelProperties: { ...Dataset.type.modelProperties, - tableName: { - serializedName: "typeProperties.tableName", + location: { + required: true, + serializedName: "typeProperties.location", type: { - name: "Object" + name: "Composite", + className: "DatasetLocation", + additionalProperties: { + type: { + name: "Object" + } + } } }, - keyspace: { - serializedName: "typeProperties.keyspace", + orcCompressionCodec: { + serializedName: "typeProperties.orcCompressionCodec", type: { - name: "Object" + name: "String" } } }, @@ -11926,31 +11830,44 @@ export const CassandraTableDataset: msRest.CompositeMapper = { } }; -export const AzureSqlDWTableDataset: msRest.CompositeMapper = { - serializedName: "AzureSqlDWTable", +export const JsonDataset: msRest.CompositeMapper = { + serializedName: "Json", type: { name: "Composite", polymorphicDiscriminator: Dataset.type.polymorphicDiscriminator, uberParent: "Dataset", - className: "AzureSqlDWTableDataset", + className: "JsonDataset", modelProperties: { ...Dataset.type.modelProperties, - tableName: { - serializedName: "typeProperties.tableName", + location: { + required: true, + serializedName: "typeProperties.location", type: { - name: "Object" + name: "Composite", + className: "DatasetLocation", + additionalProperties: { + type: { + name: "Object" + } + } } }, - azureSqlDWTableDatasetSchema: { - serializedName: "typeProperties.schema", + encodingName: { + serializedName: "typeProperties.encodingName", type: { name: "Object" } }, - table: { - serializedName: "typeProperties.table", + compression: { + serializedName: "typeProperties.compression", type: { - name: "Object" + name: "Composite", + className: "DatasetCompression", + additionalProperties: { + type: { + name: "Object" + } + } } } }, @@ -11958,29 +11875,78 @@ export const AzureSqlDWTableDataset: msRest.CompositeMapper = { } }; -export const AzureSqlMITableDataset: msRest.CompositeMapper = { - serializedName: "AzureSqlMITable", +export const DelimitedTextDataset: msRest.CompositeMapper = { + serializedName: "DelimitedText", type: { name: "Composite", polymorphicDiscriminator: Dataset.type.polymorphicDiscriminator, uberParent: "Dataset", - className: "AzureSqlMITableDataset", + className: "DelimitedTextDataset", modelProperties: { ...Dataset.type.modelProperties, - tableName: { - serializedName: "typeProperties.tableName", + location: { + required: true, + serializedName: "typeProperties.location", + type: { + name: "Composite", + className: "DatasetLocation", + additionalProperties: { + type: { + name: "Object" + } + } + } + }, + columnDelimiter: { + serializedName: "typeProperties.columnDelimiter", type: { name: "Object" } }, - azureSqlMITableDatasetSchema: { - serializedName: "typeProperties.schema", + rowDelimiter: { + serializedName: "typeProperties.rowDelimiter", type: { name: "Object" } }, - table: { - serializedName: "typeProperties.table", + encodingName: { + serializedName: "typeProperties.encodingName", + type: { + name: "Object" + } + }, + compressionCodec: { + serializedName: "typeProperties.compressionCodec", + type: { + name: "Object" + } + }, + compressionLevel: { + serializedName: "typeProperties.compressionLevel", + type: { + name: "Object" + } + }, + quoteChar: { + serializedName: "typeProperties.quoteChar", + type: { + name: "Object" + } + }, + escapeChar: { + serializedName: "typeProperties.escapeChar", + type: { + name: "Object" + } + }, + firstRowAsHeader: { + serializedName: "typeProperties.firstRowAsHeader", + type: { + name: "Object" + } + }, + nullValue: { + serializedName: "typeProperties.nullValue", type: { name: "Object" } @@ -11990,29 +11956,30 @@ export const AzureSqlMITableDataset: msRest.CompositeMapper = { } }; -export const AzureSqlTableDataset: msRest.CompositeMapper = { - serializedName: "AzureSqlTable", +export const ParquetDataset: msRest.CompositeMapper = { + serializedName: "Parquet", type: { name: "Composite", polymorphicDiscriminator: Dataset.type.polymorphicDiscriminator, uberParent: "Dataset", - className: "AzureSqlTableDataset", + className: "ParquetDataset", modelProperties: { ...Dataset.type.modelProperties, - tableName: { - serializedName: "typeProperties.tableName", - type: { - name: "Object" - } - }, - azureSqlTableDatasetSchema: { - serializedName: "typeProperties.schema", + location: { + required: true, + serializedName: "typeProperties.location", type: { - name: "Object" + name: "Composite", + className: "DatasetLocation", + additionalProperties: { + type: { + name: "Object" + } + } } }, - table: { - serializedName: "typeProperties.table", + compressionCodec: { + serializedName: "typeProperties.compressionCodec", type: { name: "Object" } @@ -12022,20 +11989,42 @@ export const AzureSqlTableDataset: msRest.CompositeMapper = { } }; -export const AzureTableDataset: msRest.CompositeMapper = { - serializedName: "AzureTable", +export const AvroDataset: msRest.CompositeMapper = { + serializedName: "Avro", type: { name: "Composite", polymorphicDiscriminator: Dataset.type.polymorphicDiscriminator, uberParent: "Dataset", - className: "AzureTableDataset", + className: "AvroDataset", modelProperties: { ...Dataset.type.modelProperties, - tableName: { + location: { required: true, - serializedName: "typeProperties.tableName", + serializedName: "typeProperties.location", type: { - name: "Object" + name: "Composite", + className: "DatasetLocation", + additionalProperties: { + type: { + name: "Object" + } + } + } + }, + avroCompressionCodec: { + serializedName: "typeProperties.avroCompressionCodec", + type: { + name: "String" + } + }, + avroCompressionLevel: { + serializedName: "typeProperties.avroCompressionLevel", + constraints: { + InclusiveMaximum: 9, + InclusiveMinimum: 1 + }, + type: { + name: "Number" } } }, @@ -12043,29 +12032,36 @@ export const AzureTableDataset: msRest.CompositeMapper = { } }; -export const AzureBlobDataset: msRest.CompositeMapper = { - serializedName: "AzureBlob", +export const AmazonS3Dataset: msRest.CompositeMapper = { + serializedName: "AmazonS3Object", type: { name: "Composite", polymorphicDiscriminator: Dataset.type.polymorphicDiscriminator, uberParent: "Dataset", - className: "AzureBlobDataset", + className: "AmazonS3Dataset", modelProperties: { ...Dataset.type.modelProperties, - folderPath: { - serializedName: "typeProperties.folderPath", + bucketName: { + required: true, + serializedName: "typeProperties.bucketName", type: { name: "Object" } }, - tableRootLocation: { - serializedName: "typeProperties.tableRootLocation", + key: { + serializedName: "typeProperties.key", type: { name: "Object" } }, - fileName: { - serializedName: "typeProperties.fileName", + prefix: { + serializedName: "typeProperties.prefix", + type: { + name: "Object" + } + }, + version: { + serializedName: "typeProperties.version", type: { name: "Object" } @@ -12101,588 +12097,665 @@ export const AzureBlobDataset: msRest.CompositeMapper = { className: "DatasetCompression", additionalProperties: { type: { - name: "Object" + name: "Object" + } + } + } + } + }, + additionalProperties: Dataset.type.additionalProperties + } +}; + +export const ChainingTrigger: msRest.CompositeMapper = { + serializedName: "ChainingTrigger", + type: { + name: "Composite", + polymorphicDiscriminator: Trigger.type.polymorphicDiscriminator, + uberParent: "Trigger", + className: "ChainingTrigger", + modelProperties: { + ...Trigger.type.modelProperties, + pipelineProperty: { + required: true, + serializedName: "pipeline", + defaultValue: {}, + type: { + name: "Composite", + className: "TriggerPipelineReference" + } + }, + dependsOn: { + required: true, + serializedName: "typeProperties.dependsOn", + type: { + name: "Sequence", + element: { + type: { + name: "Composite", + className: "PipelineReference" } } } + }, + runDimension: { + required: true, + serializedName: "typeProperties.runDimension", + type: { + name: "String" + } } }, - additionalProperties: Dataset.type.additionalProperties + additionalProperties: Trigger.type.additionalProperties } }; -export const DatasetLocation: msRest.CompositeMapper = { - serializedName: "DatasetLocation", +export const RerunTumblingWindowTrigger: msRest.CompositeMapper = { + serializedName: "RerunTumblingWindowTrigger", type: { name: "Composite", - polymorphicDiscriminator: { - serializedName: "type", - clientName: "type" - }, - uberParent: "DatasetLocation", - className: "DatasetLocation", + polymorphicDiscriminator: Trigger.type.polymorphicDiscriminator, + uberParent: "Trigger", + className: "RerunTumblingWindowTrigger", modelProperties: { - folderPath: { - serializedName: "folderPath", + ...Trigger.type.modelProperties, + parentTrigger: { + required: true, + serializedName: "typeProperties.parentTrigger", type: { name: "Object" } }, - fileName: { - serializedName: "fileName", + requestedStartTime: { + required: true, + serializedName: "typeProperties.requestedStartTime", type: { - name: "Object" + name: "DateTime" } }, - type: { + requestedEndTime: { required: true, - serializedName: "type", + serializedName: "typeProperties.requestedEndTime", type: { - name: "String" + name: "DateTime" + } + }, + rerunConcurrency: { + required: true, + serializedName: "typeProperties.rerunConcurrency", + constraints: { + InclusiveMaximum: 50, + InclusiveMinimum: 1 + }, + type: { + name: "Number" } } }, - additionalProperties: { - type: { - name: "Object" - } - } + additionalProperties: Trigger.type.additionalProperties } }; -export const HdfsLocation: msRest.CompositeMapper = { - serializedName: "HdfsLocation", +export const DependencyReference: msRest.CompositeMapper = { + serializedName: "DependencyReference", type: { name: "Composite", - polymorphicDiscriminator: DatasetLocation.type.polymorphicDiscriminator, - uberParent: "DatasetLocation", - className: "HdfsLocation", - modelProperties: { - ...DatasetLocation.type.modelProperties + polymorphicDiscriminator: { + serializedName: "type", + clientName: "type" }, - additionalProperties: DatasetLocation.type.additionalProperties - } -}; - -export const HttpServerLocation: msRest.CompositeMapper = { - serializedName: "HttpServerLocation", - type: { - name: "Composite", - polymorphicDiscriminator: DatasetLocation.type.polymorphicDiscriminator, - uberParent: "DatasetLocation", - className: "HttpServerLocation", + uberParent: "DependencyReference", + className: "DependencyReference", modelProperties: { - ...DatasetLocation.type.modelProperties, - relativeUrl: { - serializedName: "relativeUrl", + type: { + required: true, + serializedName: "type", type: { - name: "Object" + name: "String" } } - }, - additionalProperties: DatasetLocation.type.additionalProperties - } -}; - -export const SftpLocation: msRest.CompositeMapper = { - serializedName: "SftpLocation", - type: { - name: "Composite", - polymorphicDiscriminator: DatasetLocation.type.polymorphicDiscriminator, - uberParent: "DatasetLocation", - className: "SftpLocation", - modelProperties: { - ...DatasetLocation.type.modelProperties - }, - additionalProperties: DatasetLocation.type.additionalProperties - } -}; - -export const FtpServerLocation: msRest.CompositeMapper = { - serializedName: "FtpServerLocation", - type: { - name: "Composite", - polymorphicDiscriminator: DatasetLocation.type.polymorphicDiscriminator, - uberParent: "DatasetLocation", - className: "FtpServerLocation", - modelProperties: { - ...DatasetLocation.type.modelProperties - }, - additionalProperties: DatasetLocation.type.additionalProperties + } } }; -export const GoogleCloudStorageLocation: msRest.CompositeMapper = { - serializedName: "GoogleCloudStorageLocation", +export const SelfDependencyTumblingWindowTriggerReference: msRest.CompositeMapper = { + serializedName: "SelfDependencyTumblingWindowTriggerReference", type: { name: "Composite", - polymorphicDiscriminator: DatasetLocation.type.polymorphicDiscriminator, - uberParent: "DatasetLocation", - className: "GoogleCloudStorageLocation", + polymorphicDiscriminator: DependencyReference.type.polymorphicDiscriminator, + uberParent: "DependencyReference", + className: "SelfDependencyTumblingWindowTriggerReference", modelProperties: { - ...DatasetLocation.type.modelProperties, - bucketName: { - serializedName: "bucketName", + ...DependencyReference.type.modelProperties, + offset: { + required: true, + serializedName: "offset", + constraints: { + MaxLength: 15, + MinLength: 8, + Pattern: /-((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9]))/ + }, type: { - name: "Object" + name: "String" } }, - version: { - serializedName: "version", + size: { + serializedName: "size", + constraints: { + MaxLength: 15, + MinLength: 8, + Pattern: /((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9]))/ + }, type: { - name: "Object" + name: "String" } } - }, - additionalProperties: DatasetLocation.type.additionalProperties - } -}; - -export const AzureFileStorageLocation: msRest.CompositeMapper = { - serializedName: "AzureFileStorageLocation", - type: { - name: "Composite", - polymorphicDiscriminator: DatasetLocation.type.polymorphicDiscriminator, - uberParent: "DatasetLocation", - className: "AzureFileStorageLocation", - modelProperties: { - ...DatasetLocation.type.modelProperties - }, - additionalProperties: DatasetLocation.type.additionalProperties - } -}; - -export const FileServerLocation: msRest.CompositeMapper = { - serializedName: "FileServerLocation", - type: { - name: "Composite", - polymorphicDiscriminator: DatasetLocation.type.polymorphicDiscriminator, - uberParent: "DatasetLocation", - className: "FileServerLocation", - modelProperties: { - ...DatasetLocation.type.modelProperties - }, - additionalProperties: DatasetLocation.type.additionalProperties + } } }; -export const AmazonS3Location: msRest.CompositeMapper = { - serializedName: "AmazonS3Location", +export const TriggerReference: msRest.CompositeMapper = { + serializedName: "TriggerReference", type: { name: "Composite", - polymorphicDiscriminator: DatasetLocation.type.polymorphicDiscriminator, - uberParent: "DatasetLocation", - className: "AmazonS3Location", + className: "TriggerReference", modelProperties: { - ...DatasetLocation.type.modelProperties, - bucketName: { - serializedName: "bucketName", + type: { + required: true, + isConstant: true, + serializedName: "type", + defaultValue: 'TriggerReference', type: { - name: "Object" + name: "String" } }, - version: { - serializedName: "version", + referenceName: { + required: true, + serializedName: "referenceName", type: { - name: "Object" + name: "String" } } - }, - additionalProperties: DatasetLocation.type.additionalProperties - } -}; - -export const AzureDataLakeStoreLocation: msRest.CompositeMapper = { - serializedName: "AzureDataLakeStoreLocation", - type: { - name: "Composite", - polymorphicDiscriminator: DatasetLocation.type.polymorphicDiscriminator, - uberParent: "DatasetLocation", - className: "AzureDataLakeStoreLocation", - modelProperties: { - ...DatasetLocation.type.modelProperties - }, - additionalProperties: DatasetLocation.type.additionalProperties + } } }; -export const AzureBlobFSLocation: msRest.CompositeMapper = { - serializedName: "AzureBlobFSLocation", +export const TriggerDependencyReference: msRest.CompositeMapper = { + serializedName: "TriggerDependencyReference", type: { name: "Composite", - polymorphicDiscriminator: DatasetLocation.type.polymorphicDiscriminator, - uberParent: "DatasetLocation", - className: "AzureBlobFSLocation", + polymorphicDiscriminator: DependencyReference.type.polymorphicDiscriminator, + uberParent: "DependencyReference", + className: "TriggerDependencyReference", modelProperties: { - ...DatasetLocation.type.modelProperties, - fileSystem: { - serializedName: "fileSystem", + ...DependencyReference.type.modelProperties, + referenceTrigger: { + required: true, + serializedName: "referenceTrigger", + defaultValue: {}, type: { - name: "Object" + name: "Composite", + className: "TriggerReference" } } - }, - additionalProperties: DatasetLocation.type.additionalProperties + } } }; -export const AzureBlobStorageLocation: msRest.CompositeMapper = { - serializedName: "AzureBlobStorageLocation", +export const TumblingWindowTriggerDependencyReference: msRest.CompositeMapper = { + serializedName: "TumblingWindowTriggerDependencyReference", type: { name: "Composite", - polymorphicDiscriminator: DatasetLocation.type.polymorphicDiscriminator, - uberParent: "DatasetLocation", - className: "AzureBlobStorageLocation", + polymorphicDiscriminator: DependencyReference.type.polymorphicDiscriminator, + uberParent: "DependencyReference", + className: "TumblingWindowTriggerDependencyReference", modelProperties: { - ...DatasetLocation.type.modelProperties, - container: { - serializedName: "container", + ...TriggerDependencyReference.type.modelProperties, + offset: { + serializedName: "offset", + constraints: { + MaxLength: 15, + MinLength: 8, + Pattern: /-?((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9]))/ + }, type: { - name: "Object" + name: "String" + } + }, + size: { + serializedName: "size", + constraints: { + MaxLength: 15, + MinLength: 8, + Pattern: /((\d+)\.)?(\d\d):(60|([0-5][0-9])):(60|([0-5][0-9]))/ + }, + type: { + name: "String" } } - }, - additionalProperties: DatasetLocation.type.additionalProperties + } } }; -export const BinaryDataset: msRest.CompositeMapper = { - serializedName: "Binary", +export const RetryPolicy: msRest.CompositeMapper = { + serializedName: "RetryPolicy", type: { name: "Composite", - polymorphicDiscriminator: Dataset.type.polymorphicDiscriminator, - uberParent: "Dataset", - className: "BinaryDataset", + className: "RetryPolicy", modelProperties: { - ...Dataset.type.modelProperties, - location: { - required: true, - serializedName: "typeProperties.location", + count: { + serializedName: "count", type: { - name: "Composite", - className: "DatasetLocation", - additionalProperties: { - type: { - name: "Object" - } - } + name: "Object" } }, - compression: { - serializedName: "typeProperties.compression", + intervalInSeconds: { + serializedName: "intervalInSeconds", + constraints: { + InclusiveMaximum: 86400, + InclusiveMinimum: 30 + }, type: { - name: "Composite", - className: "DatasetCompression", - additionalProperties: { - type: { - name: "Object" - } - } + name: "Number" } } - }, - additionalProperties: Dataset.type.additionalProperties + } } }; -export const OrcDataset: msRest.CompositeMapper = { - serializedName: "Orc", +export const TumblingWindowTrigger: msRest.CompositeMapper = { + serializedName: "TumblingWindowTrigger", type: { name: "Composite", - polymorphicDiscriminator: Dataset.type.polymorphicDiscriminator, - uberParent: "Dataset", - className: "OrcDataset", + polymorphicDiscriminator: Trigger.type.polymorphicDiscriminator, + uberParent: "Trigger", + className: "TumblingWindowTrigger", modelProperties: { - ...Dataset.type.modelProperties, - location: { + ...Trigger.type.modelProperties, + pipelineProperty: { required: true, - serializedName: "typeProperties.location", + serializedName: "pipeline", + defaultValue: {}, type: { name: "Composite", - className: "DatasetLocation", - additionalProperties: { - type: { - name: "Object" - } - } + className: "TriggerPipelineReference" } }, - orcCompressionCodec: { - serializedName: "typeProperties.orcCompressionCodec", + frequency: { + required: true, + serializedName: "typeProperties.frequency", type: { name: "String" } - } - }, - additionalProperties: Dataset.type.additionalProperties - } -}; - -export const JsonDataset: msRest.CompositeMapper = { - serializedName: "Json", - type: { - name: "Composite", - polymorphicDiscriminator: Dataset.type.polymorphicDiscriminator, - uberParent: "Dataset", - className: "JsonDataset", - modelProperties: { - ...Dataset.type.modelProperties, - location: { + }, + interval: { required: true, - serializedName: "typeProperties.location", + serializedName: "typeProperties.interval", type: { - name: "Composite", - className: "DatasetLocation", - additionalProperties: { - type: { - name: "Object" - } - } + name: "Number" } }, - encodingName: { - serializedName: "typeProperties.encodingName", + startTime: { + required: true, + serializedName: "typeProperties.startTime", + type: { + name: "DateTime" + } + }, + endTime: { + serializedName: "typeProperties.endTime", + type: { + name: "DateTime" + } + }, + delay: { + serializedName: "typeProperties.delay", type: { name: "Object" } }, - compression: { - serializedName: "typeProperties.compression", + maxConcurrency: { + required: true, + serializedName: "typeProperties.maxConcurrency", + constraints: { + InclusiveMaximum: 50, + InclusiveMinimum: 1 + }, + type: { + name: "Number" + } + }, + retryPolicy: { + serializedName: "typeProperties.retryPolicy", type: { name: "Composite", - className: "DatasetCompression", - additionalProperties: { + className: "RetryPolicy" + } + }, + dependsOn: { + serializedName: "typeProperties.dependsOn", + type: { + name: "Sequence", + element: { type: { - name: "Object" + name: "Composite", + className: "DependencyReference" } } } } }, - additionalProperties: Dataset.type.additionalProperties + additionalProperties: Trigger.type.additionalProperties } }; -export const DelimitedTextDataset: msRest.CompositeMapper = { - serializedName: "DelimitedText", +export const MultiplePipelineTrigger: msRest.CompositeMapper = { + serializedName: "MultiplePipelineTrigger", type: { name: "Composite", - polymorphicDiscriminator: Dataset.type.polymorphicDiscriminator, - uberParent: "Dataset", - className: "DelimitedTextDataset", + polymorphicDiscriminator: Trigger.type.polymorphicDiscriminator, + uberParent: "Trigger", + className: "MultiplePipelineTrigger", modelProperties: { - ...Dataset.type.modelProperties, - location: { - required: true, - serializedName: "typeProperties.location", + ...Trigger.type.modelProperties, + pipelines: { + serializedName: "pipelines", type: { - name: "Composite", - className: "DatasetLocation", - additionalProperties: { + name: "Sequence", + element: { type: { - name: "Object" + name: "Composite", + className: "TriggerPipelineReference" } } } - }, - columnDelimiter: { - serializedName: "typeProperties.columnDelimiter", - type: { - name: "Object" - } - }, - rowDelimiter: { - serializedName: "typeProperties.rowDelimiter", + } + }, + additionalProperties: Trigger.type.additionalProperties + } +}; + +export const BlobEventsTrigger: msRest.CompositeMapper = { + serializedName: "BlobEventsTrigger", + type: { + name: "Composite", + polymorphicDiscriminator: Trigger.type.polymorphicDiscriminator, + uberParent: "Trigger", + className: "BlobEventsTrigger", + modelProperties: { + ...MultiplePipelineTrigger.type.modelProperties, + blobPathBeginsWith: { + serializedName: "typeProperties.blobPathBeginsWith", type: { - name: "Object" + name: "String" } }, - encodingName: { - serializedName: "typeProperties.encodingName", + blobPathEndsWith: { + serializedName: "typeProperties.blobPathEndsWith", type: { - name: "Object" + name: "String" } }, - compressionCodec: { - serializedName: "typeProperties.compressionCodec", + ignoreEmptyBlobs: { + serializedName: "typeProperties.ignoreEmptyBlobs", type: { - name: "Object" + name: "Boolean" } }, - compressionLevel: { - serializedName: "typeProperties.compressionLevel", + events: { + required: true, + serializedName: "typeProperties.events", type: { - name: "Object" + name: "Sequence", + element: { + type: { + name: "String" + } + } } }, - quoteChar: { - serializedName: "typeProperties.quoteChar", + scope: { + required: true, + serializedName: "typeProperties.scope", type: { - name: "Object" + name: "String" } - }, - escapeChar: { - serializedName: "typeProperties.escapeChar", + } + }, + additionalProperties: Trigger.type.additionalProperties + } +}; + +export const BlobTrigger: msRest.CompositeMapper = { + serializedName: "BlobTrigger", + type: { + name: "Composite", + polymorphicDiscriminator: Trigger.type.polymorphicDiscriminator, + uberParent: "Trigger", + className: "BlobTrigger", + modelProperties: { + ...MultiplePipelineTrigger.type.modelProperties, + folderPath: { + required: true, + serializedName: "typeProperties.folderPath", type: { - name: "Object" + name: "String" } }, - firstRowAsHeader: { - serializedName: "typeProperties.firstRowAsHeader", + maxConcurrency: { + required: true, + serializedName: "typeProperties.maxConcurrency", type: { - name: "Object" + name: "Number" } }, - nullValue: { - serializedName: "typeProperties.nullValue", + linkedService: { + required: true, + serializedName: "typeProperties.linkedService", + defaultValue: {}, type: { - name: "Object" + name: "Composite", + className: "LinkedServiceReference" } } }, - additionalProperties: Dataset.type.additionalProperties + additionalProperties: Trigger.type.additionalProperties } }; -export const ParquetDataset: msRest.CompositeMapper = { - serializedName: "Parquet", +export const RecurrenceScheduleOccurrence: msRest.CompositeMapper = { + serializedName: "RecurrenceScheduleOccurrence", type: { name: "Composite", - polymorphicDiscriminator: Dataset.type.polymorphicDiscriminator, - uberParent: "Dataset", - className: "ParquetDataset", + className: "RecurrenceScheduleOccurrence", modelProperties: { - ...Dataset.type.modelProperties, - location: { - required: true, - serializedName: "typeProperties.location", + day: { + serializedName: "day", type: { - name: "Composite", - className: "DatasetLocation", - additionalProperties: { - type: { - name: "Object" - } - } + name: "Enum", + allowedValues: [ + "Sunday", + "Monday", + "Tuesday", + "Wednesday", + "Thursday", + "Friday", + "Saturday" + ] } }, - compressionCodec: { - serializedName: "typeProperties.compressionCodec", + occurrence: { + serializedName: "occurrence", type: { - name: "Object" + name: "Number" } } }, - additionalProperties: Dataset.type.additionalProperties + additionalProperties: { + type: { + name: "Object" + } + } } }; -export const AvroDataset: msRest.CompositeMapper = { - serializedName: "Avro", +export const RecurrenceSchedule: msRest.CompositeMapper = { + serializedName: "RecurrenceSchedule", type: { name: "Composite", - polymorphicDiscriminator: Dataset.type.polymorphicDiscriminator, - uberParent: "Dataset", - className: "AvroDataset", + className: "RecurrenceSchedule", modelProperties: { - ...Dataset.type.modelProperties, - location: { - required: true, - serializedName: "typeProperties.location", + minutes: { + serializedName: "minutes", type: { - name: "Composite", - className: "DatasetLocation", - additionalProperties: { + name: "Sequence", + element: { type: { - name: "Object" + name: "Number" } } } }, - avroCompressionCodec: { - serializedName: "typeProperties.avroCompressionCodec", + hours: { + serializedName: "hours", type: { - name: "String" + name: "Sequence", + element: { + type: { + name: "Number" + } + } } }, - avroCompressionLevel: { - serializedName: "typeProperties.avroCompressionLevel", - constraints: { - InclusiveMaximum: 9, - InclusiveMinimum: 1 - }, + weekDays: { + serializedName: "weekDays", type: { - name: "Number" + name: "Sequence", + element: { + type: { + name: "Enum", + allowedValues: [ + "Sunday", + "Monday", + "Tuesday", + "Wednesday", + "Thursday", + "Friday", + "Saturday" + ] + } + } + } + }, + monthDays: { + serializedName: "monthDays", + type: { + name: "Sequence", + element: { + type: { + name: "Number" + } + } + } + }, + monthlyOccurrences: { + serializedName: "monthlyOccurrences", + type: { + name: "Sequence", + element: { + type: { + name: "Composite", + className: "RecurrenceScheduleOccurrence", + additionalProperties: { + type: { + name: "Object" + } + } + } + } } } }, - additionalProperties: Dataset.type.additionalProperties + additionalProperties: { + type: { + name: "Object" + } + } } }; -export const AmazonS3Dataset: msRest.CompositeMapper = { - serializedName: "AmazonS3Object", +export const ScheduleTriggerRecurrence: msRest.CompositeMapper = { + serializedName: "ScheduleTriggerRecurrence", type: { name: "Composite", - polymorphicDiscriminator: Dataset.type.polymorphicDiscriminator, - uberParent: "Dataset", - className: "AmazonS3Dataset", + className: "ScheduleTriggerRecurrence", modelProperties: { - ...Dataset.type.modelProperties, - bucketName: { - required: true, - serializedName: "typeProperties.bucketName", - type: { - name: "Object" - } - }, - key: { - serializedName: "typeProperties.key", + frequency: { + serializedName: "frequency", type: { - name: "Object" + name: "String" } }, - prefix: { - serializedName: "typeProperties.prefix", + interval: { + serializedName: "interval", type: { - name: "Object" + name: "Number" } }, - version: { - serializedName: "typeProperties.version", + startTime: { + serializedName: "startTime", type: { - name: "Object" + name: "DateTime" } }, - modifiedDatetimeStart: { - serializedName: "typeProperties.modifiedDatetimeStart", + endTime: { + serializedName: "endTime", type: { - name: "Object" + name: "DateTime" } }, - modifiedDatetimeEnd: { - serializedName: "typeProperties.modifiedDatetimeEnd", + timeZone: { + serializedName: "timeZone", type: { - name: "Object" + name: "String" } }, - format: { - serializedName: "typeProperties.format", + schedule: { + serializedName: "schedule", type: { name: "Composite", - className: "DatasetStorageFormat", + className: "RecurrenceSchedule", additionalProperties: { type: { name: "Object" } } } - }, - compression: { - serializedName: "typeProperties.compression", + } + }, + additionalProperties: { + type: { + name: "Object" + } + } + } +}; + +export const ScheduleTrigger: msRest.CompositeMapper = { + serializedName: "ScheduleTrigger", + type: { + name: "Composite", + polymorphicDiscriminator: Trigger.type.polymorphicDiscriminator, + uberParent: "Trigger", + className: "ScheduleTrigger", + modelProperties: { + ...MultiplePipelineTrigger.type.modelProperties, + recurrence: { + required: true, + serializedName: "typeProperties.recurrence", type: { name: "Composite", - className: "DatasetCompression", + className: "ScheduleTriggerRecurrence", additionalProperties: { type: { name: "Object" @@ -12691,7 +12764,7 @@ export const AmazonS3Dataset: msRest.CompositeMapper = { } } }, - additionalProperties: Dataset.type.additionalProperties + additionalProperties: Trigger.type.additionalProperties } }; @@ -13409,6 +13482,69 @@ export const WebActivity: msRest.CompositeMapper = { } }; +export const ExportSettings: msRest.CompositeMapper = { + serializedName: "ExportSettings", + type: { + name: "Composite", + polymorphicDiscriminator: { + serializedName: "type", + clientName: "type" + }, + uberParent: "ExportSettings", + className: "ExportSettings", + modelProperties: { + type: { + required: true, + serializedName: "type", + type: { + name: "String" + } + } + }, + additionalProperties: { + type: { + name: "Object" + } + } + } +}; + +export const SnowflakeExportCopyCommand: msRest.CompositeMapper = { + serializedName: "SnowflakeExportCopyCommand", + type: { + name: "Composite", + polymorphicDiscriminator: ExportSettings.type.polymorphicDiscriminator, + uberParent: "ExportSettings", + className: "SnowflakeExportCopyCommand", + modelProperties: { + ...ExportSettings.type.modelProperties, + additionalCopyOptions: { + serializedName: "additionalCopyOptions", + type: { + name: "Dictionary", + value: { + type: { + name: "Object" + } + } + } + }, + additionalFormatOptions: { + serializedName: "additionalFormatOptions", + type: { + name: "Dictionary", + value: { + type: { + name: "Object" + } + } + } + } + }, + additionalProperties: ExportSettings.type.additionalProperties + } +}; + export const CopySource: msRest.CompositeMapper = { serializedName: "CopySource", type: { @@ -13454,6 +13590,34 @@ export const CopySource: msRest.CompositeMapper = { } }; +export const SnowflakeSource: msRest.CompositeMapper = { + serializedName: "SnowflakeSource", + type: { + name: "Composite", + polymorphicDiscriminator: CopySource.type.polymorphicDiscriminator, + uberParent: "CopySource", + className: "SnowflakeSource", + modelProperties: { + ...CopySource.type.modelProperties, + query: { + serializedName: "query", + type: { + name: "Object" + } + }, + exportSettings: { + serializedName: "exportSettings", + type: { + name: "Composite", + className: "SnowflakeExportCopyCommand", + additionalProperties: ExportSettings.type.additionalProperties + } + } + }, + additionalProperties: CopySource.type.additionalProperties + } +}; + export const HttpSource: msRest.CompositeMapper = { serializedName: "HttpSource", type: { @@ -13576,6 +13740,28 @@ export const Office365Source: msRest.CompositeMapper = { } }; +export const AdditionalColumns: msRest.CompositeMapper = { + serializedName: "AdditionalColumns", + type: { + name: "Composite", + className: "AdditionalColumns", + modelProperties: { + name: { + serializedName: "name", + type: { + name: "Object" + } + }, + value: { + serializedName: "value", + type: { + name: "Object" + } + } + } + } +}; + export const MongoDbCursorMethodsProperties: msRest.CompositeMapper = { serializedName: "MongoDbCursorMethodsProperties", type: { @@ -13653,6 +13839,18 @@ export const CosmosDbMongoDbApiSource: msRest.CompositeMapper = { type: { name: "Object" } + }, + additionalColumns: { + serializedName: "additionalColumns", + type: { + name: "Sequence", + element: { + type: { + name: "Composite", + className: "AdditionalColumns" + } + } + } } }, additionalProperties: CopySource.type.additionalProperties @@ -13697,6 +13895,18 @@ export const MongoDbV2Source: msRest.CompositeMapper = { type: { name: "Object" } + }, + additionalColumns: { + serializedName: "additionalColumns", + type: { + name: "Sequence", + element: { + type: { + name: "Composite", + className: "AdditionalColumns" + } + } + } } }, additionalProperties: CopySource.type.additionalProperties @@ -13717,6 +13927,18 @@ export const MongoDbSource: msRest.CompositeMapper = { type: { name: "Object" } + }, + additionalColumns: { + serializedName: "additionalColumns", + type: { + name: "Sequence", + element: { + type: { + name: "Composite", + className: "AdditionalColumns" + } + } + } } }, additionalProperties: CopySource.type.additionalProperties @@ -13731,7 +13953,19 @@ export const WebSource: msRest.CompositeMapper = { uberParent: "CopySource", className: "WebSource", modelProperties: { - ...CopySource.type.modelProperties + ...CopySource.type.modelProperties, + additionalColumns: { + serializedName: "additionalColumns", + type: { + name: "Sequence", + element: { + type: { + name: "Composite", + className: "AdditionalColumns" + } + } + } + } }, additionalProperties: CopySource.type.additionalProperties } @@ -13804,6 +14038,18 @@ export const OracleSource: msRest.CompositeMapper = { name: "Composite", className: "OraclePartitionSettings" } + }, + additionalColumns: { + serializedName: "additionalColumns", + type: { + name: "Sequence", + element: { + type: { + name: "Composite", + className: "AdditionalColumns" + } + } + } } }, additionalProperties: CopySource.type.additionalProperties @@ -13837,6 +14083,18 @@ export const AzureDataExplorerSource: msRest.CompositeMapper = { type: { name: "Object" } + }, + additionalColumns: { + serializedName: "additionalColumns", + type: { + name: "Sequence", + element: { + type: { + name: "Composite", + className: "AdditionalColumns" + } + } + } } }, additionalProperties: CopySource.type.additionalProperties @@ -13914,6 +14172,18 @@ export const FileSystemSource: msRest.CompositeMapper = { type: { name: "Object" } + }, + additionalColumns: { + serializedName: "additionalColumns", + type: { + name: "Sequence", + element: { + type: { + name: "Composite", + className: "AdditionalColumns" + } + } + } } }, additionalProperties: CopySource.type.additionalProperties @@ -13964,6 +14234,18 @@ export const RestSource: msRest.CompositeMapper = { type: { name: "Object" } + }, + additionalColumns: { + serializedName: "additionalColumns", + type: { + name: "Sequence", + element: { + type: { + name: "Composite", + className: "AdditionalColumns" + } + } + } } }, additionalProperties: CopySource.type.additionalProperties @@ -13990,6 +14272,18 @@ export const SalesforceServiceCloudSource: msRest.CompositeMapper = { type: { name: "String" } + }, + additionalColumns: { + serializedName: "additionalColumns", + type: { + name: "Sequence", + element: { + type: { + name: "Composite", + className: "AdditionalColumns" + } + } + } } }, additionalProperties: CopySource.type.additionalProperties @@ -14010,6 +14304,18 @@ export const ODataSource: msRest.CompositeMapper = { type: { name: "Object" } + }, + additionalColumns: { + serializedName: "additionalColumns", + type: { + name: "Sequence", + element: { + type: { + name: "Composite", + className: "AdditionalColumns" + } + } + } } }, additionalProperties: CopySource.type.additionalProperties @@ -14030,6 +14336,18 @@ export const MicrosoftAccessSource: msRest.CompositeMapper = { type: { name: "Object" } + }, + additionalColumns: { + serializedName: "additionalColumns", + type: { + name: "Sequence", + element: { + type: { + name: "Composite", + className: "AdditionalColumns" + } + } + } } }, additionalProperties: CopySource.type.additionalProperties @@ -14050,6 +14368,18 @@ export const RelationalSource: msRest.CompositeMapper = { type: { name: "Object" } + }, + additionalColumns: { + serializedName: "additionalColumns", + type: { + name: "Sequence", + element: { + type: { + name: "Composite", + className: "AdditionalColumns" + } + } + } } }, additionalProperties: CopySource.type.additionalProperties @@ -14070,6 +14400,18 @@ export const CommonDataServiceForAppsSource: msRest.CompositeMapper = { type: { name: "Object" } + }, + additionalColumns: { + serializedName: "additionalColumns", + type: { + name: "Sequence", + element: { + type: { + name: "Composite", + className: "AdditionalColumns" + } + } + } } }, additionalProperties: CopySource.type.additionalProperties @@ -14090,6 +14432,18 @@ export const DynamicsCrmSource: msRest.CompositeMapper = { type: { name: "Object" } + }, + additionalColumns: { + serializedName: "additionalColumns", + type: { + name: "Sequence", + element: { + type: { + name: "Composite", + className: "AdditionalColumns" + } + } + } } }, additionalProperties: CopySource.type.additionalProperties @@ -14110,6 +14464,18 @@ export const DynamicsSource: msRest.CompositeMapper = { type: { name: "Object" } + }, + additionalColumns: { + serializedName: "additionalColumns", + type: { + name: "Sequence", + element: { + type: { + name: "Composite", + className: "AdditionalColumns" + } + } + } } }, additionalProperties: CopySource.type.additionalProperties @@ -14142,6 +14508,18 @@ export const CosmosDbSqlApiSource: msRest.CompositeMapper = { type: { name: "Object" } + }, + additionalColumns: { + serializedName: "additionalColumns", + type: { + name: "Sequence", + element: { + type: { + name: "Composite", + className: "AdditionalColumns" + } + } + } } }, additionalProperties: CopySource.type.additionalProperties @@ -14174,6 +14552,18 @@ export const DocumentDbCollectionSource: msRest.CompositeMapper = { type: { name: "Object" } + }, + additionalColumns: { + serializedName: "additionalColumns", + type: { + name: "Sequence", + element: { + type: { + name: "Composite", + className: "AdditionalColumns" + } + } + } } }, additionalProperties: CopySource.type.additionalProperties @@ -14252,6 +14642,18 @@ export const TabularSource: msRest.CompositeMapper = { type: { name: "Object" } + }, + additionalColumns: { + serializedName: "additionalColumns", + type: { + name: "Sequence", + element: { + type: { + name: "Composite", + className: "AdditionalColumns" + } + } + } } }, additionalProperties: CopySource.type.additionalProperties @@ -15331,6 +15733,12 @@ export const SqlSource: msRest.CompositeMapper = { } } } + }, + isolationLevel: { + serializedName: "isolationLevel", + type: { + name: "Object" + } } }, additionalProperties: CopySource.type.additionalProperties @@ -15807,6 +16215,12 @@ export const HdfsReadSettings: msRest.CompositeMapper = { name: "Object" } }, + fileListPath: { + serializedName: "fileListPath", + type: { + name: "Object" + } + }, enablePartitionDiscovery: { serializedName: "enablePartitionDiscovery", type: { @@ -15902,6 +16316,12 @@ export const SftpReadSettings: msRest.CompositeMapper = { name: "Object" } }, + fileListPath: { + serializedName: "fileListPath", + type: { + name: "Object" + } + }, modifiedDatetimeStart: { serializedName: "modifiedDatetimeStart", type: { @@ -15946,6 +16366,12 @@ export const FtpReadSettings: msRest.CompositeMapper = { name: "Object" } }, + fileListPath: { + serializedName: "fileListPath", + type: { + name: "Object" + } + }, useBinaryTransfer: { serializedName: "useBinaryTransfer", type: { @@ -15990,6 +16416,12 @@ export const GoogleCloudStorageReadSettings: msRest.CompositeMapper = { name: "Object" } }, + fileListPath: { + serializedName: "fileListPath", + type: { + name: "Object" + } + }, enablePartitionDiscovery: { serializedName: "enablePartitionDiscovery", type: { @@ -16040,6 +16472,12 @@ export const AzureFileStorageReadSettings: msRest.CompositeMapper = { name: "Object" } }, + fileListPath: { + serializedName: "fileListPath", + type: { + name: "Object" + } + }, enablePartitionDiscovery: { serializedName: "enablePartitionDiscovery", type: { @@ -16090,6 +16528,12 @@ export const FileServerReadSettings: msRest.CompositeMapper = { name: "Object" } }, + fileListPath: { + serializedName: "fileListPath", + type: { + name: "Object" + } + }, enablePartitionDiscovery: { serializedName: "enablePartitionDiscovery", type: { @@ -16146,6 +16590,12 @@ export const AmazonS3ReadSettings: msRest.CompositeMapper = { name: "Object" } }, + fileListPath: { + serializedName: "fileListPath", + type: { + name: "Object" + } + }, enablePartitionDiscovery: { serializedName: "enablePartitionDiscovery", type: { @@ -16196,6 +16646,12 @@ export const AzureDataLakeStoreReadSettings: msRest.CompositeMapper = { name: "Object" } }, + fileListPath: { + serializedName: "fileListPath", + type: { + name: "Object" + } + }, enablePartitionDiscovery: { serializedName: "enablePartitionDiscovery", type: { @@ -16246,6 +16702,12 @@ export const AzureBlobFSReadSettings: msRest.CompositeMapper = { name: "Object" } }, + fileListPath: { + serializedName: "fileListPath", + type: { + name: "Object" + } + }, enablePartitionDiscovery: { serializedName: "enablePartitionDiscovery", type: { @@ -16302,6 +16764,12 @@ export const AzureBlobStorageReadSettings: msRest.CompositeMapper = { name: "Object" } }, + fileListPath: { + serializedName: "fileListPath", + type: { + name: "Object" + } + }, enablePartitionDiscovery: { serializedName: "enablePartitionDiscovery", type: { @@ -16371,6 +16839,18 @@ export const OrcSource: msRest.CompositeMapper = { } } } + }, + additionalColumns: { + serializedName: "additionalColumns", + type: { + name: "Sequence", + element: { + type: { + name: "Composite", + className: "AdditionalColumns" + } + } + } } }, additionalProperties: CopySource.type.additionalProperties @@ -16397,6 +16877,18 @@ export const JsonSource: msRest.CompositeMapper = { } } } + }, + additionalColumns: { + serializedName: "additionalColumns", + type: { + name: "Sequence", + element: { + type: { + name: "Composite", + className: "AdditionalColumns" + } + } + } } }, additionalProperties: CopySource.type.additionalProperties @@ -16478,6 +16970,18 @@ export const DelimitedTextSource: msRest.CompositeMapper = { className: "DelimitedTextReadSettings", additionalProperties: FormatReadSettings.type.additionalProperties } + }, + additionalColumns: { + serializedName: "additionalColumns", + type: { + name: "Sequence", + element: { + type: { + name: "Composite", + className: "AdditionalColumns" + } + } + } } }, additionalProperties: CopySource.type.additionalProperties @@ -16504,6 +17008,18 @@ export const ParquetSource: msRest.CompositeMapper = { } } } + }, + additionalColumns: { + serializedName: "additionalColumns", + type: { + name: "Sequence", + element: { + type: { + name: "Composite", + className: "AdditionalColumns" + } + } + } } }, additionalProperties: CopySource.type.additionalProperties @@ -16530,6 +17046,18 @@ export const AvroSource: msRest.CompositeMapper = { } } } + }, + additionalColumns: { + serializedName: "additionalColumns", + type: { + name: "Sequence", + element: { + type: { + name: "Composite", + className: "AdditionalColumns" + } + } + } } }, additionalProperties: CopySource.type.additionalProperties @@ -17625,6 +18153,28 @@ export const HDInsightHiveActivity: msRest.CompositeMapper = { } }; +export const SkipErrorFile: msRest.CompositeMapper = { + serializedName: "SkipErrorFile", + type: { + name: "Composite", + className: "SkipErrorFile", + modelProperties: { + fileMissing: { + serializedName: "fileMissing", + type: { + name: "Object" + } + }, + dataInconsistency: { + serializedName: "dataInconsistency", + type: { + name: "Object" + } + } + } + } +}; + export const RedirectIncompatibleRowSettings: msRest.CompositeMapper = { serializedName: "RedirectIncompatibleRowSettings", type: { @@ -18113,6 +18663,97 @@ export const OracleSink: msRest.CompositeMapper = { } }; +export const ImportSettings: msRest.CompositeMapper = { + serializedName: "ImportSettings", + type: { + name: "Composite", + polymorphicDiscriminator: { + serializedName: "type", + clientName: "type" + }, + uberParent: "ImportSettings", + className: "ImportSettings", + modelProperties: { + type: { + required: true, + serializedName: "type", + type: { + name: "String" + } + } + }, + additionalProperties: { + type: { + name: "Object" + } + } + } +}; + +export const SnowflakeImportCopyCommand: msRest.CompositeMapper = { + serializedName: "SnowflakeImportCopyCommand", + type: { + name: "Composite", + polymorphicDiscriminator: ImportSettings.type.polymorphicDiscriminator, + uberParent: "ImportSettings", + className: "SnowflakeImportCopyCommand", + modelProperties: { + ...ImportSettings.type.modelProperties, + additionalCopyOptions: { + serializedName: "additionalCopyOptions", + type: { + name: "Dictionary", + value: { + type: { + name: "Object" + } + } + } + }, + additionalFormatOptions: { + serializedName: "additionalFormatOptions", + type: { + name: "Dictionary", + value: { + type: { + name: "Object" + } + } + } + } + }, + additionalProperties: ImportSettings.type.additionalProperties + } +}; + +export const SnowflakeSink: msRest.CompositeMapper = { + serializedName: "SnowflakeSink", + type: { + name: "Composite", + polymorphicDiscriminator: CopySink.type.polymorphicDiscriminator, + uberParent: "CopySink", + className: "SnowflakeSink", + modelProperties: { + ...CopySink.type.modelProperties, + preCopyScript: { + serializedName: "preCopyScript", + type: { + name: "Object" + } + }, + importSettings: { + serializedName: "importSettings", + type: { + name: "Composite", + className: "SnowflakeImportCopyCommand", + additionalProperties: ImportSettings.type.additionalProperties + } + } + }, + additionalProperties: CopySink.type.additionalProperties + } +}; + export const DWCopyCommandDefaultValue: msRest.CompositeMapper = { serializedName: "DWCopyCommandDefaultValue", type: { @@ -18653,7 +19294,13 @@ export const AzureDataLakeStoreWriteSettings: msRest.CompositeMapper = { uberParent: "StoreWriteSettings", className: "AzureDataLakeStoreWriteSettings", modelProperties: { - ...StoreWriteSettings.type.modelProperties + ...StoreWriteSettings.type.modelProperties, + expiryDateTime: { + serializedName: "expiryDateTime", + type: { + name: "Object" + } + } }, additionalProperties: StoreWriteSettings.type.additionalProperties } @@ -18713,6 +19360,12 @@ export const SftpWriteSettings: msRest.CompositeMapper = { type: { name: "Object" } + }, + useTempFileRename: { + serializedName: "useTempFileRename", + type: { + name: "Object" + } } }, additionalProperties: StoreWriteSettings.type.additionalProperties @@ -19200,6 +19853,18 @@ export const CopyActivity: msRest.CompositeMapper = { } } }, + logStorageSettings: { + serializedName: "typeProperties.logStorageSettings", + type: { + name: "Composite", + className: "LogStorageSettings", + additionalProperties: { + type: { + name: "Object" + } + } + } + }, preserveRules: { serializedName: "typeProperties.preserveRules", type: { @@ -19222,6 +19887,19 @@ export const CopyActivity: msRest.CompositeMapper = { } } }, + validateDataConsistency: { + serializedName: "typeProperties.validateDataConsistency", + type: { + name: "Object" + } + }, + skipErrorFile: { + serializedName: "typeProperties.skipErrorFile", + type: { + name: "Composite", + className: "SkipErrorFile" + } + }, inputs: { serializedName: "inputs", type: { @@ -21730,36 +22408,6 @@ export const TriggerListResponse: msRest.CompositeMapper = { } }; -export const RerunTriggerListResponse: msRest.CompositeMapper = { - serializedName: "RerunTriggerListResponse", - type: { - name: "Composite", - className: "RerunTriggerListResponse", - modelProperties: { - value: { - required: true, - serializedName: "", - type: { - name: "Sequence", - element: { - type: { - name: "Composite", - className: "RerunTriggerResource" - } - } - } - }, - nextLink: { - readOnly: true, - serializedName: "nextLink", - type: { - name: "String" - } - } - } - } -}; - export const DataFlowListResponse: msRest.CompositeMapper = { serializedName: "DataFlowListResponse", type: { @@ -21835,19 +22483,9 @@ export const discriminators = { 'Trigger' : Trigger, 'FactoryRepoConfiguration.FactoryVSTSConfiguration' : FactoryVSTSConfiguration, 'FactoryRepoConfiguration.FactoryGitHubConfiguration' : FactoryGitHubConfiguration, - 'Trigger.RerunTumblingWindowTrigger' : RerunTumblingWindowTrigger, 'DataFlow' : DataFlow, 'DataFlow.MappingDataFlow' : MappingDataFlow, - 'Trigger.ChainingTrigger' : ChainingTrigger, - 'DependencyReference.SelfDependencyTumblingWindowTriggerReference' : SelfDependencyTumblingWindowTriggerReference, - 'DependencyReference.TumblingWindowTriggerDependencyReference' : TumblingWindowTriggerDependencyReference, - 'DependencyReference.TriggerDependencyReference' : TriggerDependencyReference, - 'DependencyReference' : DependencyReference, - 'Trigger.TumblingWindowTrigger' : TumblingWindowTrigger, - 'Trigger.BlobEventsTrigger' : BlobEventsTrigger, - 'Trigger.BlobTrigger' : BlobTrigger, - 'Trigger.ScheduleTrigger' : ScheduleTrigger, - 'Trigger.MultiplePipelineTrigger' : MultiplePipelineTrigger, + 'LinkedService.Snowflake' : SnowflakeLinkedService, 'LinkedService.AzureFunction' : AzureFunctionLinkedService, 'LinkedService.AzureDataExplorer' : AzureDataExplorerLinkedService, 'LinkedService.SapTable' : SapTableLinkedService, @@ -21946,6 +22584,7 @@ export const discriminators = { 'LinkedService.AzureTableStorage' : AzureTableStorageLinkedService, 'LinkedService.AzureBlobStorage' : AzureBlobStorageLinkedService, 'LinkedService.AzureStorage' : AzureStorageLinkedService, + 'Dataset.SnowflakeTable' : SnowflakeDataset, 'Dataset.GoogleAdWordsObject' : GoogleAdWordsObjectDataset, 'Dataset.AzureDataExplorerTable' : AzureDataExplorerTableDataset, 'Dataset.OracleServiceCloudObject' : OracleServiceCloudObjectDataset, @@ -22056,6 +22695,17 @@ export const discriminators = { 'Dataset.Parquet' : ParquetDataset, 'Dataset.Avro' : AvroDataset, 'Dataset.AmazonS3Object' : AmazonS3Dataset, + 'Trigger.ChainingTrigger' : ChainingTrigger, + 'Trigger.RerunTumblingWindowTrigger' : RerunTumblingWindowTrigger, + 'DependencyReference.SelfDependencyTumblingWindowTriggerReference' : SelfDependencyTumblingWindowTriggerReference, + 'DependencyReference.TumblingWindowTriggerDependencyReference' : TumblingWindowTriggerDependencyReference, + 'DependencyReference.TriggerDependencyReference' : TriggerDependencyReference, + 'DependencyReference' : DependencyReference, + 'Trigger.TumblingWindowTrigger' : TumblingWindowTrigger, + 'Trigger.BlobEventsTrigger' : BlobEventsTrigger, + 'Trigger.BlobTrigger' : BlobTrigger, + 'Trigger.ScheduleTrigger' : ScheduleTrigger, + 'Trigger.MultiplePipelineTrigger' : MultiplePipelineTrigger, 'Activity.ExecuteDataFlow' : ExecuteDataFlowActivity, 'Activity.AzureFunctionActivity' : AzureFunctionActivity, 'Activity.DatabricksSparkPython' : DatabricksSparkPythonActivity, @@ -22067,6 +22717,9 @@ export const discriminators = { 'Activity.AzureMLBatchExecution' : AzureMLBatchExecutionActivity, 'Activity.GetMetadata' : GetMetadataActivity, 'Activity.WebActivity' : WebActivity, + 'ExportSettings' : ExportSettings, + 'ExportSettings.SnowflakeExportCopyCommand' : SnowflakeExportCopyCommand, + 'CopySource.SnowflakeSource' : SnowflakeSource, 'CopySource.HttpSource' : HttpSource, 'CopySource.AzureBlobFSSource' : AzureBlobFSSource, 'CopySource.AzureDataLakeStoreSource' : AzureDataLakeStoreSource, @@ -22194,6 +22847,9 @@ export const discriminators = { 'CopySink.AzureBlobFSSink' : AzureBlobFSSink, 'CopySink.AzureDataLakeStoreSink' : AzureDataLakeStoreSink, 'CopySink.OracleSink' : OracleSink, + 'ImportSettings' : ImportSettings, + 'ImportSettings.SnowflakeImportCopyCommand' : SnowflakeImportCopyCommand, + 'CopySink.SnowflakeSink' : SnowflakeSink, 'CopySink.SqlDWSink' : SqlDWSink, 'CopySink.SqlMISink' : SqlMISink, 'CopySink.AzureSqlSink' : AzureSqlSink, diff --git a/sdk/datafactory/arm-datafactory/src/models/parameters.ts b/sdk/datafactory/arm-datafactory/src/models/parameters.ts index aafb78063d22..f636c40d2c55 100644 --- a/sdk/datafactory/arm-datafactory/src/models/parameters.ts +++ b/sdk/datafactory/arm-datafactory/src/models/parameters.ts @@ -216,21 +216,6 @@ export const referencePipelineRunId: msRest.OperationQueryParameter = { } } }; -export const rerunTriggerName: msRest.OperationURLParameter = { - parameterPath: "rerunTriggerName", - mapper: { - required: true, - serializedName: "rerunTriggerName", - constraints: { - MaxLength: 260, - MinLength: 1, - Pattern: /^[A-Za-z0-9_][^<>*#.%&:\\+?\/]*$/ - }, - type: { - name: "String" - } - } -}; export const resourceGroupName: msRest.OperationURLParameter = { parameterPath: "resourceGroupName", mapper: { @@ -268,6 +253,18 @@ export const startActivityName: msRest.OperationQueryParameter = { } } }; +export const startFromFailure: msRest.OperationQueryParameter = { + parameterPath: [ + "options", + "startFromFailure" + ], + mapper: { + serializedName: "startFromFailure", + type: { + name: "Boolean" + } + } +}; export const subscriptionId: msRest.OperationURLParameter = { parameterPath: "subscriptionId", mapper: { diff --git a/sdk/datafactory/arm-datafactory/src/models/pipelinesMappers.ts b/sdk/datafactory/arm-datafactory/src/models/pipelinesMappers.ts index 86d02b87ff9f..16f592add8b6 100644 --- a/sdk/datafactory/arm-datafactory/src/models/pipelinesMappers.ts +++ b/sdk/datafactory/arm-datafactory/src/models/pipelinesMappers.ts @@ -11,6 +11,7 @@ export { Activity, ActivityDependency, ActivityPolicy, + AdditionalColumns, AmazonMWSLinkedService, AmazonMWSObjectDataset, AmazonMWSSource, @@ -200,6 +201,7 @@ export { ExecutePipelineActivity, ExecuteSSISPackageActivity, ExecutionActivity, + ExportSettings, Expression, Factory, FactoryGitHubConfiguration, @@ -262,6 +264,7 @@ export { ImpalaLinkedService, ImpalaObjectDataset, ImpalaSource, + ImportSettings, InformixLinkedService, InformixSink, InformixSource, @@ -375,7 +378,6 @@ export { RedshiftUnloadSettings, RelationalSource, RelationalTableDataset, - RerunTriggerResource, RerunTumblingWindowTrigger, Resource, ResponsysLinkedService, @@ -435,6 +437,13 @@ export { ShopifyLinkedService, ShopifyObjectDataset, ShopifySource, + SkipErrorFile, + SnowflakeDataset, + SnowflakeExportCopyCommand, + SnowflakeImportCopyCommand, + SnowflakeLinkedService, + SnowflakeSink, + SnowflakeSource, SparkLinkedService, SparkObjectDataset, SparkSource, diff --git a/sdk/datafactory/arm-datafactory/src/models/rerunTriggersMappers.ts b/sdk/datafactory/arm-datafactory/src/models/rerunTriggersMappers.ts deleted file mode 100644 index f02c88a4f465..000000000000 --- a/sdk/datafactory/arm-datafactory/src/models/rerunTriggersMappers.ts +++ /dev/null @@ -1,510 +0,0 @@ -/* - * Copyright (c) Microsoft Corporation. All rights reserved. - * Licensed under the MIT License. See License.txt in the project root for license information. - * - * Code generated by Microsoft (R) AutoRest Code Generator. - * Changes may cause incorrect behavior and will be lost if the code is regenerated. - */ - -export { - discriminators, - Activity, - ActivityDependency, - ActivityPolicy, - AmazonMWSLinkedService, - AmazonMWSObjectDataset, - AmazonMWSSource, - AmazonRedshiftLinkedService, - AmazonRedshiftSource, - AmazonRedshiftTableDataset, - AmazonS3Dataset, - AmazonS3LinkedService, - AmazonS3Location, - AmazonS3ReadSettings, - AppendVariableActivity, - AvroDataset, - AvroFormat, - AvroSink, - AvroSource, - AvroWriteSettings, - AzureBatchLinkedService, - AzureBlobDataset, - AzureBlobFSDataset, - AzureBlobFSLinkedService, - AzureBlobFSLocation, - AzureBlobFSReadSettings, - AzureBlobFSSink, - AzureBlobFSSource, - AzureBlobFSWriteSettings, - AzureBlobStorageLinkedService, - AzureBlobStorageLocation, - AzureBlobStorageReadSettings, - AzureBlobStorageWriteSettings, - AzureDatabricksLinkedService, - AzureDataExplorerCommandActivity, - AzureDataExplorerLinkedService, - AzureDataExplorerSink, - AzureDataExplorerSource, - AzureDataExplorerTableDataset, - AzureDataLakeAnalyticsLinkedService, - AzureDataLakeStoreDataset, - AzureDataLakeStoreLinkedService, - AzureDataLakeStoreLocation, - AzureDataLakeStoreReadSettings, - AzureDataLakeStoreSink, - AzureDataLakeStoreSource, - AzureDataLakeStoreWriteSettings, - AzureFileStorageLinkedService, - AzureFileStorageLocation, - AzureFileStorageReadSettings, - AzureFunctionActivity, - AzureFunctionLinkedService, - AzureKeyVaultLinkedService, - AzureKeyVaultSecretReference, - AzureMariaDBLinkedService, - AzureMariaDBSource, - AzureMariaDBTableDataset, - AzureMLBatchExecutionActivity, - AzureMLExecutePipelineActivity, - AzureMLLinkedService, - AzureMLServiceLinkedService, - AzureMLUpdateResourceActivity, - AzureMLWebServiceFile, - AzureMySqlLinkedService, - AzureMySqlSink, - AzureMySqlSource, - AzureMySqlTableDataset, - AzurePostgreSqlLinkedService, - AzurePostgreSqlSink, - AzurePostgreSqlSource, - AzurePostgreSqlTableDataset, - AzureQueueSink, - AzureSearchIndexDataset, - AzureSearchIndexSink, - AzureSearchLinkedService, - AzureSqlDatabaseLinkedService, - AzureSqlDWLinkedService, - AzureSqlDWTableDataset, - AzureSqlMILinkedService, - AzureSqlMITableDataset, - AzureSqlSink, - AzureSqlSource, - AzureSqlTableDataset, - AzureStorageLinkedService, - AzureTableDataset, - AzureTableSink, - AzureTableSource, - AzureTableStorageLinkedService, - BaseResource, - BinaryDataset, - BinarySink, - BinarySource, - BlobEventsTrigger, - BlobSink, - BlobSource, - BlobTrigger, - CassandraLinkedService, - CassandraSource, - CassandraTableDataset, - ChainingTrigger, - CloudError, - CmdkeySetup, - CommonDataServiceForAppsEntityDataset, - CommonDataServiceForAppsLinkedService, - CommonDataServiceForAppsSink, - CommonDataServiceForAppsSource, - ComponentSetup, - ConcurLinkedService, - ConcurObjectDataset, - ConcurSource, - ControlActivity, - CopyActivity, - CopySink, - CopySource, - CosmosDbLinkedService, - CosmosDbMongoDbApiCollectionDataset, - CosmosDbMongoDbApiLinkedService, - CosmosDbMongoDbApiSink, - CosmosDbMongoDbApiSource, - CosmosDbSqlApiCollectionDataset, - CosmosDbSqlApiSink, - CosmosDbSqlApiSource, - CouchbaseLinkedService, - CouchbaseSource, - CouchbaseTableDataset, - CustomActivity, - CustomActivityReferenceObject, - CustomDataset, - CustomDataSourceLinkedService, - CustomSetupBase, - DatabricksNotebookActivity, - DatabricksSparkJarActivity, - DatabricksSparkPythonActivity, - DataFlow, - DataFlowFolder, - DataFlowReference, - DataFlowResource, - DataFlowSink, - DataFlowSource, - DataFlowStagingInfo, - DataLakeAnalyticsUSQLActivity, - Dataset, - DatasetBZip2Compression, - DatasetCompression, - DatasetDeflateCompression, - DatasetFolder, - DatasetGZipCompression, - DatasetLocation, - DatasetReference, - DatasetResource, - DatasetStorageFormat, - DatasetZipDeflateCompression, - Db2LinkedService, - Db2Source, - Db2TableDataset, - DeleteActivity, - DelimitedTextDataset, - DelimitedTextReadSettings, - DelimitedTextSink, - DelimitedTextSource, - DelimitedTextWriteSettings, - DependencyReference, - DistcpSettings, - DocumentDbCollectionDataset, - DocumentDbCollectionSink, - DocumentDbCollectionSource, - DrillLinkedService, - DrillSource, - DrillTableDataset, - DWCopyCommandDefaultValue, - DWCopyCommandSettings, - DynamicsAXLinkedService, - DynamicsAXResourceDataset, - DynamicsAXSource, - DynamicsCrmEntityDataset, - DynamicsCrmLinkedService, - DynamicsCrmSink, - DynamicsCrmSource, - DynamicsEntityDataset, - DynamicsLinkedService, - DynamicsSink, - DynamicsSource, - EloquaLinkedService, - EloquaObjectDataset, - EloquaSource, - EntityReference, - EnvironmentVariableSetup, - ExecuteDataFlowActivity, - ExecuteDataFlowActivityTypePropertiesCompute, - ExecutePipelineActivity, - ExecuteSSISPackageActivity, - ExecutionActivity, - Expression, - Factory, - FactoryGitHubConfiguration, - FactoryIdentity, - FactoryRepoConfiguration, - FactoryVSTSConfiguration, - FileServerLinkedService, - FileServerLocation, - FileServerReadSettings, - FileServerWriteSettings, - FileShareDataset, - FileSystemSink, - FileSystemSource, - FilterActivity, - ForEachActivity, - FormatReadSettings, - FormatWriteSettings, - FtpReadSettings, - FtpServerLinkedService, - FtpServerLocation, - GetMetadataActivity, - GoogleAdWordsLinkedService, - GoogleAdWordsObjectDataset, - GoogleAdWordsSource, - GoogleBigQueryLinkedService, - GoogleBigQueryObjectDataset, - GoogleBigQuerySource, - GoogleCloudStorageLinkedService, - GoogleCloudStorageLocation, - GoogleCloudStorageReadSettings, - GreenplumLinkedService, - GreenplumSource, - GreenplumTableDataset, - HBaseLinkedService, - HBaseObjectDataset, - HBaseSource, - HdfsLinkedService, - HdfsLocation, - HdfsReadSettings, - HdfsSource, - HDInsightHiveActivity, - HDInsightLinkedService, - HDInsightMapReduceActivity, - HDInsightOnDemandLinkedService, - HDInsightPigActivity, - HDInsightSparkActivity, - HDInsightStreamingActivity, - HiveLinkedService, - HiveObjectDataset, - HiveSource, - HttpDataset, - HttpLinkedService, - HttpReadSettings, - HttpServerLocation, - HttpSource, - HubspotLinkedService, - HubspotObjectDataset, - HubspotSource, - IfConditionActivity, - ImpalaLinkedService, - ImpalaObjectDataset, - ImpalaSource, - InformixLinkedService, - InformixSink, - InformixSource, - InformixTableDataset, - IntegrationRuntime, - IntegrationRuntimeComputeProperties, - IntegrationRuntimeCustomSetupScriptProperties, - IntegrationRuntimeDataFlowProperties, - IntegrationRuntimeDataProxyProperties, - IntegrationRuntimeReference, - IntegrationRuntimeResource, - IntegrationRuntimeSsisCatalogInfo, - IntegrationRuntimeSsisProperties, - IntegrationRuntimeVNetProperties, - JiraLinkedService, - JiraObjectDataset, - JiraSource, - JsonDataset, - JsonFormat, - JsonSink, - JsonSource, - JsonWriteSettings, - LinkedIntegrationRuntimeKeyAuthorization, - LinkedIntegrationRuntimeRbacAuthorization, - LinkedIntegrationRuntimeType, - LinkedService, - LinkedServiceReference, - LinkedServiceResource, - LogStorageSettings, - LookupActivity, - MagentoLinkedService, - MagentoObjectDataset, - MagentoSource, - ManagedIntegrationRuntime, - MappingDataFlow, - MariaDBLinkedService, - MariaDBSource, - MariaDBTableDataset, - MarketoLinkedService, - MarketoObjectDataset, - MarketoSource, - MicrosoftAccessLinkedService, - MicrosoftAccessSink, - MicrosoftAccessSource, - MicrosoftAccessTableDataset, - MongoDbCollectionDataset, - MongoDbCursorMethodsProperties, - MongoDbLinkedService, - MongoDbSource, - MongoDbV2CollectionDataset, - MongoDbV2LinkedService, - MongoDbV2Source, - MultiplePipelineTrigger, - MySqlLinkedService, - MySqlSource, - MySqlTableDataset, - NetezzaLinkedService, - NetezzaPartitionSettings, - NetezzaSource, - NetezzaTableDataset, - ODataLinkedService, - ODataResourceDataset, - ODataSource, - OdbcLinkedService, - OdbcSink, - OdbcSource, - OdbcTableDataset, - Office365Dataset, - Office365LinkedService, - Office365Source, - OracleLinkedService, - OraclePartitionSettings, - OracleServiceCloudLinkedService, - OracleServiceCloudObjectDataset, - OracleServiceCloudSource, - OracleSink, - OracleSource, - OracleTableDataset, - OrcDataset, - OrcFormat, - OrcSink, - OrcSource, - ParameterSpecification, - ParquetDataset, - ParquetFormat, - ParquetSink, - ParquetSource, - PaypalLinkedService, - PaypalObjectDataset, - PaypalSource, - PhoenixLinkedService, - PhoenixObjectDataset, - PhoenixSource, - PipelineFolder, - PipelineReference, - PipelineResource, - PolybaseSettings, - PostgreSqlLinkedService, - PostgreSqlSource, - PostgreSqlTableDataset, - PrestoLinkedService, - PrestoObjectDataset, - PrestoSource, - QuickBooksLinkedService, - QuickBooksObjectDataset, - QuickBooksSource, - RecurrenceSchedule, - RecurrenceScheduleOccurrence, - RedirectIncompatibleRowSettings, - RedshiftUnloadSettings, - RelationalSource, - RelationalTableDataset, - RerunTriggerListResponse, - RerunTriggerResource, - RerunTumblingWindowTrigger, - RerunTumblingWindowTriggerActionParameters, - Resource, - ResponsysLinkedService, - ResponsysObjectDataset, - ResponsysSource, - RestResourceDataset, - RestServiceLinkedService, - RestSource, - RetryPolicy, - SalesforceLinkedService, - SalesforceMarketingCloudLinkedService, - SalesforceMarketingCloudObjectDataset, - SalesforceMarketingCloudSource, - SalesforceObjectDataset, - SalesforceServiceCloudLinkedService, - SalesforceServiceCloudObjectDataset, - SalesforceServiceCloudSink, - SalesforceServiceCloudSource, - SalesforceSink, - SalesforceSource, - SapBwCubeDataset, - SapBWLinkedService, - SapBwSource, - SapCloudForCustomerLinkedService, - SapCloudForCustomerResourceDataset, - SapCloudForCustomerSink, - SapCloudForCustomerSource, - SapEccLinkedService, - SapEccResourceDataset, - SapEccSource, - SapHanaLinkedService, - SapHanaPartitionSettings, - SapHanaSource, - SapHanaTableDataset, - SapOpenHubLinkedService, - SapOpenHubSource, - SapOpenHubTableDataset, - SapTableLinkedService, - SapTablePartitionSettings, - SapTableResourceDataset, - SapTableSource, - ScheduleTrigger, - ScheduleTriggerRecurrence, - ScriptAction, - SecretBase, - SecureString, - SelfDependencyTumblingWindowTriggerReference, - SelfHostedIntegrationRuntime, - ServiceNowLinkedService, - ServiceNowObjectDataset, - ServiceNowSource, - SetVariableActivity, - SftpLocation, - SftpReadSettings, - SftpServerLinkedService, - SftpWriteSettings, - ShopifyLinkedService, - ShopifyObjectDataset, - ShopifySource, - SparkLinkedService, - SparkObjectDataset, - SparkSource, - SqlDWSink, - SqlDWSource, - SqlMISink, - SqlMISource, - SqlServerLinkedService, - SqlServerSink, - SqlServerSource, - SqlServerStoredProcedureActivity, - SqlServerTableDataset, - SqlSink, - SqlSource, - SquareLinkedService, - SquareObjectDataset, - SquareSource, - SSISAccessCredential, - SSISChildPackage, - SSISExecutionCredential, - SSISExecutionParameter, - SSISLogLocation, - SSISPackageLocation, - SSISPropertyOverride, - StagingSettings, - StoredProcedureParameter, - StoreReadSettings, - StoreWriteSettings, - SubResource, - SwitchActivity, - SwitchCase, - SybaseLinkedService, - SybaseSource, - SybaseTableDataset, - TabularSource, - TeradataLinkedService, - TeradataPartitionSettings, - TeradataSource, - TeradataTableDataset, - TextFormat, - Transformation, - Trigger, - TriggerDependencyReference, - TriggerPipelineReference, - TriggerReference, - TriggerResource, - TumblingWindowTrigger, - TumblingWindowTriggerDependencyReference, - UntilActivity, - UserProperty, - ValidationActivity, - VariableSpecification, - VerticaLinkedService, - VerticaSource, - VerticaTableDataset, - WaitActivity, - WebActivity, - WebActivityAuthentication, - WebAnonymousAuthentication, - WebBasicAuthentication, - WebClientCertificateAuthentication, - WebHookActivity, - WebLinkedService, - WebLinkedServiceTypeProperties, - WebSource, - WebTableDataset, - XeroLinkedService, - XeroObjectDataset, - XeroSource, - ZohoLinkedService, - ZohoObjectDataset, - ZohoSource -} from "../models/mappers"; diff --git a/sdk/datafactory/arm-datafactory/src/models/triggersMappers.ts b/sdk/datafactory/arm-datafactory/src/models/triggersMappers.ts index ce8f2bc7097d..163a42a3998b 100644 --- a/sdk/datafactory/arm-datafactory/src/models/triggersMappers.ts +++ b/sdk/datafactory/arm-datafactory/src/models/triggersMappers.ts @@ -11,6 +11,7 @@ export { Activity, ActivityDependency, ActivityPolicy, + AdditionalColumns, AmazonMWSLinkedService, AmazonMWSObjectDataset, AmazonMWSSource, @@ -199,6 +200,7 @@ export { ExecutePipelineActivity, ExecuteSSISPackageActivity, ExecutionActivity, + ExportSettings, Expression, Factory, FactoryGitHubConfiguration, @@ -261,6 +263,7 @@ export { ImpalaLinkedService, ImpalaObjectDataset, ImpalaSource, + ImportSettings, InformixLinkedService, InformixSink, InformixSource, @@ -373,7 +376,6 @@ export { RedshiftUnloadSettings, RelationalSource, RelationalTableDataset, - RerunTriggerResource, RerunTumblingWindowTrigger, Resource, ResponsysLinkedService, @@ -433,6 +435,13 @@ export { ShopifyLinkedService, ShopifyObjectDataset, ShopifySource, + SkipErrorFile, + SnowflakeDataset, + SnowflakeExportCopyCommand, + SnowflakeImportCopyCommand, + SnowflakeLinkedService, + SnowflakeSink, + SnowflakeSource, SparkLinkedService, SparkObjectDataset, SparkSource, @@ -476,8 +485,10 @@ export { Transformation, Trigger, TriggerDependencyReference, + TriggerFilterParameters, TriggerListResponse, TriggerPipelineReference, + TriggerQueryResponse, TriggerReference, TriggerResource, TriggerSubscriptionOperationStatus, diff --git a/sdk/datafactory/arm-datafactory/src/operations/index.ts b/sdk/datafactory/arm-datafactory/src/operations/index.ts index feda4fb51a5e..a4e051342b30 100644 --- a/sdk/datafactory/arm-datafactory/src/operations/index.ts +++ b/sdk/datafactory/arm-datafactory/src/operations/index.ts @@ -21,6 +21,5 @@ export * from "./pipelineRuns"; export * from "./activityRuns"; export * from "./triggers"; export * from "./triggerRuns"; -export * from "./rerunTriggers"; export * from "./dataFlows"; export * from "./dataFlowDebugSession"; diff --git a/sdk/datafactory/arm-datafactory/src/operations/pipelines.ts b/sdk/datafactory/arm-datafactory/src/operations/pipelines.ts index e0d7c1b37157..0b0d35920ac5 100644 --- a/sdk/datafactory/arm-datafactory/src/operations/pipelines.ts +++ b/sdk/datafactory/arm-datafactory/src/operations/pipelines.ts @@ -362,7 +362,8 @@ const createRunOperationSpec: msRest.OperationSpec = { Parameters.apiVersion, Parameters.referencePipelineRunId, Parameters.isRecovery, - Parameters.startActivityName + Parameters.startActivityName, + Parameters.startFromFailure ], headerParameters: [ Parameters.acceptLanguage diff --git a/sdk/datafactory/arm-datafactory/src/operations/rerunTriggers.ts b/sdk/datafactory/arm-datafactory/src/operations/rerunTriggers.ts deleted file mode 100644 index 22693ccbabba..000000000000 --- a/sdk/datafactory/arm-datafactory/src/operations/rerunTriggers.ts +++ /dev/null @@ -1,406 +0,0 @@ -/* - * Copyright (c) Microsoft Corporation. All rights reserved. - * Licensed under the MIT License. See License.txt in the project root for - * license information. - * - * Code generated by Microsoft (R) AutoRest Code Generator. - * Changes may cause incorrect behavior and will be lost if the code is - * regenerated. - */ - -import * as msRest from "@azure/ms-rest-js"; -import * as msRestAzure from "@azure/ms-rest-azure-js"; -import * as Models from "../models"; -import * as Mappers from "../models/rerunTriggersMappers"; -import * as Parameters from "../models/parameters"; -import { DataFactoryManagementClientContext } from "../dataFactoryManagementClientContext"; - -/** Class representing a RerunTriggers. */ -export class RerunTriggers { - private readonly client: DataFactoryManagementClientContext; - - /** - * Create a RerunTriggers. - * @param {DataFactoryManagementClientContext} client Reference to the service client. - */ - constructor(client: DataFactoryManagementClientContext) { - this.client = client; - } - - /** - * Creates a rerun trigger. - * @param resourceGroupName The resource group name. - * @param factoryName The factory name. - * @param triggerName The trigger name. - * @param rerunTriggerName The rerun trigger name. - * @param rerunTumblingWindowTriggerActionParameters Rerun tumbling window trigger action - * parameters. - * @param [options] The optional parameters - * @returns Promise - */ - create(resourceGroupName: string, factoryName: string, triggerName: string, rerunTriggerName: string, rerunTumblingWindowTriggerActionParameters: Models.RerunTumblingWindowTriggerActionParameters, options?: msRest.RequestOptionsBase): Promise; - /** - * @param resourceGroupName The resource group name. - * @param factoryName The factory name. - * @param triggerName The trigger name. - * @param rerunTriggerName The rerun trigger name. - * @param rerunTumblingWindowTriggerActionParameters Rerun tumbling window trigger action - * parameters. - * @param callback The callback - */ - create(resourceGroupName: string, factoryName: string, triggerName: string, rerunTriggerName: string, rerunTumblingWindowTriggerActionParameters: Models.RerunTumblingWindowTriggerActionParameters, callback: msRest.ServiceCallback): void; - /** - * @param resourceGroupName The resource group name. - * @param factoryName The factory name. - * @param triggerName The trigger name. - * @param rerunTriggerName The rerun trigger name. - * @param rerunTumblingWindowTriggerActionParameters Rerun tumbling window trigger action - * parameters. - * @param options The optional parameters - * @param callback The callback - */ - create(resourceGroupName: string, factoryName: string, triggerName: string, rerunTriggerName: string, rerunTumblingWindowTriggerActionParameters: Models.RerunTumblingWindowTriggerActionParameters, options: msRest.RequestOptionsBase, callback: msRest.ServiceCallback): void; - create(resourceGroupName: string, factoryName: string, triggerName: string, rerunTriggerName: string, rerunTumblingWindowTriggerActionParameters: Models.RerunTumblingWindowTriggerActionParameters, options?: msRest.RequestOptionsBase | msRest.ServiceCallback, callback?: msRest.ServiceCallback): Promise { - return this.client.sendOperationRequest( - { - resourceGroupName, - factoryName, - triggerName, - rerunTriggerName, - rerunTumblingWindowTriggerActionParameters, - options - }, - createOperationSpec, - callback) as Promise; - } - - /** - * Starts a trigger. - * @param resourceGroupName The resource group name. - * @param factoryName The factory name. - * @param triggerName The trigger name. - * @param rerunTriggerName The rerun trigger name. - * @param [options] The optional parameters - * @returns Promise - */ - start(resourceGroupName: string, factoryName: string, triggerName: string, rerunTriggerName: string, options?: msRest.RequestOptionsBase): Promise { - return this.beginStart(resourceGroupName,factoryName,triggerName,rerunTriggerName,options) - .then(lroPoller => lroPoller.pollUntilFinished()); - } - - /** - * Stops a trigger. - * @param resourceGroupName The resource group name. - * @param factoryName The factory name. - * @param triggerName The trigger name. - * @param rerunTriggerName The rerun trigger name. - * @param [options] The optional parameters - * @returns Promise - */ - stop(resourceGroupName: string, factoryName: string, triggerName: string, rerunTriggerName: string, options?: msRest.RequestOptionsBase): Promise { - return this.beginStop(resourceGroupName,factoryName,triggerName,rerunTriggerName,options) - .then(lroPoller => lroPoller.pollUntilFinished()); - } - - /** - * Cancels a trigger. - * @param resourceGroupName The resource group name. - * @param factoryName The factory name. - * @param triggerName The trigger name. - * @param rerunTriggerName The rerun trigger name. - * @param [options] The optional parameters - * @returns Promise - */ - cancel(resourceGroupName: string, factoryName: string, triggerName: string, rerunTriggerName: string, options?: msRest.RequestOptionsBase): Promise { - return this.beginCancel(resourceGroupName,factoryName,triggerName,rerunTriggerName,options) - .then(lroPoller => lroPoller.pollUntilFinished()); - } - - /** - * Lists rerun triggers by an original trigger name. - * @param resourceGroupName The resource group name. - * @param factoryName The factory name. - * @param triggerName The trigger name. - * @param [options] The optional parameters - * @returns Promise - */ - listByTrigger(resourceGroupName: string, factoryName: string, triggerName: string, options?: msRest.RequestOptionsBase): Promise; - /** - * @param resourceGroupName The resource group name. - * @param factoryName The factory name. - * @param triggerName The trigger name. - * @param callback The callback - */ - listByTrigger(resourceGroupName: string, factoryName: string, triggerName: string, callback: msRest.ServiceCallback): void; - /** - * @param resourceGroupName The resource group name. - * @param factoryName The factory name. - * @param triggerName The trigger name. - * @param options The optional parameters - * @param callback The callback - */ - listByTrigger(resourceGroupName: string, factoryName: string, triggerName: string, options: msRest.RequestOptionsBase, callback: msRest.ServiceCallback): void; - listByTrigger(resourceGroupName: string, factoryName: string, triggerName: string, options?: msRest.RequestOptionsBase | msRest.ServiceCallback, callback?: msRest.ServiceCallback): Promise { - return this.client.sendOperationRequest( - { - resourceGroupName, - factoryName, - triggerName, - options - }, - listByTriggerOperationSpec, - callback) as Promise; - } - - /** - * Starts a trigger. - * @param resourceGroupName The resource group name. - * @param factoryName The factory name. - * @param triggerName The trigger name. - * @param rerunTriggerName The rerun trigger name. - * @param [options] The optional parameters - * @returns Promise - */ - beginStart(resourceGroupName: string, factoryName: string, triggerName: string, rerunTriggerName: string, options?: msRest.RequestOptionsBase): Promise { - return this.client.sendLRORequest( - { - resourceGroupName, - factoryName, - triggerName, - rerunTriggerName, - options - }, - beginStartOperationSpec, - options); - } - - /** - * Stops a trigger. - * @param resourceGroupName The resource group name. - * @param factoryName The factory name. - * @param triggerName The trigger name. - * @param rerunTriggerName The rerun trigger name. - * @param [options] The optional parameters - * @returns Promise - */ - beginStop(resourceGroupName: string, factoryName: string, triggerName: string, rerunTriggerName: string, options?: msRest.RequestOptionsBase): Promise { - return this.client.sendLRORequest( - { - resourceGroupName, - factoryName, - triggerName, - rerunTriggerName, - options - }, - beginStopOperationSpec, - options); - } - - /** - * Cancels a trigger. - * @param resourceGroupName The resource group name. - * @param factoryName The factory name. - * @param triggerName The trigger name. - * @param rerunTriggerName The rerun trigger name. - * @param [options] The optional parameters - * @returns Promise - */ - beginCancel(resourceGroupName: string, factoryName: string, triggerName: string, rerunTriggerName: string, options?: msRest.RequestOptionsBase): Promise { - return this.client.sendLRORequest( - { - resourceGroupName, - factoryName, - triggerName, - rerunTriggerName, - options - }, - beginCancelOperationSpec, - options); - } - - /** - * Lists rerun triggers by an original trigger name. - * @param nextPageLink The NextLink from the previous successful call to List operation. - * @param [options] The optional parameters - * @returns Promise - */ - listByTriggerNext(nextPageLink: string, options?: msRest.RequestOptionsBase): Promise; - /** - * @param nextPageLink The NextLink from the previous successful call to List operation. - * @param callback The callback - */ - listByTriggerNext(nextPageLink: string, callback: msRest.ServiceCallback): void; - /** - * @param nextPageLink The NextLink from the previous successful call to List operation. - * @param options The optional parameters - * @param callback The callback - */ - listByTriggerNext(nextPageLink: string, options: msRest.RequestOptionsBase, callback: msRest.ServiceCallback): void; - listByTriggerNext(nextPageLink: string, options?: msRest.RequestOptionsBase | msRest.ServiceCallback, callback?: msRest.ServiceCallback): Promise { - return this.client.sendOperationRequest( - { - nextPageLink, - options - }, - listByTriggerNextOperationSpec, - callback) as Promise; - } -} - -// Operation Specifications -const serializer = new msRest.Serializer(Mappers); -const createOperationSpec: msRest.OperationSpec = { - httpMethod: "PUT", - path: "subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.DataFactory/factories/{factoryName}/triggers/{triggerName}/rerunTriggers/{rerunTriggerName}", - urlParameters: [ - Parameters.subscriptionId, - Parameters.resourceGroupName, - Parameters.factoryName, - Parameters.triggerName, - Parameters.rerunTriggerName - ], - queryParameters: [ - Parameters.apiVersion - ], - headerParameters: [ - Parameters.acceptLanguage - ], - requestBody: { - parameterPath: "rerunTumblingWindowTriggerActionParameters", - mapper: { - ...Mappers.RerunTumblingWindowTriggerActionParameters, - required: true - } - }, - responses: { - 200: { - bodyMapper: Mappers.TriggerResource - }, - default: { - bodyMapper: Mappers.CloudError - } - }, - serializer -}; - -const listByTriggerOperationSpec: msRest.OperationSpec = { - httpMethod: "GET", - path: "subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.DataFactory/factories/{factoryName}/triggers/{triggerName}/rerunTriggers", - urlParameters: [ - Parameters.subscriptionId, - Parameters.resourceGroupName, - Parameters.factoryName, - Parameters.triggerName - ], - queryParameters: [ - Parameters.apiVersion - ], - headerParameters: [ - Parameters.acceptLanguage - ], - responses: { - 200: { - bodyMapper: Mappers.RerunTriggerListResponse - }, - default: { - bodyMapper: Mappers.CloudError - } - }, - serializer -}; - -const beginStartOperationSpec: msRest.OperationSpec = { - httpMethod: "POST", - path: "subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.DataFactory/factories/{factoryName}/triggers/{triggerName}/rerunTriggers/{rerunTriggerName}/start", - urlParameters: [ - Parameters.subscriptionId, - Parameters.resourceGroupName, - Parameters.factoryName, - Parameters.triggerName, - Parameters.rerunTriggerName - ], - queryParameters: [ - Parameters.apiVersion - ], - headerParameters: [ - Parameters.acceptLanguage - ], - responses: { - 200: {}, - default: { - bodyMapper: Mappers.CloudError - } - }, - serializer -}; - -const beginStopOperationSpec: msRest.OperationSpec = { - httpMethod: "POST", - path: "subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.DataFactory/factories/{factoryName}/triggers/{triggerName}/rerunTriggers/{rerunTriggerName}/stop", - urlParameters: [ - Parameters.subscriptionId, - Parameters.resourceGroupName, - Parameters.factoryName, - Parameters.triggerName, - Parameters.rerunTriggerName - ], - queryParameters: [ - Parameters.apiVersion - ], - headerParameters: [ - Parameters.acceptLanguage - ], - responses: { - 200: {}, - default: { - bodyMapper: Mappers.CloudError - } - }, - serializer -}; - -const beginCancelOperationSpec: msRest.OperationSpec = { - httpMethod: "POST", - path: "subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.DataFactory/factories/{factoryName}/triggers/{triggerName}/rerunTriggers/{rerunTriggerName}/cancel", - urlParameters: [ - Parameters.subscriptionId, - Parameters.resourceGroupName, - Parameters.factoryName, - Parameters.triggerName, - Parameters.rerunTriggerName - ], - queryParameters: [ - Parameters.apiVersion - ], - headerParameters: [ - Parameters.acceptLanguage - ], - responses: { - 200: {}, - default: { - bodyMapper: Mappers.CloudError - } - }, - serializer -}; - -const listByTriggerNextOperationSpec: msRest.OperationSpec = { - httpMethod: "GET", - baseUrl: "https://management.azure.com", - path: "{nextLink}", - urlParameters: [ - Parameters.nextPageLink - ], - headerParameters: [ - Parameters.acceptLanguage - ], - responses: { - 200: { - bodyMapper: Mappers.RerunTriggerListResponse - }, - default: { - bodyMapper: Mappers.CloudError - } - }, - serializer -}; diff --git a/sdk/datafactory/arm-datafactory/src/operations/triggers.ts b/sdk/datafactory/arm-datafactory/src/operations/triggers.ts index 49a41a8960d2..5218d661220e 100644 --- a/sdk/datafactory/arm-datafactory/src/operations/triggers.ts +++ b/sdk/datafactory/arm-datafactory/src/operations/triggers.ts @@ -59,6 +59,42 @@ export class Triggers { callback) as Promise; } + /** + * Query triggers. + * @param resourceGroupName The resource group name. + * @param factoryName The factory name. + * @param filterParameters Parameters to filter the triggers. + * @param [options] The optional parameters + * @returns Promise + */ + queryByFactory(resourceGroupName: string, factoryName: string, filterParameters: Models.TriggerFilterParameters, options?: msRest.RequestOptionsBase): Promise; + /** + * @param resourceGroupName The resource group name. + * @param factoryName The factory name. + * @param filterParameters Parameters to filter the triggers. + * @param callback The callback + */ + queryByFactory(resourceGroupName: string, factoryName: string, filterParameters: Models.TriggerFilterParameters, callback: msRest.ServiceCallback): void; + /** + * @param resourceGroupName The resource group name. + * @param factoryName The factory name. + * @param filterParameters Parameters to filter the triggers. + * @param options The optional parameters + * @param callback The callback + */ + queryByFactory(resourceGroupName: string, factoryName: string, filterParameters: Models.TriggerFilterParameters, options: msRest.RequestOptionsBase, callback: msRest.ServiceCallback): void; + queryByFactory(resourceGroupName: string, factoryName: string, filterParameters: Models.TriggerFilterParameters, options?: msRest.RequestOptionsBase | msRest.ServiceCallback, callback?: msRest.ServiceCallback): Promise { + return this.client.sendOperationRequest( + { + resourceGroupName, + factoryName, + filterParameters, + options + }, + queryByFactoryOperationSpec, + callback) as Promise; + } + /** * Creates or updates a trigger. * @param resourceGroupName The resource group name. @@ -395,6 +431,38 @@ const listByFactoryOperationSpec: msRest.OperationSpec = { serializer }; +const queryByFactoryOperationSpec: msRest.OperationSpec = { + httpMethod: "POST", + path: "subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.DataFactory/factories/{factoryName}/querytriggers", + urlParameters: [ + Parameters.subscriptionId, + Parameters.resourceGroupName, + Parameters.factoryName + ], + queryParameters: [ + Parameters.apiVersion + ], + headerParameters: [ + Parameters.acceptLanguage + ], + requestBody: { + parameterPath: "filterParameters", + mapper: { + ...Mappers.TriggerFilterParameters, + required: true + } + }, + responses: { + 200: { + bodyMapper: Mappers.TriggerQueryResponse + }, + default: { + bodyMapper: Mappers.CloudError + } + }, + serializer +}; + const createOrUpdateOperationSpec: msRest.OperationSpec = { httpMethod: "PUT", path: "subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.DataFactory/factories/{factoryName}/triggers/{triggerName}",