From 38b59412a309e5b8d2cac3fbee68276d775ca02c Mon Sep 17 00:00:00 2001 From: Laurent Mazuel Date: Mon, 20 Nov 2017 20:42:42 +0000 Subject: [PATCH] Generated from 75ff961199b2af69541c91c2c9dba0c3b85ea1e7 --- .../azure/mgmt/datafactory/models/__init__.py | 228 +++++++++++++++++- .../models/amazon_mws_linked_service.py | 92 +++++++ .../models/amazon_mws_object_dataset.py | 40 +++ .../datafactory/models/amazon_mws_source.py | 46 ++++ .../models/azure_data_lake_store_dataset.py | 8 +- .../models/azure_my_sql_linked_service.py | 50 ++++ .../datafactory/models/azure_my_sql_source.py | 46 ++++ .../models/azure_my_sql_table_dataset.py | 53 ++++ .../azure_postgre_sql_linked_service.py | 49 ++++ .../models/azure_postgre_sql_source.py | 46 ++++ .../models/azure_postgre_sql_table_dataset.py | 40 +++ .../mgmt/datafactory/models/blob_trigger.py | 12 +- .../models/concur_linked_service.py | 76 ++++++ .../models/concur_object_dataset.py | 40 +++ .../mgmt/datafactory/models/concur_source.py | 46 ++++ .../mgmt/datafactory/models/copy_sink.py | 10 +- .../mgmt/datafactory/models/copy_source.py | 16 +- .../models/couchbase_linked_service.py | 49 ++++ .../datafactory/models/couchbase_source.py | 46 ++++ .../models/couchbase_table_dataset.py | 40 +++ .../data_factory_management_client_enums.py | 100 ++++++++ .../azure/mgmt/datafactory/models/dataset.py | 16 +- .../datafactory/models/dataset_compression.py | 6 +- .../dataset_date_time_partition_value.py | 42 ---- .../datafactory/models/dataset_partition.py | 32 --- ....py => dataset_zip_deflate_compression.py} | 23 +- .../models/drill_linked_service.py | 49 ++++ .../mgmt/datafactory/models/drill_source.py | 46 ++++ .../datafactory/models/drill_table_dataset.py | 40 +++ .../models/dynamics_linked_service.py | 14 +- .../models/eloqua_linked_service.py | 77 ++++++ .../models/eloqua_object_dataset.py | 40 +++ .../mgmt/datafactory/models/eloqua_source.py | 46 ++++ .../models/google_big_query_linked_service.py | 99 ++++++++ .../models/google_big_query_object_dataset.py | 40 +++ .../models/google_big_query_source.py | 46 ++++ .../models/greenplum_linked_service.py | 49 ++++ .../datafactory/models/greenplum_source.py | 46 ++++ .../models/greenplum_table_dataset.py | 40 +++ .../models/hbase_linked_service.py | 99 ++++++++ .../models/hbase_object_dataset.py | 40 +++ .../mgmt/datafactory/models/hbase_source.py | 46 ++++ .../mgmt/datafactory/models/hdfs_source.py | 2 +- .../datafactory/models/hive_linked_service.py | 132 ++++++++++ .../datafactory/models/hive_object_dataset.py | 40 +++ .../mgmt/datafactory/models/hive_source.py | 46 ++++ .../mgmt/datafactory/models/http_dataset.py | 6 +- .../models/hubspot_linked_service.py | 81 +++++++ .../models/hubspot_object_dataset.py | 40 +++ .../mgmt/datafactory/models/hubspot_source.py | 46 ++++ .../models/impala_linked_service.py | 102 ++++++++ .../models/impala_object_dataset.py | 40 +++ .../mgmt/datafactory/models/impala_source.py | 46 ++++ .../datafactory/models/jira_linked_service.py | 83 +++++++ .../datafactory/models/jira_object_dataset.py | 40 +++ .../mgmt/datafactory/models/jira_source.py | 46 ++++ .../mgmt/datafactory/models/json_format.py | 22 +- .../mgmt/datafactory/models/linked_service.py | 28 ++- .../models/magento_linked_service.py | 71 ++++++ .../models/magento_object_dataset.py | 40 +++ .../mgmt/datafactory/models/magento_source.py | 46 ++++ .../models/maria_db_linked_service.py | 49 ++++ .../datafactory/models/maria_db_source.py | 46 ++++ .../models/maria_db_table_dataset.py | 40 +++ .../models/marketo_linked_service.py | 76 ++++++ .../models/marketo_object_dataset.py | 40 +++ .../mgmt/datafactory/models/marketo_source.py | 46 ++++ .../models/multiple_pipeline_trigger.py | 58 +++++ .../models/paypal_linked_service.py | 76 ++++++ .../models/paypal_object_dataset.py | 40 +++ .../mgmt/datafactory/models/paypal_source.py | 46 ++++ .../models/phoenix_linked_service.py | 107 ++++++++ .../models/phoenix_object_dataset.py | 40 +++ .../mgmt/datafactory/models/phoenix_source.py | 46 ++++ .../models/presto_linked_service.py | 116 +++++++++ .../models/presto_object_dataset.py | 40 +++ .../mgmt/datafactory/models/presto_source.py | 46 ++++ .../models/quick_books_linked_service.py | 70 ++++++ .../models/quick_books_object_dataset.py | 40 +++ .../datafactory/models/quick_books_source.py | 46 ++++ .../redirect_incompatible_row_settings.py | 12 +- .../mgmt/datafactory/models/retry_policy.py | 37 +++ .../models/salesforce_linked_service.py | 6 +- .../models/salesforce_object_dataset.py | 53 ++++ .../datafactory/models/salesforce_sink.py | 73 ++++++ .../datafactory/models/salesforce_source.py | 46 ++++ .../sap_cloud_for_customer_linked_service.py | 62 +++++ ...sap_cloud_for_customer_resource_dataset.py | 54 +++++ .../models/sap_cloud_for_customer_sink.py | 56 +++++ .../models/sap_cloud_for_customer_source.py | 46 ++++ .../datafactory/models/schedule_trigger.py | 12 +- .../models/service_now_linked_service.py | 92 +++++++ .../models/service_now_object_dataset.py | 40 +++ .../datafactory/models/service_now_source.py | 46 ++++ .../models/shopify_linked_service.py | 72 ++++++ .../models/shopify_object_dataset.py | 40 +++ .../mgmt/datafactory/models/shopify_source.py | 46 ++++ .../models/spark_linked_service.py | 117 +++++++++ .../models/spark_object_dataset.py | 40 +++ .../mgmt/datafactory/models/spark_source.py | 46 ++++ .../models/square_linked_service.py | 82 +++++++ .../models/square_object_dataset.py | 40 +++ .../mgmt/datafactory/models/square_source.py | 46 ++++ .../azure/mgmt/datafactory/models/trigger.py | 11 +- .../models/tumbling_window_trigger.py | 94 ++++++++ .../datafactory/models/xero_linked_service.py | 78 ++++++ .../datafactory/models/xero_object_dataset.py | 40 +++ .../mgmt/datafactory/models/xero_source.py | 46 ++++ .../datafactory/models/zoho_linked_service.py | 71 ++++++ .../datafactory/models/zoho_object_dataset.py | 40 +++ .../mgmt/datafactory/models/zoho_source.py | 46 ++++ .../azure/mgmt/datafactory/version.py | 2 +- 112 files changed, 5513 insertions(+), 173 deletions(-) create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_mws_linked_service.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_mws_object_dataset.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_mws_source.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_my_sql_linked_service.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_my_sql_source.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_my_sql_table_dataset.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_postgre_sql_linked_service.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_postgre_sql_source.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_postgre_sql_table_dataset.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/concur_linked_service.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/concur_object_dataset.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/concur_source.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/couchbase_linked_service.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/couchbase_source.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/couchbase_table_dataset.py delete mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_date_time_partition_value.py delete mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_partition.py rename azure-mgmt-datafactory/azure/mgmt/datafactory/models/{dataset_partition_value.py => dataset_zip_deflate_compression.py} (55%) create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/drill_linked_service.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/drill_source.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/drill_table_dataset.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/eloqua_linked_service.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/eloqua_object_dataset.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/eloqua_source.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/google_big_query_linked_service.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/google_big_query_object_dataset.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/google_big_query_source.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/greenplum_linked_service.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/greenplum_source.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/greenplum_table_dataset.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/hbase_linked_service.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/hbase_object_dataset.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/hbase_source.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/hive_linked_service.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/hive_object_dataset.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/hive_source.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/hubspot_linked_service.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/hubspot_object_dataset.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/hubspot_source.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/impala_linked_service.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/impala_object_dataset.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/impala_source.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/jira_linked_service.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/jira_object_dataset.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/jira_source.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/magento_linked_service.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/magento_object_dataset.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/magento_source.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/maria_db_linked_service.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/maria_db_source.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/maria_db_table_dataset.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/marketo_linked_service.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/marketo_object_dataset.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/marketo_source.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/multiple_pipeline_trigger.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/paypal_linked_service.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/paypal_object_dataset.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/paypal_source.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/phoenix_linked_service.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/phoenix_object_dataset.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/phoenix_source.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/presto_linked_service.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/presto_object_dataset.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/presto_source.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/quick_books_linked_service.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/quick_books_object_dataset.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/quick_books_source.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/retry_policy.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_object_dataset.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_sink.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_source.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_cloud_for_customer_linked_service.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_cloud_for_customer_resource_dataset.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_cloud_for_customer_sink.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_cloud_for_customer_source.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/service_now_linked_service.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/service_now_object_dataset.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/service_now_source.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/shopify_linked_service.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/shopify_object_dataset.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/shopify_source.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/spark_linked_service.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/spark_object_dataset.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/spark_source.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/square_linked_service.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/square_object_dataset.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/square_source.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/tumbling_window_trigger.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/xero_linked_service.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/xero_object_dataset.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/xero_source.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/zoho_linked_service.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/zoho_object_dataset.py create mode 100644 azure-mgmt-datafactory/azure/mgmt/datafactory/models/zoho_source.py diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py index 6d8711b055f7..e7231a706224 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/__init__.py @@ -32,12 +32,12 @@ from .activity_dependency import ActivityDependency from .activity import Activity from .pipeline_resource import PipelineResource -from .pipeline_reference import PipelineReference -from .trigger_pipeline_reference import TriggerPipelineReference from .trigger import Trigger from .trigger_resource import TriggerResource from .create_run_response import CreateRunResponse from .error_response import ErrorResponse, ErrorResponseException +from .pipeline_reference import PipelineReference +from .trigger_pipeline_reference import TriggerPipelineReference from .factory_update_parameters import FactoryUpdateParameters from .dataset_reference import DatasetReference from .pipeline_run_query_filter import PipelineRunQueryFilter @@ -57,6 +57,32 @@ from .operation_list_response import OperationListResponse from .azure_data_lake_analytics_linked_service import AzureDataLakeAnalyticsLinkedService from .hd_insight_on_demand_linked_service import HDInsightOnDemandLinkedService +from .zoho_linked_service import ZohoLinkedService +from .xero_linked_service import XeroLinkedService +from .square_linked_service import SquareLinkedService +from .spark_linked_service import SparkLinkedService +from .shopify_linked_service import ShopifyLinkedService +from .service_now_linked_service import ServiceNowLinkedService +from .quick_books_linked_service import QuickBooksLinkedService +from .presto_linked_service import PrestoLinkedService +from .phoenix_linked_service import PhoenixLinkedService +from .paypal_linked_service import PaypalLinkedService +from .marketo_linked_service import MarketoLinkedService +from .maria_db_linked_service import MariaDBLinkedService +from .magento_linked_service import MagentoLinkedService +from .jira_linked_service import JiraLinkedService +from .impala_linked_service import ImpalaLinkedService +from .hubspot_linked_service import HubspotLinkedService +from .hive_linked_service import HiveLinkedService +from .hbase_linked_service import HBaseLinkedService +from .greenplum_linked_service import GreenplumLinkedService +from .google_big_query_linked_service import GoogleBigQueryLinkedService +from .eloqua_linked_service import EloquaLinkedService +from .drill_linked_service import DrillLinkedService +from .couchbase_linked_service import CouchbaseLinkedService +from .concur_linked_service import ConcurLinkedService +from .azure_postgre_sql_linked_service import AzurePostgreSqlLinkedService +from .amazon_mws_linked_service import AmazonMWSLinkedService from .sap_hana_linked_service import SapHanaLinkedService from .sap_bw_linked_service import SapBWLinkedService from .sftp_server_linked_service import SftpServerLinkedService @@ -66,6 +92,7 @@ from .custom_data_source_linked_service import CustomDataSourceLinkedService from .amazon_redshift_linked_service import AmazonRedshiftLinkedService from .amazon_s3_linked_service import AmazonS3LinkedService +from .sap_cloud_for_customer_linked_service import SapCloudForCustomerLinkedService from .salesforce_linked_service import SalesforceLinkedService from .azure_data_lake_store_linked_service import AzureDataLakeStoreLinkedService from .mongo_db_linked_service import MongoDbLinkedService @@ -84,6 +111,7 @@ from .sybase_linked_service import SybaseLinkedService from .postgre_sql_linked_service import PostgreSqlLinkedService from .my_sql_linked_service import MySqlLinkedService +from .azure_my_sql_linked_service import AzureMySqlLinkedService from .oracle_linked_service import OracleLinkedService from .file_server_linked_service import FileServerLinkedService from .hd_insight_linked_service import HDInsightLinkedService @@ -95,6 +123,33 @@ from .sql_server_linked_service import SqlServerLinkedService from .azure_sql_dw_linked_service import AzureSqlDWLinkedService from .azure_storage_linked_service import AzureStorageLinkedService +from .zoho_object_dataset import ZohoObjectDataset +from .xero_object_dataset import XeroObjectDataset +from .square_object_dataset import SquareObjectDataset +from .spark_object_dataset import SparkObjectDataset +from .shopify_object_dataset import ShopifyObjectDataset +from .service_now_object_dataset import ServiceNowObjectDataset +from .quick_books_object_dataset import QuickBooksObjectDataset +from .presto_object_dataset import PrestoObjectDataset +from .phoenix_object_dataset import PhoenixObjectDataset +from .paypal_object_dataset import PaypalObjectDataset +from .marketo_object_dataset import MarketoObjectDataset +from .maria_db_table_dataset import MariaDBTableDataset +from .magento_object_dataset import MagentoObjectDataset +from .jira_object_dataset import JiraObjectDataset +from .impala_object_dataset import ImpalaObjectDataset +from .hubspot_object_dataset import HubspotObjectDataset +from .hive_object_dataset import HiveObjectDataset +from .hbase_object_dataset import HBaseObjectDataset +from .greenplum_table_dataset import GreenplumTableDataset +from .google_big_query_object_dataset import GoogleBigQueryObjectDataset +from .eloqua_object_dataset import EloquaObjectDataset +from .drill_table_dataset import DrillTableDataset +from .couchbase_table_dataset import CouchbaseTableDataset +from .concur_object_dataset import ConcurObjectDataset +from .azure_postgre_sql_table_dataset import AzurePostgreSqlTableDataset +from .amazon_mws_object_dataset import AmazonMWSObjectDataset +from .dataset_zip_deflate_compression import DatasetZipDeflateCompression from .dataset_deflate_compression import DatasetDeflateCompression from .dataset_gzip_compression import DatasetGZipCompression from .dataset_bzip2_compression import DatasetBZip2Compression @@ -105,14 +160,14 @@ from .json_format import JsonFormat from .text_format import TextFormat from .dataset_storage_format import DatasetStorageFormat -from .dataset_date_time_partition_value import DatasetDateTimePartitionValue -from .dataset_partition_value import DatasetPartitionValue -from .dataset_partition import DatasetPartition from .http_dataset import HttpDataset from .azure_search_index_dataset import AzureSearchIndexDataset from .web_table_dataset import WebTableDataset from .sql_server_table_dataset import SqlServerTableDataset +from .sap_cloud_for_customer_resource_dataset import SapCloudForCustomerResourceDataset +from .salesforce_object_dataset import SalesforceObjectDataset from .relational_table_dataset import RelationalTableDataset +from .azure_my_sql_table_dataset import AzureMySqlTableDataset from .oracle_table_dataset import OracleTableDataset from .odata_resource_dataset import ODataResourceDataset from .mongo_db_collection_dataset import MongoDbCollectionDataset @@ -127,11 +182,14 @@ from .azure_table_dataset import AzureTableDataset from .azure_blob_dataset import AzureBlobDataset from .amazon_s3_dataset import AmazonS3Dataset +from .retry_policy import RetryPolicy +from .tumbling_window_trigger import TumblingWindowTrigger from .blob_trigger import BlobTrigger from .recurrence_schedule_occurrence import RecurrenceScheduleOccurrence from .recurrence_schedule import RecurrenceSchedule from .schedule_trigger_recurrence import ScheduleTriggerRecurrence from .schedule_trigger import ScheduleTrigger +from .multiple_pipeline_trigger import MultiplePipelineTrigger from .activity_policy import ActivityPolicy from .data_lake_analytics_usql_activity import DataLakeAnalyticsUSQLActivity from .azure_ml_update_resource_activity import AzureMLUpdateResourceActivity @@ -142,18 +200,47 @@ from .web_activity import WebActivity from .redshift_unload_settings import RedshiftUnloadSettings from .amazon_redshift_source import AmazonRedshiftSource +from .zoho_source import ZohoSource +from .xero_source import XeroSource +from .square_source import SquareSource +from .spark_source import SparkSource +from .shopify_source import ShopifySource +from .service_now_source import ServiceNowSource +from .quick_books_source import QuickBooksSource +from .presto_source import PrestoSource +from .phoenix_source import PhoenixSource +from .paypal_source import PaypalSource +from .marketo_source import MarketoSource +from .maria_db_source import MariaDBSource +from .magento_source import MagentoSource +from .jira_source import JiraSource +from .impala_source import ImpalaSource +from .hubspot_source import HubspotSource +from .hive_source import HiveSource +from .hbase_source import HBaseSource +from .greenplum_source import GreenplumSource +from .google_big_query_source import GoogleBigQuerySource +from .eloqua_source import EloquaSource +from .drill_source import DrillSource +from .couchbase_source import CouchbaseSource +from .concur_source import ConcurSource +from .azure_postgre_sql_source import AzurePostgreSqlSource +from .amazon_mws_source import AmazonMWSSource from .http_source import HttpSource from .azure_data_lake_store_source import AzureDataLakeStoreSource from .mongo_db_source import MongoDbSource from .cassandra_source import CassandraSource from .web_source import WebSource from .oracle_source import OracleSource +from .azure_my_sql_source import AzureMySqlSource from .distcp_settings import DistcpSettings from .hdfs_source import HdfsSource from .file_system_source import FileSystemSource from .sql_dw_source import SqlDWSource from .stored_procedure_parameter import StoredProcedureParameter from .sql_source import SqlSource +from .sap_cloud_for_customer_source import SapCloudForCustomerSource +from .salesforce_source import SalesforceSource from .relational_source import RelationalSource from .dynamics_source import DynamicsSource from .document_db_collection_source import DocumentDbCollectionSource @@ -173,6 +260,7 @@ from .staging_settings import StagingSettings from .tabular_translator import TabularTranslator from .copy_translator import CopyTranslator +from .salesforce_sink import SalesforceSink from .dynamics_sink import DynamicsSink from .odbc_sink import OdbcSink from .azure_search_index_sink import AzureSearchIndexSink @@ -186,6 +274,7 @@ from .blob_sink import BlobSink from .azure_table_sink import AzureTableSink from .azure_queue_sink import AzureQueueSink +from .sap_cloud_for_customer_sink import SapCloudForCustomerSink from .copy_sink import CopySink from .copy_activity import CopyActivity from .execution_activity import ExecutionActivity @@ -231,6 +320,18 @@ PipelineRunQueryOrderByField, PipelineRunQueryOrder, TriggerRunStatus, + SparkServerType, + SparkThriftTransportProtocol, + SparkAuthenticationType, + ServiceNowAuthenticationType, + PrestoAuthenticationType, + PhoenixAuthenticationType, + ImpalaAuthenticationType, + HiveServerType, + HiveThriftTransportProtocol, + HiveAuthenticationType, + HBaseAuthenticationType, + GoogleBigQueryAuthenticationType, SapHanaAuthenticationType, SftpAuthenticationType, FtpAuthenticationType, @@ -242,6 +343,7 @@ SybaseAuthenticationType, DatasetCompressionLevel, JsonFormatFilePattern, + TumblingWindowFrequency, DayOfWeek, DaysOfWeek, RecurrenceFrequency, @@ -249,9 +351,11 @@ CassandraSourceReadConsistencyLevels, StoredProcedureParameterType, HDInsightActivityDebugInfoOption, + SalesforceSinkWriteBehavior, AzureSearchIndexWriteBehaviorType, CopyBehaviorType, PolybaseSettingsRejectType, + SapCloudForCustomerSinkWriteBehavior, SelfHostedIntegrationRuntimeNodeStatus, IntegrationRuntimeUpdateResult, IntegrationRuntimeInternalChannelEncryptionMode, @@ -285,12 +389,12 @@ 'ActivityDependency', 'Activity', 'PipelineResource', - 'PipelineReference', - 'TriggerPipelineReference', 'Trigger', 'TriggerResource', 'CreateRunResponse', 'ErrorResponse', 'ErrorResponseException', + 'PipelineReference', + 'TriggerPipelineReference', 'FactoryUpdateParameters', 'DatasetReference', 'PipelineRunQueryFilter', @@ -310,6 +414,32 @@ 'OperationListResponse', 'AzureDataLakeAnalyticsLinkedService', 'HDInsightOnDemandLinkedService', + 'ZohoLinkedService', + 'XeroLinkedService', + 'SquareLinkedService', + 'SparkLinkedService', + 'ShopifyLinkedService', + 'ServiceNowLinkedService', + 'QuickBooksLinkedService', + 'PrestoLinkedService', + 'PhoenixLinkedService', + 'PaypalLinkedService', + 'MarketoLinkedService', + 'MariaDBLinkedService', + 'MagentoLinkedService', + 'JiraLinkedService', + 'ImpalaLinkedService', + 'HubspotLinkedService', + 'HiveLinkedService', + 'HBaseLinkedService', + 'GreenplumLinkedService', + 'GoogleBigQueryLinkedService', + 'EloquaLinkedService', + 'DrillLinkedService', + 'CouchbaseLinkedService', + 'ConcurLinkedService', + 'AzurePostgreSqlLinkedService', + 'AmazonMWSLinkedService', 'SapHanaLinkedService', 'SapBWLinkedService', 'SftpServerLinkedService', @@ -319,6 +449,7 @@ 'CustomDataSourceLinkedService', 'AmazonRedshiftLinkedService', 'AmazonS3LinkedService', + 'SapCloudForCustomerLinkedService', 'SalesforceLinkedService', 'AzureDataLakeStoreLinkedService', 'MongoDbLinkedService', @@ -337,6 +468,7 @@ 'SybaseLinkedService', 'PostgreSqlLinkedService', 'MySqlLinkedService', + 'AzureMySqlLinkedService', 'OracleLinkedService', 'FileServerLinkedService', 'HDInsightLinkedService', @@ -348,6 +480,33 @@ 'SqlServerLinkedService', 'AzureSqlDWLinkedService', 'AzureStorageLinkedService', + 'ZohoObjectDataset', + 'XeroObjectDataset', + 'SquareObjectDataset', + 'SparkObjectDataset', + 'ShopifyObjectDataset', + 'ServiceNowObjectDataset', + 'QuickBooksObjectDataset', + 'PrestoObjectDataset', + 'PhoenixObjectDataset', + 'PaypalObjectDataset', + 'MarketoObjectDataset', + 'MariaDBTableDataset', + 'MagentoObjectDataset', + 'JiraObjectDataset', + 'ImpalaObjectDataset', + 'HubspotObjectDataset', + 'HiveObjectDataset', + 'HBaseObjectDataset', + 'GreenplumTableDataset', + 'GoogleBigQueryObjectDataset', + 'EloquaObjectDataset', + 'DrillTableDataset', + 'CouchbaseTableDataset', + 'ConcurObjectDataset', + 'AzurePostgreSqlTableDataset', + 'AmazonMWSObjectDataset', + 'DatasetZipDeflateCompression', 'DatasetDeflateCompression', 'DatasetGZipCompression', 'DatasetBZip2Compression', @@ -358,14 +517,14 @@ 'JsonFormat', 'TextFormat', 'DatasetStorageFormat', - 'DatasetDateTimePartitionValue', - 'DatasetPartitionValue', - 'DatasetPartition', 'HttpDataset', 'AzureSearchIndexDataset', 'WebTableDataset', 'SqlServerTableDataset', + 'SapCloudForCustomerResourceDataset', + 'SalesforceObjectDataset', 'RelationalTableDataset', + 'AzureMySqlTableDataset', 'OracleTableDataset', 'ODataResourceDataset', 'MongoDbCollectionDataset', @@ -380,11 +539,14 @@ 'AzureTableDataset', 'AzureBlobDataset', 'AmazonS3Dataset', + 'RetryPolicy', + 'TumblingWindowTrigger', 'BlobTrigger', 'RecurrenceScheduleOccurrence', 'RecurrenceSchedule', 'ScheduleTriggerRecurrence', 'ScheduleTrigger', + 'MultiplePipelineTrigger', 'ActivityPolicy', 'DataLakeAnalyticsUSQLActivity', 'AzureMLUpdateResourceActivity', @@ -395,18 +557,47 @@ 'WebActivity', 'RedshiftUnloadSettings', 'AmazonRedshiftSource', + 'ZohoSource', + 'XeroSource', + 'SquareSource', + 'SparkSource', + 'ShopifySource', + 'ServiceNowSource', + 'QuickBooksSource', + 'PrestoSource', + 'PhoenixSource', + 'PaypalSource', + 'MarketoSource', + 'MariaDBSource', + 'MagentoSource', + 'JiraSource', + 'ImpalaSource', + 'HubspotSource', + 'HiveSource', + 'HBaseSource', + 'GreenplumSource', + 'GoogleBigQuerySource', + 'EloquaSource', + 'DrillSource', + 'CouchbaseSource', + 'ConcurSource', + 'AzurePostgreSqlSource', + 'AmazonMWSSource', 'HttpSource', 'AzureDataLakeStoreSource', 'MongoDbSource', 'CassandraSource', 'WebSource', 'OracleSource', + 'AzureMySqlSource', 'DistcpSettings', 'HdfsSource', 'FileSystemSource', 'SqlDWSource', 'StoredProcedureParameter', 'SqlSource', + 'SapCloudForCustomerSource', + 'SalesforceSource', 'RelationalSource', 'DynamicsSource', 'DocumentDbCollectionSource', @@ -426,6 +617,7 @@ 'StagingSettings', 'TabularTranslator', 'CopyTranslator', + 'SalesforceSink', 'DynamicsSink', 'OdbcSink', 'AzureSearchIndexSink', @@ -439,6 +631,7 @@ 'BlobSink', 'AzureTableSink', 'AzureQueueSink', + 'SapCloudForCustomerSink', 'CopySink', 'CopyActivity', 'ExecutionActivity', @@ -483,6 +676,18 @@ 'PipelineRunQueryOrderByField', 'PipelineRunQueryOrder', 'TriggerRunStatus', + 'SparkServerType', + 'SparkThriftTransportProtocol', + 'SparkAuthenticationType', + 'ServiceNowAuthenticationType', + 'PrestoAuthenticationType', + 'PhoenixAuthenticationType', + 'ImpalaAuthenticationType', + 'HiveServerType', + 'HiveThriftTransportProtocol', + 'HiveAuthenticationType', + 'HBaseAuthenticationType', + 'GoogleBigQueryAuthenticationType', 'SapHanaAuthenticationType', 'SftpAuthenticationType', 'FtpAuthenticationType', @@ -494,6 +699,7 @@ 'SybaseAuthenticationType', 'DatasetCompressionLevel', 'JsonFormatFilePattern', + 'TumblingWindowFrequency', 'DayOfWeek', 'DaysOfWeek', 'RecurrenceFrequency', @@ -501,9 +707,11 @@ 'CassandraSourceReadConsistencyLevels', 'StoredProcedureParameterType', 'HDInsightActivityDebugInfoOption', + 'SalesforceSinkWriteBehavior', 'AzureSearchIndexWriteBehaviorType', 'CopyBehaviorType', 'PolybaseSettingsRejectType', + 'SapCloudForCustomerSinkWriteBehavior', 'SelfHostedIntegrationRuntimeNodeStatus', 'IntegrationRuntimeUpdateResult', 'IntegrationRuntimeInternalChannelEncryptionMode', diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_mws_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_mws_linked_service.py new file mode 100644 index 000000000000..3055e03fd46a --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_mws_linked_service.py @@ -0,0 +1,92 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .linked_service import LinkedService + + +class AmazonMWSLinkedService(LinkedService): + """Amazon Marketplace Web Service linked service. + + :param connect_via: The integration runtime reference. + :type connect_via: + ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :param description: Linked service description. + :type description: str + :param type: Constant filled by server. + :type type: str + :param endpoint: The endpoint of the Amazon MWS server, (i.e. + mws.amazonservices.com) + :type endpoint: object + :param marketplace_id: The Amazon Marketplace ID you want to retrieve data + from. To retrive data from multiple Marketplace IDs, seperate them with a + comma (,). (i.e. A2EUQ1WTGCTBG2) + :type marketplace_id: object + :param seller_id: The Amazon seller ID. + :type seller_id: object + :param mws_auth_token: The Amazon MWS authentication token. + :type mws_auth_token: ~azure.mgmt.datafactory.models.SecretBase + :param access_key_id: The access key id used to access data. + :type access_key_id: object + :param secret_key: The secret key used to access data. + :type secret_key: ~azure.mgmt.datafactory.models.SecretBase + :param use_encrypted_endpoints: Specifies whether the data source + endpoints are encrypted using HTTPS. The default value is true. + :type use_encrypted_endpoints: object + :param use_host_verification: Specifies whether to require the host name + in the server's certificate to match the host name of the server when + connecting over SSL. The default value is true. + :type use_host_verification: object + :param use_peer_verification: Specifies whether to verify the identity of + the server when connecting over SSL. The default value is true. + :type use_peer_verification: object + :param encrypted_credential: The encrypted credential used for + authentication. Credentials are encrypted using the integration runtime + credential manager. Type: string (or Expression with resultType string). + :type encrypted_credential: object + """ + + _validation = { + 'type': {'required': True}, + 'endpoint': {'required': True}, + 'marketplace_id': {'required': True}, + 'seller_id': {'required': True}, + 'access_key_id': {'required': True}, + } + + _attribute_map = { + 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, + 'description': {'key': 'description', 'type': 'str'}, + 'type': {'key': 'type', 'type': 'str'}, + 'endpoint': {'key': 'typeProperties.endpoint', 'type': 'object'}, + 'marketplace_id': {'key': 'typeProperties.marketplaceID', 'type': 'object'}, + 'seller_id': {'key': 'typeProperties.sellerID', 'type': 'object'}, + 'mws_auth_token': {'key': 'typeProperties.mwsAuthToken', 'type': 'SecretBase'}, + 'access_key_id': {'key': 'typeProperties.accessKeyId', 'type': 'object'}, + 'secret_key': {'key': 'typeProperties.secretKey', 'type': 'SecretBase'}, + 'use_encrypted_endpoints': {'key': 'typeProperties.useEncryptedEndpoints', 'type': 'object'}, + 'use_host_verification': {'key': 'typeProperties.useHostVerification', 'type': 'object'}, + 'use_peer_verification': {'key': 'typeProperties.usePeerVerification', 'type': 'object'}, + 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, + } + + def __init__(self, endpoint, marketplace_id, seller_id, access_key_id, connect_via=None, description=None, mws_auth_token=None, secret_key=None, use_encrypted_endpoints=None, use_host_verification=None, use_peer_verification=None, encrypted_credential=None): + super(AmazonMWSLinkedService, self).__init__(connect_via=connect_via, description=description) + self.endpoint = endpoint + self.marketplace_id = marketplace_id + self.seller_id = seller_id + self.mws_auth_token = mws_auth_token + self.access_key_id = access_key_id + self.secret_key = secret_key + self.use_encrypted_endpoints = use_encrypted_endpoints + self.use_host_verification = use_host_verification + self.use_peer_verification = use_peer_verification + self.encrypted_credential = encrypted_credential + self.type = 'AmazonMWS' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_mws_object_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_mws_object_dataset.py new file mode 100644 index 000000000000..1d01da91822d --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_mws_object_dataset.py @@ -0,0 +1,40 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .dataset import Dataset + + +class AmazonMWSObjectDataset(Dataset): + """Amazon Marketplace Web Service dataset. + + :param description: Dataset description. + :type description: str + :param structure: Columns that define the structure of the dataset. Type: + array (or Expression with resultType array), itemType: DatasetDataElement. + :type structure: object + :param linked_service_name: Linked service reference. + :type linked_service_name: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param parameters: Parameters for dataset. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param type: Constant filled by server. + :type type: str + """ + + _validation = { + 'linked_service_name': {'required': True}, + 'type': {'required': True}, + } + + def __init__(self, linked_service_name, description=None, structure=None, parameters=None): + super(AmazonMWSObjectDataset, self).__init__(description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters) + self.type = 'AmazonMWSObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_mws_source.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_mws_source.py new file mode 100644 index 000000000000..4333a2c92394 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/amazon_mws_source.py @@ -0,0 +1,46 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .copy_source import CopySource + + +class AmazonMWSSource(CopySource): + """A copy activity Amazon Marketplace Web Service source. + + :param source_retry_count: Source retry count. Type: integer (or + Expression with resultType integer). + :type source_retry_count: object + :param source_retry_wait: Source retry wait. Type: string (or Expression + with resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type source_retry_wait: object + :param type: Constant filled by server. + :type type: str + :param query: A query to retrieve data from source. Type: string (or + Expression with resultType string). + :type query: object + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, + 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, + 'query': {'key': 'query', 'type': 'object'}, + } + + def __init__(self, source_retry_count=None, source_retry_wait=None, query=None): + super(AmazonMWSSource, self).__init__(source_retry_count=source_retry_count, source_retry_wait=source_retry_wait) + self.query = query + self.type = 'AmazonMWSSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_data_lake_store_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_data_lake_store_dataset.py index e7baf6015c7f..5ed9af729242 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_data_lake_store_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_data_lake_store_dataset.py @@ -39,10 +39,6 @@ class AzureDataLakeStoreDataset(Dataset): :param compression: The data compression method used for the item(s) in the Azure Data Lake Store. :type compression: ~azure.mgmt.datafactory.models.DatasetCompression - :param partitioned_by: Specify a dynamic path and filename for time series - data. - :type partitioned_by: - list[~azure.mgmt.datafactory.models.DatasetPartition] """ _validation = { @@ -61,14 +57,12 @@ class AzureDataLakeStoreDataset(Dataset): 'file_name': {'key': 'typeProperties.fileName', 'type': 'object'}, 'format': {'key': 'typeProperties.format', 'type': 'DatasetStorageFormat'}, 'compression': {'key': 'typeProperties.compression', 'type': 'DatasetCompression'}, - 'partitioned_by': {'key': 'typeProperties.partitionedBy', 'type': '[DatasetPartition]'}, } - def __init__(self, linked_service_name, folder_path, description=None, structure=None, parameters=None, file_name=None, format=None, compression=None, partitioned_by=None): + def __init__(self, linked_service_name, folder_path, description=None, structure=None, parameters=None, file_name=None, format=None, compression=None): super(AzureDataLakeStoreDataset, self).__init__(description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters) self.folder_path = folder_path self.file_name = file_name self.format = format self.compression = compression - self.partitioned_by = partitioned_by self.type = 'AzureDataLakeStoreFile' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_my_sql_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_my_sql_linked_service.py new file mode 100644 index 000000000000..deaf8f28ff2f --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_my_sql_linked_service.py @@ -0,0 +1,50 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .linked_service import LinkedService + + +class AzureMySqlLinkedService(LinkedService): + """Azure MySQL database linked service. + + :param connect_via: The integration runtime reference. + :type connect_via: + ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :param description: Linked service description. + :type description: str + :param type: Constant filled by server. + :type type: str + :param connection_string: The connection string. + :type connection_string: ~azure.mgmt.datafactory.models.SecureString + :param encrypted_credential: The encrypted credential used for + authentication. Credentials are encrypted using the integration runtime + credential manager. Type: string (or Expression with resultType string). + :type encrypted_credential: object + """ + + _validation = { + 'type': {'required': True}, + 'connection_string': {'required': True}, + } + + _attribute_map = { + 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, + 'description': {'key': 'description', 'type': 'str'}, + 'type': {'key': 'type', 'type': 'str'}, + 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'SecureString'}, + 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, + } + + def __init__(self, connection_string, connect_via=None, description=None, encrypted_credential=None): + super(AzureMySqlLinkedService, self).__init__(connect_via=connect_via, description=description) + self.connection_string = connection_string + self.encrypted_credential = encrypted_credential + self.type = 'AzureMySql' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_my_sql_source.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_my_sql_source.py new file mode 100644 index 000000000000..76f9c3b97781 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_my_sql_source.py @@ -0,0 +1,46 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .copy_source import CopySource + + +class AzureMySqlSource(CopySource): + """A copy activity Azure MySQL source. + + :param source_retry_count: Source retry count. Type: integer (or + Expression with resultType integer). + :type source_retry_count: object + :param source_retry_wait: Source retry wait. Type: string (or Expression + with resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type source_retry_wait: object + :param type: Constant filled by server. + :type type: str + :param query: Database query. Type: string (or Expression with resultType + string). + :type query: object + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, + 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, + 'query': {'key': 'query', 'type': 'object'}, + } + + def __init__(self, source_retry_count=None, source_retry_wait=None, query=None): + super(AzureMySqlSource, self).__init__(source_retry_count=source_retry_count, source_retry_wait=source_retry_wait) + self.query = query + self.type = 'AzureMySqlSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_my_sql_table_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_my_sql_table_dataset.py new file mode 100644 index 000000000000..d4b71feb4338 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_my_sql_table_dataset.py @@ -0,0 +1,53 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .dataset import Dataset + + +class AzureMySqlTableDataset(Dataset): + """The Azure MySQL database dataset. + + :param description: Dataset description. + :type description: str + :param structure: Columns that define the structure of the dataset. Type: + array (or Expression with resultType array), itemType: DatasetDataElement. + :type structure: object + :param linked_service_name: Linked service reference. + :type linked_service_name: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param parameters: Parameters for dataset. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param type: Constant filled by server. + :type type: str + :param table_name: The Azure MySQL database table name. Type: string (or + Expression with resultType string). + :type table_name: object + """ + + _validation = { + 'linked_service_name': {'required': True}, + 'type': {'required': True}, + } + + _attribute_map = { + 'description': {'key': 'description', 'type': 'str'}, + 'structure': {'key': 'structure', 'type': 'object'}, + 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'type': {'key': 'type', 'type': 'str'}, + 'table_name': {'key': 'typeProperties.tableName', 'type': 'object'}, + } + + def __init__(self, linked_service_name, description=None, structure=None, parameters=None, table_name=None): + super(AzureMySqlTableDataset, self).__init__(description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters) + self.table_name = table_name + self.type = 'AzureMySqlTable' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_postgre_sql_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_postgre_sql_linked_service.py new file mode 100644 index 000000000000..f5cee5707630 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_postgre_sql_linked_service.py @@ -0,0 +1,49 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .linked_service import LinkedService + + +class AzurePostgreSqlLinkedService(LinkedService): + """Azure PostgreSQL linked service. + + :param connect_via: The integration runtime reference. + :type connect_via: + ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :param description: Linked service description. + :type description: str + :param type: Constant filled by server. + :type type: str + :param connection_string: An ODBC connection string. + :type connection_string: ~azure.mgmt.datafactory.models.SecretBase + :param encrypted_credential: The encrypted credential used for + authentication. Credentials are encrypted using the integration runtime + credential manager. Type: string (or Expression with resultType string). + :type encrypted_credential: object + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, + 'description': {'key': 'description', 'type': 'str'}, + 'type': {'key': 'type', 'type': 'str'}, + 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'SecretBase'}, + 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, + } + + def __init__(self, connect_via=None, description=None, connection_string=None, encrypted_credential=None): + super(AzurePostgreSqlLinkedService, self).__init__(connect_via=connect_via, description=description) + self.connection_string = connection_string + self.encrypted_credential = encrypted_credential + self.type = 'AzurePostgreSql' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_postgre_sql_source.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_postgre_sql_source.py new file mode 100644 index 000000000000..080247b481bd --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_postgre_sql_source.py @@ -0,0 +1,46 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .copy_source import CopySource + + +class AzurePostgreSqlSource(CopySource): + """A copy activity Azure PostgreSQL source. + + :param source_retry_count: Source retry count. Type: integer (or + Expression with resultType integer). + :type source_retry_count: object + :param source_retry_wait: Source retry wait. Type: string (or Expression + with resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type source_retry_wait: object + :param type: Constant filled by server. + :type type: str + :param query: A query to retrieve data from source. Type: string (or + Expression with resultType string). + :type query: object + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, + 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, + 'query': {'key': 'query', 'type': 'object'}, + } + + def __init__(self, source_retry_count=None, source_retry_wait=None, query=None): + super(AzurePostgreSqlSource, self).__init__(source_retry_count=source_retry_count, source_retry_wait=source_retry_wait) + self.query = query + self.type = 'AzurePostgreSqlSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_postgre_sql_table_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_postgre_sql_table_dataset.py new file mode 100644 index 000000000000..7b2db4ffdeb9 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/azure_postgre_sql_table_dataset.py @@ -0,0 +1,40 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .dataset import Dataset + + +class AzurePostgreSqlTableDataset(Dataset): + """Azure PostgreSQL dataset. + + :param description: Dataset description. + :type description: str + :param structure: Columns that define the structure of the dataset. Type: + array (or Expression with resultType array), itemType: DatasetDataElement. + :type structure: object + :param linked_service_name: Linked service reference. + :type linked_service_name: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param parameters: Parameters for dataset. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param type: Constant filled by server. + :type type: str + """ + + _validation = { + 'linked_service_name': {'required': True}, + 'type': {'required': True}, + } + + def __init__(self, linked_service_name, description=None, structure=None, parameters=None): + super(AzurePostgreSqlTableDataset, self).__init__(description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters) + self.type = 'AzurePostgreSqlTable' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/blob_trigger.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/blob_trigger.py index fabaca733e0c..365a86a86f88 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/blob_trigger.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/blob_trigger.py @@ -9,10 +9,10 @@ # regenerated. # -------------------------------------------------------------------------- -from .trigger import Trigger +from .multiple_pipeline_trigger import MultiplePipelineTrigger -class BlobTrigger(Trigger): +class BlobTrigger(MultiplePipelineTrigger): """Trigger that runs everytime the selected Blob container changes. Variables are only populated by the server, and will be ignored when @@ -20,9 +20,6 @@ class BlobTrigger(Trigger): :param description: Trigger description. :type description: str - :param pipelines: Pipelines that need to be started. - :type pipelines: - list[~azure.mgmt.datafactory.models.TriggerPipelineReference] :ivar runtime_state: Indicates if trigger is running or not. Updated when Start/Stop APIs are called on the Trigger. Possible values include: 'Started', 'Stopped', 'Disabled' @@ -30,6 +27,9 @@ class BlobTrigger(Trigger): ~azure.mgmt.datafactory.models.TriggerRuntimeState :param type: Constant filled by server. :type type: str + :param pipelines: Pipelines that need to be started. + :type pipelines: + list[~azure.mgmt.datafactory.models.TriggerPipelineReference] :param folder_path: The path of the container/folder that will trigger the pipeline. :type folder_path: str @@ -48,9 +48,9 @@ class BlobTrigger(Trigger): _attribute_map = { 'description': {'key': 'description', 'type': 'str'}, - 'pipelines': {'key': 'pipelines', 'type': '[TriggerPipelineReference]'}, 'runtime_state': {'key': 'runtimeState', 'type': 'str'}, 'type': {'key': 'type', 'type': 'str'}, + 'pipelines': {'key': 'pipelines', 'type': '[TriggerPipelineReference]'}, 'folder_path': {'key': 'typeProperties.folderPath', 'type': 'str'}, 'max_concurrency': {'key': 'typeProperties.maxConcurrency', 'type': 'int'}, 'linked_service': {'key': 'typeProperties.linkedService', 'type': 'LinkedServiceReference'}, diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/concur_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/concur_linked_service.py new file mode 100644 index 000000000000..4a13671ba7ae --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/concur_linked_service.py @@ -0,0 +1,76 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .linked_service import LinkedService + + +class ConcurLinkedService(LinkedService): + """Concur Serivce linked service. + + :param connect_via: The integration runtime reference. + :type connect_via: + ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :param description: Linked service description. + :type description: str + :param type: Constant filled by server. + :type type: str + :param client_id: Application client_id supplied by Concur App Management. + :type client_id: object + :param username: The user name that you use to access Concur Service. + :type username: object + :param password: The password corresponding to the user name that you + provided in the username field. + :type password: ~azure.mgmt.datafactory.models.SecretBase + :param use_encrypted_endpoints: Specifies whether the data source + endpoints are encrypted using HTTPS. The default value is true. + :type use_encrypted_endpoints: object + :param use_host_verification: Specifies whether to require the host name + in the server's certificate to match the host name of the server when + connecting over SSL. The default value is true. + :type use_host_verification: object + :param use_peer_verification: Specifies whether to verify the identity of + the server when connecting over SSL. The default value is true. + :type use_peer_verification: object + :param encrypted_credential: The encrypted credential used for + authentication. Credentials are encrypted using the integration runtime + credential manager. Type: string (or Expression with resultType string). + :type encrypted_credential: object + """ + + _validation = { + 'type': {'required': True}, + 'client_id': {'required': True}, + 'username': {'required': True}, + } + + _attribute_map = { + 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, + 'description': {'key': 'description', 'type': 'str'}, + 'type': {'key': 'type', 'type': 'str'}, + 'client_id': {'key': 'typeProperties.clientId', 'type': 'object'}, + 'username': {'key': 'typeProperties.username', 'type': 'object'}, + 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, + 'use_encrypted_endpoints': {'key': 'typeProperties.useEncryptedEndpoints', 'type': 'object'}, + 'use_host_verification': {'key': 'typeProperties.useHostVerification', 'type': 'object'}, + 'use_peer_verification': {'key': 'typeProperties.usePeerVerification', 'type': 'object'}, + 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, + } + + def __init__(self, client_id, username, connect_via=None, description=None, password=None, use_encrypted_endpoints=None, use_host_verification=None, use_peer_verification=None, encrypted_credential=None): + super(ConcurLinkedService, self).__init__(connect_via=connect_via, description=description) + self.client_id = client_id + self.username = username + self.password = password + self.use_encrypted_endpoints = use_encrypted_endpoints + self.use_host_verification = use_host_verification + self.use_peer_verification = use_peer_verification + self.encrypted_credential = encrypted_credential + self.type = 'Concur' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/concur_object_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/concur_object_dataset.py new file mode 100644 index 000000000000..f4f2bb407759 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/concur_object_dataset.py @@ -0,0 +1,40 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .dataset import Dataset + + +class ConcurObjectDataset(Dataset): + """Concur Serivce dataset. + + :param description: Dataset description. + :type description: str + :param structure: Columns that define the structure of the dataset. Type: + array (or Expression with resultType array), itemType: DatasetDataElement. + :type structure: object + :param linked_service_name: Linked service reference. + :type linked_service_name: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param parameters: Parameters for dataset. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param type: Constant filled by server. + :type type: str + """ + + _validation = { + 'linked_service_name': {'required': True}, + 'type': {'required': True}, + } + + def __init__(self, linked_service_name, description=None, structure=None, parameters=None): + super(ConcurObjectDataset, self).__init__(description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters) + self.type = 'ConcurObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/concur_source.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/concur_source.py new file mode 100644 index 000000000000..b55b0471f2ef --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/concur_source.py @@ -0,0 +1,46 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .copy_source import CopySource + + +class ConcurSource(CopySource): + """A copy activity Concur Serivce source. + + :param source_retry_count: Source retry count. Type: integer (or + Expression with resultType integer). + :type source_retry_count: object + :param source_retry_wait: Source retry wait. Type: string (or Expression + with resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type source_retry_wait: object + :param type: Constant filled by server. + :type type: str + :param query: A query to retrieve data from source. Type: string (or + Expression with resultType string). + :type query: object + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, + 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, + 'query': {'key': 'query', 'type': 'object'}, + } + + def __init__(self, source_retry_count=None, source_retry_wait=None, query=None): + super(ConcurSource, self).__init__(source_retry_count=source_retry_count, source_retry_wait=source_retry_wait) + self.query = query + self.type = 'ConcurSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/copy_sink.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/copy_sink.py index ec8b62e38a06..02e8acca16ea 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/copy_sink.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/copy_sink.py @@ -16,10 +16,10 @@ class CopySink(Model): """A copy activity sink. You probably want to use the sub-classes and not this class directly. Known - sub-classes are: DynamicsSink, OdbcSink, AzureSearchIndexSink, - AzureDataLakeStoreSink, OracleSink, SqlDWSink, SqlSink, - DocumentDbCollectionSink, FileSystemSink, BlobSink, AzureTableSink, - AzureQueueSink + sub-classes are: SalesforceSink, DynamicsSink, OdbcSink, + AzureSearchIndexSink, AzureDataLakeStoreSink, OracleSink, SqlDWSink, + SqlSink, DocumentDbCollectionSink, FileSystemSink, BlobSink, + AzureTableSink, AzureQueueSink, SapCloudForCustomerSink :param write_batch_size: Write batch size. Type: integer (or Expression with resultType integer), minimum: 0. @@ -52,7 +52,7 @@ class CopySink(Model): } _subtype_map = { - 'type': {'DynamicsSink': 'DynamicsSink', 'OdbcSink': 'OdbcSink', 'AzureSearchIndexSink': 'AzureSearchIndexSink', 'AzureDataLakeStoreSink': 'AzureDataLakeStoreSink', 'OracleSink': 'OracleSink', 'SqlDWSink': 'SqlDWSink', 'SqlSink': 'SqlSink', 'DocumentDbCollectionSink': 'DocumentDbCollectionSink', 'FileSystemSink': 'FileSystemSink', 'BlobSink': 'BlobSink', 'AzureTableSink': 'AzureTableSink', 'AzureQueueSink': 'AzureQueueSink'} + 'type': {'SalesforceSink': 'SalesforceSink', 'DynamicsSink': 'DynamicsSink', 'OdbcSink': 'OdbcSink', 'AzureSearchIndexSink': 'AzureSearchIndexSink', 'AzureDataLakeStoreSink': 'AzureDataLakeStoreSink', 'OracleSink': 'OracleSink', 'SqlDWSink': 'SqlDWSink', 'SqlSink': 'SqlSink', 'DocumentDbCollectionSink': 'DocumentDbCollectionSink', 'FileSystemSink': 'FileSystemSink', 'BlobSink': 'BlobSink', 'AzureTableSink': 'AzureTableSink', 'AzureQueueSink': 'AzureQueueSink', 'SapCloudForCustomerSink': 'SapCloudForCustomerSink'} } def __init__(self, write_batch_size=None, write_batch_timeout=None, sink_retry_count=None, sink_retry_wait=None): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/copy_source.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/copy_source.py index c72f33c7f4c8..f120795e93a8 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/copy_source.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/copy_source.py @@ -16,11 +16,17 @@ class CopySource(Model): """A copy activity source. You probably want to use the sub-classes and not this class directly. Known - sub-classes are: AmazonRedshiftSource, HttpSource, + sub-classes are: AmazonRedshiftSource, ZohoSource, XeroSource, + SquareSource, SparkSource, ShopifySource, ServiceNowSource, + QuickBooksSource, PrestoSource, PhoenixSource, PaypalSource, MarketoSource, + MariaDBSource, MagentoSource, JiraSource, ImpalaSource, HubspotSource, + HiveSource, HBaseSource, GreenplumSource, GoogleBigQuerySource, + EloquaSource, DrillSource, CouchbaseSource, ConcurSource, + AzurePostgreSqlSource, AmazonMWSSource, HttpSource, AzureDataLakeStoreSource, MongoDbSource, CassandraSource, WebSource, - OracleSource, HdfsSource, FileSystemSource, SqlDWSource, SqlSource, - RelationalSource, DynamicsSource, DocumentDbCollectionSource, BlobSource, - AzureTableSource + OracleSource, AzureMySqlSource, HdfsSource, FileSystemSource, SqlDWSource, + SqlSource, SapCloudForCustomerSource, SalesforceSource, RelationalSource, + DynamicsSource, DocumentDbCollectionSource, BlobSource, AzureTableSource :param source_retry_count: Source retry count. Type: integer (or Expression with resultType integer). @@ -44,7 +50,7 @@ class CopySource(Model): } _subtype_map = { - 'type': {'AmazonRedshiftSource': 'AmazonRedshiftSource', 'HttpSource': 'HttpSource', 'AzureDataLakeStoreSource': 'AzureDataLakeStoreSource', 'MongoDbSource': 'MongoDbSource', 'CassandraSource': 'CassandraSource', 'WebSource': 'WebSource', 'OracleSource': 'OracleSource', 'HdfsSource': 'HdfsSource', 'FileSystemSource': 'FileSystemSource', 'SqlDWSource': 'SqlDWSource', 'SqlSource': 'SqlSource', 'RelationalSource': 'RelationalSource', 'DynamicsSource': 'DynamicsSource', 'DocumentDbCollectionSource': 'DocumentDbCollectionSource', 'BlobSource': 'BlobSource', 'AzureTableSource': 'AzureTableSource'} + 'type': {'AmazonRedshiftSource': 'AmazonRedshiftSource', 'ZohoSource': 'ZohoSource', 'XeroSource': 'XeroSource', 'SquareSource': 'SquareSource', 'SparkSource': 'SparkSource', 'ShopifySource': 'ShopifySource', 'ServiceNowSource': 'ServiceNowSource', 'QuickBooksSource': 'QuickBooksSource', 'PrestoSource': 'PrestoSource', 'PhoenixSource': 'PhoenixSource', 'PaypalSource': 'PaypalSource', 'MarketoSource': 'MarketoSource', 'MariaDBSource': 'MariaDBSource', 'MagentoSource': 'MagentoSource', 'JiraSource': 'JiraSource', 'ImpalaSource': 'ImpalaSource', 'HubspotSource': 'HubspotSource', 'HiveSource': 'HiveSource', 'HBaseSource': 'HBaseSource', 'GreenplumSource': 'GreenplumSource', 'GoogleBigQuerySource': 'GoogleBigQuerySource', 'EloquaSource': 'EloquaSource', 'DrillSource': 'DrillSource', 'CouchbaseSource': 'CouchbaseSource', 'ConcurSource': 'ConcurSource', 'AzurePostgreSqlSource': 'AzurePostgreSqlSource', 'AmazonMWSSource': 'AmazonMWSSource', 'HttpSource': 'HttpSource', 'AzureDataLakeStoreSource': 'AzureDataLakeStoreSource', 'MongoDbSource': 'MongoDbSource', 'CassandraSource': 'CassandraSource', 'WebSource': 'WebSource', 'OracleSource': 'OracleSource', 'AzureMySqlSource': 'AzureMySqlSource', 'HdfsSource': 'HdfsSource', 'FileSystemSource': 'FileSystemSource', 'SqlDWSource': 'SqlDWSource', 'SqlSource': 'SqlSource', 'SapCloudForCustomerSource': 'SapCloudForCustomerSource', 'SalesforceSource': 'SalesforceSource', 'RelationalSource': 'RelationalSource', 'DynamicsSource': 'DynamicsSource', 'DocumentDbCollectionSource': 'DocumentDbCollectionSource', 'BlobSource': 'BlobSource', 'AzureTableSource': 'AzureTableSource'} } def __init__(self, source_retry_count=None, source_retry_wait=None): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/couchbase_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/couchbase_linked_service.py new file mode 100644 index 000000000000..7cc7d29f453d --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/couchbase_linked_service.py @@ -0,0 +1,49 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .linked_service import LinkedService + + +class CouchbaseLinkedService(LinkedService): + """Couchbase server linked service. + + :param connect_via: The integration runtime reference. + :type connect_via: + ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :param description: Linked service description. + :type description: str + :param type: Constant filled by server. + :type type: str + :param connection_string: An ODBC connection string. + :type connection_string: ~azure.mgmt.datafactory.models.SecretBase + :param encrypted_credential: The encrypted credential used for + authentication. Credentials are encrypted using the integration runtime + credential manager. Type: string (or Expression with resultType string). + :type encrypted_credential: object + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, + 'description': {'key': 'description', 'type': 'str'}, + 'type': {'key': 'type', 'type': 'str'}, + 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'SecretBase'}, + 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, + } + + def __init__(self, connect_via=None, description=None, connection_string=None, encrypted_credential=None): + super(CouchbaseLinkedService, self).__init__(connect_via=connect_via, description=description) + self.connection_string = connection_string + self.encrypted_credential = encrypted_credential + self.type = 'Couchbase' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/couchbase_source.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/couchbase_source.py new file mode 100644 index 000000000000..9befbbe03859 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/couchbase_source.py @@ -0,0 +1,46 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .copy_source import CopySource + + +class CouchbaseSource(CopySource): + """A copy activity Couchbase server source. + + :param source_retry_count: Source retry count. Type: integer (or + Expression with resultType integer). + :type source_retry_count: object + :param source_retry_wait: Source retry wait. Type: string (or Expression + with resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type source_retry_wait: object + :param type: Constant filled by server. + :type type: str + :param query: A query to retrieve data from source. Type: string (or + Expression with resultType string). + :type query: object + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, + 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, + 'query': {'key': 'query', 'type': 'object'}, + } + + def __init__(self, source_retry_count=None, source_retry_wait=None, query=None): + super(CouchbaseSource, self).__init__(source_retry_count=source_retry_count, source_retry_wait=source_retry_wait) + self.query = query + self.type = 'CouchbaseSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/couchbase_table_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/couchbase_table_dataset.py new file mode 100644 index 000000000000..b0be57c46879 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/couchbase_table_dataset.py @@ -0,0 +1,40 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .dataset import Dataset + + +class CouchbaseTableDataset(Dataset): + """Couchbase server dataset. + + :param description: Dataset description. + :type description: str + :param structure: Columns that define the structure of the dataset. Type: + array (or Expression with resultType array), itemType: DatasetDataElement. + :type structure: object + :param linked_service_name: Linked service reference. + :type linked_service_name: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param parameters: Parameters for dataset. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param type: Constant filled by server. + :type type: str + """ + + _validation = { + 'linked_service_name': {'required': True}, + 'type': {'required': True}, + } + + def __init__(self, linked_service_name, description=None, structure=None, parameters=None): + super(CouchbaseTableDataset, self).__init__(description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters) + self.type = 'CouchbaseTable' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/data_factory_management_client_enums.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/data_factory_management_client_enums.py index d2eeebeb0e92..309251f97dbe 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/data_factory_management_client_enums.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/data_factory_management_client_enums.py @@ -85,6 +85,88 @@ class TriggerRunStatus(Enum): inprogress = "Inprogress" +class SparkServerType(Enum): + + shark_server = "SharkServer" + shark_server2 = "SharkServer2" + spark_thrift_server = "SparkThriftServer" + + +class SparkThriftTransportProtocol(Enum): + + binary = "Binary" + sasl = "SASL" + http = "HTTP " + + +class SparkAuthenticationType(Enum): + + anonymous = "Anonymous" + username = "Username" + username_and_password = "UsernameAndPassword" + windows_azure_hd_insight_service = "WindowsAzureHDInsightService" + + +class ServiceNowAuthenticationType(Enum): + + basic = "Basic" + oauth2 = "OAuth2" + + +class PrestoAuthenticationType(Enum): + + anonymous = "Anonymous" + ldap = "LDAP" + + +class PhoenixAuthenticationType(Enum): + + anonymous = "Anonymous" + username_and_password = "UsernameAndPassword" + windows_azure_hd_insight_service = "WindowsAzureHDInsightService" + + +class ImpalaAuthenticationType(Enum): + + anonymous = "Anonymous" + sasl_username = "SASLUsername" + username_and_password = "UsernameAndPassword" + + +class HiveServerType(Enum): + + hive_server1 = "HiveServer1" + hive_server2 = "HiveServer2" + hive_thrift_server = "HiveThriftServer" + + +class HiveThriftTransportProtocol(Enum): + + binary = "Binary" + sasl = "SASL" + http = "HTTP " + + +class HiveAuthenticationType(Enum): + + anonymous = "Anonymous" + username = "Username" + username_and_password = "UsernameAndPassword" + windows_azure_hd_insight_service = "WindowsAzureHDInsightService" + + +class HBaseAuthenticationType(Enum): + + anonymous = "Anonymous" + basic = "Basic" + + +class GoogleBigQueryAuthenticationType(Enum): + + service_authentication = "ServiceAuthentication" + user_authentication = "UserAuthentication" + + class SapHanaAuthenticationType(Enum): basic = "Basic" @@ -153,6 +235,12 @@ class JsonFormatFilePattern(Enum): array_of_objects = "arrayOfObjects" +class TumblingWindowFrequency(Enum): + + minute = "Minute" + hour = "Hour" + + class DayOfWeek(Enum): sunday = "Sunday" @@ -224,6 +312,12 @@ class HDInsightActivityDebugInfoOption(Enum): failure = "Failure" +class SalesforceSinkWriteBehavior(Enum): + + insert = "Insert" + upsert = "Upsert" + + class AzureSearchIndexWriteBehaviorType(Enum): merge = "Merge" @@ -243,6 +337,12 @@ class PolybaseSettingsRejectType(Enum): percentage = "percentage" +class SapCloudForCustomerSinkWriteBehavior(Enum): + + insert = "Insert" + update = "Update" + + class SelfHostedIntegrationRuntimeNodeStatus(Enum): need_registration = "NeedRegistration" diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset.py index eba95d640db1..b4b5fb4c0401 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset.py @@ -17,8 +17,18 @@ class Dataset(Model): data stores, such as tables, files, folders, and documents. You probably want to use the sub-classes and not this class directly. Known - sub-classes are: HttpDataset, AzureSearchIndexDataset, WebTableDataset, - SqlServerTableDataset, RelationalTableDataset, OracleTableDataset, + sub-classes are: ZohoObjectDataset, XeroObjectDataset, SquareObjectDataset, + SparkObjectDataset, ShopifyObjectDataset, ServiceNowObjectDataset, + QuickBooksObjectDataset, PrestoObjectDataset, PhoenixObjectDataset, + PaypalObjectDataset, MarketoObjectDataset, MariaDBTableDataset, + MagentoObjectDataset, JiraObjectDataset, ImpalaObjectDataset, + HubspotObjectDataset, HiveObjectDataset, HBaseObjectDataset, + GreenplumTableDataset, GoogleBigQueryObjectDataset, EloquaObjectDataset, + DrillTableDataset, CouchbaseTableDataset, ConcurObjectDataset, + AzurePostgreSqlTableDataset, AmazonMWSObjectDataset, HttpDataset, + AzureSearchIndexDataset, WebTableDataset, SqlServerTableDataset, + SapCloudForCustomerResourceDataset, SalesforceObjectDataset, + RelationalTableDataset, AzureMySqlTableDataset, OracleTableDataset, ODataResourceDataset, MongoDbCollectionDataset, FileShareDataset, AzureDataLakeStoreDataset, DynamicsEntityDataset, DocumentDbCollectionDataset, CustomDataset, CassandraTableDataset, @@ -54,7 +64,7 @@ class Dataset(Model): } _subtype_map = { - 'type': {'HttpFile': 'HttpDataset', 'AzureSearchIndex': 'AzureSearchIndexDataset', 'WebTable': 'WebTableDataset', 'SqlServerTable': 'SqlServerTableDataset', 'RelationalTable': 'RelationalTableDataset', 'OracleTable': 'OracleTableDataset', 'ODataResource': 'ODataResourceDataset', 'MongoDbCollection': 'MongoDbCollectionDataset', 'FileShare': 'FileShareDataset', 'AzureDataLakeStoreFile': 'AzureDataLakeStoreDataset', 'DynamicsEntity': 'DynamicsEntityDataset', 'DocumentDbCollection': 'DocumentDbCollectionDataset', 'CustomDataset': 'CustomDataset', 'CassandraTable': 'CassandraTableDataset', 'AzureSqlDWTable': 'AzureSqlDWTableDataset', 'AzureSqlTable': 'AzureSqlTableDataset', 'AzureTable': 'AzureTableDataset', 'AzureBlob': 'AzureBlobDataset', 'AmazonS3Object': 'AmazonS3Dataset'} + 'type': {'ZohoObject': 'ZohoObjectDataset', 'XeroObject': 'XeroObjectDataset', 'SquareObject': 'SquareObjectDataset', 'SparkObject': 'SparkObjectDataset', 'ShopifyObject': 'ShopifyObjectDataset', 'ServiceNowObject': 'ServiceNowObjectDataset', 'QuickBooksObject': 'QuickBooksObjectDataset', 'PrestoObject': 'PrestoObjectDataset', 'PhoenixObject': 'PhoenixObjectDataset', 'PaypalObject': 'PaypalObjectDataset', 'MarketoObject': 'MarketoObjectDataset', 'MariaDBTable': 'MariaDBTableDataset', 'MagentoObject': 'MagentoObjectDataset', 'JiraObject': 'JiraObjectDataset', 'ImpalaObject': 'ImpalaObjectDataset', 'HubspotObject': 'HubspotObjectDataset', 'HiveObject': 'HiveObjectDataset', 'HBaseObject': 'HBaseObjectDataset', 'GreenplumTable': 'GreenplumTableDataset', 'GoogleBigQueryObject': 'GoogleBigQueryObjectDataset', 'EloquaObject': 'EloquaObjectDataset', 'DrillTable': 'DrillTableDataset', 'CouchbaseTable': 'CouchbaseTableDataset', 'ConcurObject': 'ConcurObjectDataset', 'AzurePostgreSqlTable': 'AzurePostgreSqlTableDataset', 'AmazonMWSObject': 'AmazonMWSObjectDataset', 'HttpFile': 'HttpDataset', 'AzureSearchIndex': 'AzureSearchIndexDataset', 'WebTable': 'WebTableDataset', 'SqlServerTable': 'SqlServerTableDataset', 'SapCloudForCustomerResource': 'SapCloudForCustomerResourceDataset', 'SalesforceObject': 'SalesforceObjectDataset', 'RelationalTable': 'RelationalTableDataset', 'AzureMySqlTable': 'AzureMySqlTableDataset', 'OracleTable': 'OracleTableDataset', 'ODataResource': 'ODataResourceDataset', 'MongoDbCollection': 'MongoDbCollectionDataset', 'FileShare': 'FileShareDataset', 'AzureDataLakeStoreFile': 'AzureDataLakeStoreDataset', 'DynamicsEntity': 'DynamicsEntityDataset', 'DocumentDbCollection': 'DocumentDbCollectionDataset', 'CustomDataset': 'CustomDataset', 'CassandraTable': 'CassandraTableDataset', 'AzureSqlDWTable': 'AzureSqlDWTableDataset', 'AzureSqlTable': 'AzureSqlTableDataset', 'AzureTable': 'AzureTableDataset', 'AzureBlob': 'AzureBlobDataset', 'AmazonS3Object': 'AmazonS3Dataset'} } def __init__(self, linked_service_name, description=None, structure=None, parameters=None): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_compression.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_compression.py index 499f0e92b7c4..d6d38d1e07db 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_compression.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_compression.py @@ -16,8 +16,8 @@ class DatasetCompression(Model): """The compression method used on a dataset. You probably want to use the sub-classes and not this class directly. Known - sub-classes are: DatasetDeflateCompression, DatasetGZipCompression, - DatasetBZip2Compression + sub-classes are: DatasetZipDeflateCompression, DatasetDeflateCompression, + DatasetGZipCompression, DatasetBZip2Compression :param type: Constant filled by server. :type type: str @@ -32,7 +32,7 @@ class DatasetCompression(Model): } _subtype_map = { - 'type': {'Deflate': 'DatasetDeflateCompression', 'GZip': 'DatasetGZipCompression', 'BZip2': 'DatasetBZip2Compression'} + 'type': {'ZipDeflate': 'DatasetZipDeflateCompression', 'Deflate': 'DatasetDeflateCompression', 'GZip': 'DatasetGZipCompression', 'BZip2': 'DatasetBZip2Compression'} } def __init__(self): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_date_time_partition_value.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_date_time_partition_value.py deleted file mode 100644 index 3bdaa7b3c877..000000000000 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_date_time_partition_value.py +++ /dev/null @@ -1,42 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for -# license information. -# -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is -# regenerated. -# -------------------------------------------------------------------------- - -from .dataset_partition_value import DatasetPartitionValue - - -class DatasetDateTimePartitionValue(DatasetPartitionValue): - """The date/time value of a partition. - - :param type: Constant filled by server. - :type type: str - :param date_property: Name of variable containing date. Type: string (or - Expression with resultType string). - :type date_property: object - :param format: Format string for the Date value. Type: string (or - Expression with resultType string). - :type format: object - """ - - _validation = { - 'type': {'required': True}, - } - - _attribute_map = { - 'type': {'key': 'type', 'type': 'str'}, - 'date_property': {'key': 'date', 'type': 'object'}, - 'format': {'key': 'format', 'type': 'object'}, - } - - def __init__(self, date_property=None, format=None): - super(DatasetDateTimePartitionValue, self).__init__() - self.date_property = date_property - self.format = format - self.type = 'DateTime' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_partition.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_partition.py deleted file mode 100644 index 70e920226324..000000000000 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_partition.py +++ /dev/null @@ -1,32 +0,0 @@ -# coding=utf-8 -# -------------------------------------------------------------------------- -# Copyright (c) Microsoft Corporation. All rights reserved. -# Licensed under the MIT License. See License.txt in the project root for -# license information. -# -# Code generated by Microsoft (R) AutoRest Code Generator. -# Changes may cause incorrect behavior and will be lost if the code is -# regenerated. -# -------------------------------------------------------------------------- - -from msrest.serialization import Model - - -class DatasetPartition(Model): - """The partition definition. - - :param name: Name of the partition. Type: string (or Expression with - resultType string). - :type name: object - :param value: Value of the partition. - :type value: ~azure.mgmt.datafactory.models.DatasetPartitionValue - """ - - _attribute_map = { - 'name': {'key': 'name', 'type': 'object'}, - 'value': {'key': 'value', 'type': 'DatasetPartitionValue'}, - } - - def __init__(self, name=None, value=None): - self.name = name - self.value = value diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_partition_value.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_zip_deflate_compression.py similarity index 55% rename from azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_partition_value.py rename to azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_zip_deflate_compression.py index 8e2192cbf9fd..0cfca966d475 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_partition_value.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dataset_zip_deflate_compression.py @@ -9,17 +9,17 @@ # regenerated. # -------------------------------------------------------------------------- -from msrest.serialization import Model +from .dataset_compression import DatasetCompression -class DatasetPartitionValue(Model): - """The value of a partition. - - You probably want to use the sub-classes and not this class directly. Known - sub-classes are: DatasetDateTimePartitionValue +class DatasetZipDeflateCompression(DatasetCompression): + """The ZipDeflate compression method used on a dataset. :param type: Constant filled by server. :type type: str + :param level: The ZipDeflate compression level. Possible values include: + 'Optimal', 'Fastest' + :type level: str or ~azure.mgmt.datafactory.models.DatasetCompressionLevel """ _validation = { @@ -28,11 +28,10 @@ class DatasetPartitionValue(Model): _attribute_map = { 'type': {'key': 'type', 'type': 'str'}, + 'level': {'key': 'level', 'type': 'str'}, } - _subtype_map = { - 'type': {'DateTime': 'DatasetDateTimePartitionValue'} - } - - def __init__(self): - self.type = None + def __init__(self, level=None): + super(DatasetZipDeflateCompression, self).__init__() + self.level = level + self.type = 'ZipDeflate' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/drill_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/drill_linked_service.py new file mode 100644 index 000000000000..f73bddd19b1a --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/drill_linked_service.py @@ -0,0 +1,49 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .linked_service import LinkedService + + +class DrillLinkedService(LinkedService): + """Drill server linked service. + + :param connect_via: The integration runtime reference. + :type connect_via: + ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :param description: Linked service description. + :type description: str + :param type: Constant filled by server. + :type type: str + :param connection_string: An ODBC connection string. + :type connection_string: ~azure.mgmt.datafactory.models.SecretBase + :param encrypted_credential: The encrypted credential used for + authentication. Credentials are encrypted using the integration runtime + credential manager. Type: string (or Expression with resultType string). + :type encrypted_credential: object + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, + 'description': {'key': 'description', 'type': 'str'}, + 'type': {'key': 'type', 'type': 'str'}, + 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'SecretBase'}, + 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, + } + + def __init__(self, connect_via=None, description=None, connection_string=None, encrypted_credential=None): + super(DrillLinkedService, self).__init__(connect_via=connect_via, description=description) + self.connection_string = connection_string + self.encrypted_credential = encrypted_credential + self.type = 'Drill' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/drill_source.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/drill_source.py new file mode 100644 index 000000000000..8eebc327c7e3 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/drill_source.py @@ -0,0 +1,46 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .copy_source import CopySource + + +class DrillSource(CopySource): + """A copy activity Drill server source. + + :param source_retry_count: Source retry count. Type: integer (or + Expression with resultType integer). + :type source_retry_count: object + :param source_retry_wait: Source retry wait. Type: string (or Expression + with resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type source_retry_wait: object + :param type: Constant filled by server. + :type type: str + :param query: A query to retrieve data from source. Type: string (or + Expression with resultType string). + :type query: object + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, + 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, + 'query': {'key': 'query', 'type': 'object'}, + } + + def __init__(self, source_retry_count=None, source_retry_wait=None, query=None): + super(DrillSource, self).__init__(source_retry_count=source_retry_count, source_retry_wait=source_retry_wait) + self.query = query + self.type = 'DrillSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/drill_table_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/drill_table_dataset.py new file mode 100644 index 000000000000..edefaff586c2 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/drill_table_dataset.py @@ -0,0 +1,40 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .dataset import Dataset + + +class DrillTableDataset(Dataset): + """Drill server dataset. + + :param description: Dataset description. + :type description: str + :param structure: Columns that define the structure of the dataset. Type: + array (or Expression with resultType array), itemType: DatasetDataElement. + :type structure: object + :param linked_service_name: Linked service reference. + :type linked_service_name: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param parameters: Parameters for dataset. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param type: Constant filled by server. + :type type: str + """ + + _validation = { + 'linked_service_name': {'required': True}, + 'type': {'required': True}, + } + + def __init__(self, linked_service_name, description=None, structure=None, parameters=None): + super(DrillTableDataset, self).__init__(description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters) + self.type = 'DrillTable' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dynamics_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dynamics_linked_service.py index 3c6c858c497f..36c83d784d5c 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dynamics_linked_service.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/dynamics_linked_service.py @@ -47,8 +47,11 @@ class DynamicsLinkedService(LinkedService): (or Expression with resultType string). :type username: object :param password: Password to access the Dynamics instance. - :type password: - ~azure.mgmt.datafactory.models.AzureKeyVaultSecretReference + :type password: ~azure.mgmt.datafactory.models.SecretBase + :param encrypted_credential: The encrypted credential used for + authentication. Credentials are encrypted using the integration runtime + credential manager. Type: string (or Expression with resultType string). + :type encrypted_credential: object """ _validation = { @@ -56,7 +59,6 @@ class DynamicsLinkedService(LinkedService): 'deployment_type': {'required': True}, 'authentication_type': {'required': True}, 'username': {'required': True}, - 'password': {'required': True}, } _attribute_map = { @@ -69,10 +71,11 @@ class DynamicsLinkedService(LinkedService): 'organization_name': {'key': 'typeProperties.organizationName', 'type': 'object'}, 'authentication_type': {'key': 'typeProperties.authenticationType', 'type': 'object'}, 'username': {'key': 'typeProperties.username', 'type': 'object'}, - 'password': {'key': 'typeProperties.password', 'type': 'AzureKeyVaultSecretReference'}, + 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, + 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, } - def __init__(self, deployment_type, authentication_type, username, password, connect_via=None, description=None, host_name=None, port=None, organization_name=None): + def __init__(self, deployment_type, authentication_type, username, connect_via=None, description=None, host_name=None, port=None, organization_name=None, password=None, encrypted_credential=None): super(DynamicsLinkedService, self).__init__(connect_via=connect_via, description=description) self.deployment_type = deployment_type self.host_name = host_name @@ -81,4 +84,5 @@ def __init__(self, deployment_type, authentication_type, username, password, con self.authentication_type = authentication_type self.username = username self.password = password + self.encrypted_credential = encrypted_credential self.type = 'Dynamics' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/eloqua_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/eloqua_linked_service.py new file mode 100644 index 000000000000..43bca05743c3 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/eloqua_linked_service.py @@ -0,0 +1,77 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .linked_service import LinkedService + + +class EloquaLinkedService(LinkedService): + """Eloqua server linked service. + + :param connect_via: The integration runtime reference. + :type connect_via: + ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :param description: Linked service description. + :type description: str + :param type: Constant filled by server. + :type type: str + :param endpoint: The endpoint of the Eloqua server. (i.e. + eloqua.example.com) + :type endpoint: object + :param username: The site name and user name of your Eloqua account in the + form: sitename/username. (i.e. Eloqua/Alice) + :type username: object + :param password: The password corresponding to the user name. + :type password: ~azure.mgmt.datafactory.models.SecretBase + :param use_encrypted_endpoints: Specifies whether the data source + endpoints are encrypted using HTTPS. The default value is true. + :type use_encrypted_endpoints: object + :param use_host_verification: Specifies whether to require the host name + in the server's certificate to match the host name of the server when + connecting over SSL. The default value is true. + :type use_host_verification: object + :param use_peer_verification: Specifies whether to verify the identity of + the server when connecting over SSL. The default value is true. + :type use_peer_verification: object + :param encrypted_credential: The encrypted credential used for + authentication. Credentials are encrypted using the integration runtime + credential manager. Type: string (or Expression with resultType string). + :type encrypted_credential: object + """ + + _validation = { + 'type': {'required': True}, + 'endpoint': {'required': True}, + 'username': {'required': True}, + } + + _attribute_map = { + 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, + 'description': {'key': 'description', 'type': 'str'}, + 'type': {'key': 'type', 'type': 'str'}, + 'endpoint': {'key': 'typeProperties.endpoint', 'type': 'object'}, + 'username': {'key': 'typeProperties.username', 'type': 'object'}, + 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, + 'use_encrypted_endpoints': {'key': 'typeProperties.useEncryptedEndpoints', 'type': 'object'}, + 'use_host_verification': {'key': 'typeProperties.useHostVerification', 'type': 'object'}, + 'use_peer_verification': {'key': 'typeProperties.usePeerVerification', 'type': 'object'}, + 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, + } + + def __init__(self, endpoint, username, connect_via=None, description=None, password=None, use_encrypted_endpoints=None, use_host_verification=None, use_peer_verification=None, encrypted_credential=None): + super(EloquaLinkedService, self).__init__(connect_via=connect_via, description=description) + self.endpoint = endpoint + self.username = username + self.password = password + self.use_encrypted_endpoints = use_encrypted_endpoints + self.use_host_verification = use_host_verification + self.use_peer_verification = use_peer_verification + self.encrypted_credential = encrypted_credential + self.type = 'Eloqua' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/eloqua_object_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/eloqua_object_dataset.py new file mode 100644 index 000000000000..c4fc6ddb0303 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/eloqua_object_dataset.py @@ -0,0 +1,40 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .dataset import Dataset + + +class EloquaObjectDataset(Dataset): + """Eloqua server dataset. + + :param description: Dataset description. + :type description: str + :param structure: Columns that define the structure of the dataset. Type: + array (or Expression with resultType array), itemType: DatasetDataElement. + :type structure: object + :param linked_service_name: Linked service reference. + :type linked_service_name: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param parameters: Parameters for dataset. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param type: Constant filled by server. + :type type: str + """ + + _validation = { + 'linked_service_name': {'required': True}, + 'type': {'required': True}, + } + + def __init__(self, linked_service_name, description=None, structure=None, parameters=None): + super(EloquaObjectDataset, self).__init__(description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters) + self.type = 'EloquaObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/eloqua_source.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/eloqua_source.py new file mode 100644 index 000000000000..a2c2982f7043 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/eloqua_source.py @@ -0,0 +1,46 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .copy_source import CopySource + + +class EloquaSource(CopySource): + """A copy activity Eloqua server source. + + :param source_retry_count: Source retry count. Type: integer (or + Expression with resultType integer). + :type source_retry_count: object + :param source_retry_wait: Source retry wait. Type: string (or Expression + with resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type source_retry_wait: object + :param type: Constant filled by server. + :type type: str + :param query: A query to retrieve data from source. Type: string (or + Expression with resultType string). + :type query: object + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, + 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, + 'query': {'key': 'query', 'type': 'object'}, + } + + def __init__(self, source_retry_count=None, source_retry_wait=None, query=None): + super(EloquaSource, self).__init__(source_retry_count=source_retry_count, source_retry_wait=source_retry_wait) + self.query = query + self.type = 'EloquaSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/google_big_query_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/google_big_query_linked_service.py new file mode 100644 index 000000000000..168d37441d04 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/google_big_query_linked_service.py @@ -0,0 +1,99 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .linked_service import LinkedService + + +class GoogleBigQueryLinkedService(LinkedService): + """Google BigQuery service linked service. + + :param connect_via: The integration runtime reference. + :type connect_via: + ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :param description: Linked service description. + :type description: str + :param type: Constant filled by server. + :type type: str + :param project: The default BigQuery project to query against. + :type project: object + :param additional_projects: A comma-separated list of public BigQuery + projects to access. + :type additional_projects: object + :param request_google_drive_scope: Whether to request access to Google + Drive. Allowing Google Drive access enables support for federated tables + that combine BigQuery data with data from Google Drive. The default value + is false. + :type request_google_drive_scope: object + :param authentication_type: The OAuth 2.0 authentication mechanism used + for authentication. ServiceAuthentication can only be used on self-hosted + IR. Possible values include: 'ServiceAuthentication', 'UserAuthentication' + :type authentication_type: str or + ~azure.mgmt.datafactory.models.GoogleBigQueryAuthenticationType + :param refresh_token: The refresh token obtained from Google for + authorizing access to BigQuery for UserAuthentication. + :type refresh_token: ~azure.mgmt.datafactory.models.SecretBase + :param email: The service account email ID that is used for + ServiceAuthentication and can only be used on self-hosted IR. + :type email: object + :param key_file_path: The full path to the .p12 key file that is used to + authenticate the service account email address and can only be used on + self-hosted IR. + :type key_file_path: object + :param trusted_cert_path: The full path of the .pem file containing + trusted CA certificates for verifying the server when connecting over SSL. + This property can only be set when using SSL on self-hosted IR. The + default value is the cacerts.pem file installed with the IR. + :type trusted_cert_path: object + :param use_system_trust_store: Specifies whether to use a CA certificate + from the system trust store or from a specified PEM file. The default + value is false. + :type use_system_trust_store: object + :param encrypted_credential: The encrypted credential used for + authentication. Credentials are encrypted using the integration runtime + credential manager. Type: string (or Expression with resultType string). + :type encrypted_credential: object + """ + + _validation = { + 'type': {'required': True}, + 'project': {'required': True}, + 'authentication_type': {'required': True}, + } + + _attribute_map = { + 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, + 'description': {'key': 'description', 'type': 'str'}, + 'type': {'key': 'type', 'type': 'str'}, + 'project': {'key': 'typeProperties.project', 'type': 'object'}, + 'additional_projects': {'key': 'typeProperties.additionalProjects', 'type': 'object'}, + 'request_google_drive_scope': {'key': 'typeProperties.requestGoogleDriveScope', 'type': 'object'}, + 'authentication_type': {'key': 'typeProperties.authenticationType', 'type': 'str'}, + 'refresh_token': {'key': 'typeProperties.refreshToken', 'type': 'SecretBase'}, + 'email': {'key': 'typeProperties.email', 'type': 'object'}, + 'key_file_path': {'key': 'typeProperties.keyFilePath', 'type': 'object'}, + 'trusted_cert_path': {'key': 'typeProperties.trustedCertPath', 'type': 'object'}, + 'use_system_trust_store': {'key': 'typeProperties.useSystemTrustStore', 'type': 'object'}, + 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, + } + + def __init__(self, project, authentication_type, connect_via=None, description=None, additional_projects=None, request_google_drive_scope=None, refresh_token=None, email=None, key_file_path=None, trusted_cert_path=None, use_system_trust_store=None, encrypted_credential=None): + super(GoogleBigQueryLinkedService, self).__init__(connect_via=connect_via, description=description) + self.project = project + self.additional_projects = additional_projects + self.request_google_drive_scope = request_google_drive_scope + self.authentication_type = authentication_type + self.refresh_token = refresh_token + self.email = email + self.key_file_path = key_file_path + self.trusted_cert_path = trusted_cert_path + self.use_system_trust_store = use_system_trust_store + self.encrypted_credential = encrypted_credential + self.type = 'GoogleBigQuery' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/google_big_query_object_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/google_big_query_object_dataset.py new file mode 100644 index 000000000000..b8ab5c0ee854 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/google_big_query_object_dataset.py @@ -0,0 +1,40 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .dataset import Dataset + + +class GoogleBigQueryObjectDataset(Dataset): + """Google BigQuery service dataset. + + :param description: Dataset description. + :type description: str + :param structure: Columns that define the structure of the dataset. Type: + array (or Expression with resultType array), itemType: DatasetDataElement. + :type structure: object + :param linked_service_name: Linked service reference. + :type linked_service_name: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param parameters: Parameters for dataset. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param type: Constant filled by server. + :type type: str + """ + + _validation = { + 'linked_service_name': {'required': True}, + 'type': {'required': True}, + } + + def __init__(self, linked_service_name, description=None, structure=None, parameters=None): + super(GoogleBigQueryObjectDataset, self).__init__(description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters) + self.type = 'GoogleBigQueryObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/google_big_query_source.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/google_big_query_source.py new file mode 100644 index 000000000000..fcf6f184ca3d --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/google_big_query_source.py @@ -0,0 +1,46 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .copy_source import CopySource + + +class GoogleBigQuerySource(CopySource): + """A copy activity Google BigQuery service source. + + :param source_retry_count: Source retry count. Type: integer (or + Expression with resultType integer). + :type source_retry_count: object + :param source_retry_wait: Source retry wait. Type: string (or Expression + with resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type source_retry_wait: object + :param type: Constant filled by server. + :type type: str + :param query: A query to retrieve data from source. Type: string (or + Expression with resultType string). + :type query: object + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, + 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, + 'query': {'key': 'query', 'type': 'object'}, + } + + def __init__(self, source_retry_count=None, source_retry_wait=None, query=None): + super(GoogleBigQuerySource, self).__init__(source_retry_count=source_retry_count, source_retry_wait=source_retry_wait) + self.query = query + self.type = 'GoogleBigQuerySource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/greenplum_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/greenplum_linked_service.py new file mode 100644 index 000000000000..b34559cb2aff --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/greenplum_linked_service.py @@ -0,0 +1,49 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .linked_service import LinkedService + + +class GreenplumLinkedService(LinkedService): + """Greenplum Database linked service. + + :param connect_via: The integration runtime reference. + :type connect_via: + ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :param description: Linked service description. + :type description: str + :param type: Constant filled by server. + :type type: str + :param connection_string: An ODBC connection string. + :type connection_string: ~azure.mgmt.datafactory.models.SecretBase + :param encrypted_credential: The encrypted credential used for + authentication. Credentials are encrypted using the integration runtime + credential manager. Type: string (or Expression with resultType string). + :type encrypted_credential: object + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, + 'description': {'key': 'description', 'type': 'str'}, + 'type': {'key': 'type', 'type': 'str'}, + 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'SecretBase'}, + 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, + } + + def __init__(self, connect_via=None, description=None, connection_string=None, encrypted_credential=None): + super(GreenplumLinkedService, self).__init__(connect_via=connect_via, description=description) + self.connection_string = connection_string + self.encrypted_credential = encrypted_credential + self.type = 'Greenplum' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/greenplum_source.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/greenplum_source.py new file mode 100644 index 000000000000..151241eb4b9c --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/greenplum_source.py @@ -0,0 +1,46 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .copy_source import CopySource + + +class GreenplumSource(CopySource): + """A copy activity Greenplum Database source. + + :param source_retry_count: Source retry count. Type: integer (or + Expression with resultType integer). + :type source_retry_count: object + :param source_retry_wait: Source retry wait. Type: string (or Expression + with resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type source_retry_wait: object + :param type: Constant filled by server. + :type type: str + :param query: A query to retrieve data from source. Type: string (or + Expression with resultType string). + :type query: object + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, + 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, + 'query': {'key': 'query', 'type': 'object'}, + } + + def __init__(self, source_retry_count=None, source_retry_wait=None, query=None): + super(GreenplumSource, self).__init__(source_retry_count=source_retry_count, source_retry_wait=source_retry_wait) + self.query = query + self.type = 'GreenplumSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/greenplum_table_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/greenplum_table_dataset.py new file mode 100644 index 000000000000..21f7736172f3 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/greenplum_table_dataset.py @@ -0,0 +1,40 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .dataset import Dataset + + +class GreenplumTableDataset(Dataset): + """Greenplum Database dataset. + + :param description: Dataset description. + :type description: str + :param structure: Columns that define the structure of the dataset. Type: + array (or Expression with resultType array), itemType: DatasetDataElement. + :type structure: object + :param linked_service_name: Linked service reference. + :type linked_service_name: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param parameters: Parameters for dataset. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param type: Constant filled by server. + :type type: str + """ + + _validation = { + 'linked_service_name': {'required': True}, + 'type': {'required': True}, + } + + def __init__(self, linked_service_name, description=None, structure=None, parameters=None): + super(GreenplumTableDataset, self).__init__(description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters) + self.type = 'GreenplumTable' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hbase_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hbase_linked_service.py new file mode 100644 index 000000000000..0405ff64b33e --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hbase_linked_service.py @@ -0,0 +1,99 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .linked_service import LinkedService + + +class HBaseLinkedService(LinkedService): + """HBase server linked service. + + :param connect_via: The integration runtime reference. + :type connect_via: + ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :param description: Linked service description. + :type description: str + :param type: Constant filled by server. + :type type: str + :param host: The IP address or host name of the HBase server. (i.e. + 192.168.222.160) + :type host: object + :param port: The TCP port that the HBase instance uses to listen for + client connections. The default value is 9090. + :type port: object + :param http_path: The partial URL corresponding to the HBase server. (i.e. + /gateway/sandbox/hbase/version) + :type http_path: object + :param authentication_type: The authentication mechanism to use to connect + to the HBase server. Possible values include: 'Anonymous', 'Basic' + :type authentication_type: str or + ~azure.mgmt.datafactory.models.HBaseAuthenticationType + :param username: The user name used to connect to the HBase instance. + :type username: object + :param password: The password corresponding to the user name. + :type password: ~azure.mgmt.datafactory.models.SecretBase + :param enable_ssl: Specifies whether the connections to the server are + encrypted using SSL. The default value is false. + :type enable_ssl: object + :param trusted_cert_path: The full path of the .pem file containing + trusted CA certificates for verifying the server when connecting over SSL. + This property can only be set when using SSL on self-hosted IR. The + default value is the cacerts.pem file installed with the IR. + :type trusted_cert_path: object + :param allow_host_name_cn_mismatch: Specifies whether to require a + CA-issued SSL certificate name to match the host name of the server when + connecting over SSL. The default value is false. + :type allow_host_name_cn_mismatch: object + :param allow_self_signed_server_cert: Specifies whether to allow + self-signed certificates from the server. The default value is false. + :type allow_self_signed_server_cert: object + :param encrypted_credential: The encrypted credential used for + authentication. Credentials are encrypted using the integration runtime + credential manager. Type: string (or Expression with resultType string). + :type encrypted_credential: object + """ + + _validation = { + 'type': {'required': True}, + 'host': {'required': True}, + 'authentication_type': {'required': True}, + } + + _attribute_map = { + 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, + 'description': {'key': 'description', 'type': 'str'}, + 'type': {'key': 'type', 'type': 'str'}, + 'host': {'key': 'typeProperties.host', 'type': 'object'}, + 'port': {'key': 'typeProperties.port', 'type': 'object'}, + 'http_path': {'key': 'typeProperties.httpPath', 'type': 'object'}, + 'authentication_type': {'key': 'typeProperties.authenticationType', 'type': 'str'}, + 'username': {'key': 'typeProperties.username', 'type': 'object'}, + 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, + 'enable_ssl': {'key': 'typeProperties.enableSsl', 'type': 'object'}, + 'trusted_cert_path': {'key': 'typeProperties.trustedCertPath', 'type': 'object'}, + 'allow_host_name_cn_mismatch': {'key': 'typeProperties.allowHostNameCNMismatch', 'type': 'object'}, + 'allow_self_signed_server_cert': {'key': 'typeProperties.allowSelfSignedServerCert', 'type': 'object'}, + 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, + } + + def __init__(self, host, authentication_type, connect_via=None, description=None, port=None, http_path=None, username=None, password=None, enable_ssl=None, trusted_cert_path=None, allow_host_name_cn_mismatch=None, allow_self_signed_server_cert=None, encrypted_credential=None): + super(HBaseLinkedService, self).__init__(connect_via=connect_via, description=description) + self.host = host + self.port = port + self.http_path = http_path + self.authentication_type = authentication_type + self.username = username + self.password = password + self.enable_ssl = enable_ssl + self.trusted_cert_path = trusted_cert_path + self.allow_host_name_cn_mismatch = allow_host_name_cn_mismatch + self.allow_self_signed_server_cert = allow_self_signed_server_cert + self.encrypted_credential = encrypted_credential + self.type = 'HBase' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hbase_object_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hbase_object_dataset.py new file mode 100644 index 000000000000..035d29a2ed7c --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hbase_object_dataset.py @@ -0,0 +1,40 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .dataset import Dataset + + +class HBaseObjectDataset(Dataset): + """HBase server dataset. + + :param description: Dataset description. + :type description: str + :param structure: Columns that define the structure of the dataset. Type: + array (or Expression with resultType array), itemType: DatasetDataElement. + :type structure: object + :param linked_service_name: Linked service reference. + :type linked_service_name: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param parameters: Parameters for dataset. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param type: Constant filled by server. + :type type: str + """ + + _validation = { + 'linked_service_name': {'required': True}, + 'type': {'required': True}, + } + + def __init__(self, linked_service_name, description=None, structure=None, parameters=None): + super(HBaseObjectDataset, self).__init__(description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters) + self.type = 'HBaseObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hbase_source.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hbase_source.py new file mode 100644 index 000000000000..03ab97d082be --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hbase_source.py @@ -0,0 +1,46 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .copy_source import CopySource + + +class HBaseSource(CopySource): + """A copy activity HBase server source. + + :param source_retry_count: Source retry count. Type: integer (or + Expression with resultType integer). + :type source_retry_count: object + :param source_retry_wait: Source retry wait. Type: string (or Expression + with resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type source_retry_wait: object + :param type: Constant filled by server. + :type type: str + :param query: A query to retrieve data from source. Type: string (or + Expression with resultType string). + :type query: object + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, + 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, + 'query': {'key': 'query', 'type': 'object'}, + } + + def __init__(self, source_retry_count=None, source_retry_wait=None, query=None): + super(HBaseSource, self).__init__(source_retry_count=source_retry_count, source_retry_wait=source_retry_wait) + self.query = query + self.type = 'HBaseSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hdfs_source.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hdfs_source.py index cfb6c0dd7a20..3ec6154496a8 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hdfs_source.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hdfs_source.py @@ -13,7 +13,7 @@ class HdfsSource(CopySource): - """A copy activity source for HDFS source. + """A copy activity HDFS source. :param source_retry_count: Source retry count. Type: integer (or Expression with resultType integer). diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hive_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hive_linked_service.py new file mode 100644 index 000000000000..5c5d00fc73dc --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hive_linked_service.py @@ -0,0 +1,132 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .linked_service import LinkedService + + +class HiveLinkedService(LinkedService): + """Hive Server linked service. + + :param connect_via: The integration runtime reference. + :type connect_via: + ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :param description: Linked service description. + :type description: str + :param type: Constant filled by server. + :type type: str + :param host: IP address or host name of the Hive server, separated by ';' + for multiple hosts (only when serviceDiscoveryMode is enable). + :type host: object + :param port: The TCP port that the Hive server uses to listen for client + connections. + :type port: object + :param server_type: The type of Hive server. Possible values include: + 'HiveServer1', 'HiveServer2', 'HiveThriftServer' + :type server_type: str or ~azure.mgmt.datafactory.models.HiveServerType + :param thrift_transport_protocol: The transport protocol to use in the + Thrift layer. Possible values include: 'Binary', 'SASL', 'HTTP ' + :type thrift_transport_protocol: str or + ~azure.mgmt.datafactory.models.HiveThriftTransportProtocol + :param authentication_type: The authentication method used to access the + Hive server. Possible values include: 'Anonymous', 'Username', + 'UsernameAndPassword', 'WindowsAzureHDInsightService' + :type authentication_type: str or + ~azure.mgmt.datafactory.models.HiveAuthenticationType + :param service_discovery_mode: true to indicate using the ZooKeeper + service, false not. + :type service_discovery_mode: object + :param zoo_keeper_name_space: The namespace on ZooKeeper under which Hive + Server 2 nodes are added. + :type zoo_keeper_name_space: object + :param use_native_query: Specifies whether the driver uses native HiveQL + queries,or converts them into an equivalent form in HiveQL. + :type use_native_query: object + :param username: The user name that you use to access Hive Server. + :type username: object + :param password: The password corresponding to the user name that you + provided in the Username field + :type password: ~azure.mgmt.datafactory.models.SecretBase + :param http_path: The partial URL corresponding to the Hive server. + :type http_path: object + :param enable_ssl: Specifies whether the connections to the server are + encrypted using SSL. The default value is false. + :type enable_ssl: object + :param trusted_cert_path: The full path of the .pem file containing + trusted CA certificates for verifying the server when connecting over SSL. + This property can only be set when using SSL on self-hosted IR. The + default value is the cacerts.pem file installed with the IR. + :type trusted_cert_path: object + :param use_system_trust_store: Specifies whether to use a CA certificate + from the system trust store or from a specified PEM file. The default + value is false. + :type use_system_trust_store: object + :param allow_host_name_cn_mismatch: Specifies whether to require a + CA-issued SSL certificate name to match the host name of the server when + connecting over SSL. The default value is false. + :type allow_host_name_cn_mismatch: object + :param allow_self_signed_server_cert: Specifies whether to allow + self-signed certificates from the server. The default value is false. + :type allow_self_signed_server_cert: object + :param encrypted_credential: The encrypted credential used for + authentication. Credentials are encrypted using the integration runtime + credential manager. Type: string (or Expression with resultType string). + :type encrypted_credential: object + """ + + _validation = { + 'type': {'required': True}, + 'host': {'required': True}, + 'authentication_type': {'required': True}, + } + + _attribute_map = { + 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, + 'description': {'key': 'description', 'type': 'str'}, + 'type': {'key': 'type', 'type': 'str'}, + 'host': {'key': 'typeProperties.host', 'type': 'object'}, + 'port': {'key': 'typeProperties.port', 'type': 'object'}, + 'server_type': {'key': 'typeProperties.serverType', 'type': 'str'}, + 'thrift_transport_protocol': {'key': 'typeProperties.thriftTransportProtocol', 'type': 'str'}, + 'authentication_type': {'key': 'typeProperties.authenticationType', 'type': 'str'}, + 'service_discovery_mode': {'key': 'typeProperties.serviceDiscoveryMode', 'type': 'object'}, + 'zoo_keeper_name_space': {'key': 'typeProperties.zooKeeperNameSpace', 'type': 'object'}, + 'use_native_query': {'key': 'typeProperties.useNativeQuery', 'type': 'object'}, + 'username': {'key': 'typeProperties.username', 'type': 'object'}, + 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, + 'http_path': {'key': 'typeProperties.httpPath', 'type': 'object'}, + 'enable_ssl': {'key': 'typeProperties.enableSsl', 'type': 'object'}, + 'trusted_cert_path': {'key': 'typeProperties.trustedCertPath', 'type': 'object'}, + 'use_system_trust_store': {'key': 'typeProperties.useSystemTrustStore', 'type': 'object'}, + 'allow_host_name_cn_mismatch': {'key': 'typeProperties.allowHostNameCNMismatch', 'type': 'object'}, + 'allow_self_signed_server_cert': {'key': 'typeProperties.allowSelfSignedServerCert', 'type': 'object'}, + 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, + } + + def __init__(self, host, authentication_type, connect_via=None, description=None, port=None, server_type=None, thrift_transport_protocol=None, service_discovery_mode=None, zoo_keeper_name_space=None, use_native_query=None, username=None, password=None, http_path=None, enable_ssl=None, trusted_cert_path=None, use_system_trust_store=None, allow_host_name_cn_mismatch=None, allow_self_signed_server_cert=None, encrypted_credential=None): + super(HiveLinkedService, self).__init__(connect_via=connect_via, description=description) + self.host = host + self.port = port + self.server_type = server_type + self.thrift_transport_protocol = thrift_transport_protocol + self.authentication_type = authentication_type + self.service_discovery_mode = service_discovery_mode + self.zoo_keeper_name_space = zoo_keeper_name_space + self.use_native_query = use_native_query + self.username = username + self.password = password + self.http_path = http_path + self.enable_ssl = enable_ssl + self.trusted_cert_path = trusted_cert_path + self.use_system_trust_store = use_system_trust_store + self.allow_host_name_cn_mismatch = allow_host_name_cn_mismatch + self.allow_self_signed_server_cert = allow_self_signed_server_cert + self.encrypted_credential = encrypted_credential + self.type = 'Hive' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hive_object_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hive_object_dataset.py new file mode 100644 index 000000000000..55c34ab48472 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hive_object_dataset.py @@ -0,0 +1,40 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .dataset import Dataset + + +class HiveObjectDataset(Dataset): + """Hive Server dataset. + + :param description: Dataset description. + :type description: str + :param structure: Columns that define the structure of the dataset. Type: + array (or Expression with resultType array), itemType: DatasetDataElement. + :type structure: object + :param linked_service_name: Linked service reference. + :type linked_service_name: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param parameters: Parameters for dataset. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param type: Constant filled by server. + :type type: str + """ + + _validation = { + 'linked_service_name': {'required': True}, + 'type': {'required': True}, + } + + def __init__(self, linked_service_name, description=None, structure=None, parameters=None): + super(HiveObjectDataset, self).__init__(description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters) + self.type = 'HiveObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hive_source.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hive_source.py new file mode 100644 index 000000000000..1113a91befd2 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hive_source.py @@ -0,0 +1,46 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .copy_source import CopySource + + +class HiveSource(CopySource): + """A copy activity Hive Server source. + + :param source_retry_count: Source retry count. Type: integer (or + Expression with resultType integer). + :type source_retry_count: object + :param source_retry_wait: Source retry wait. Type: string (or Expression + with resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type source_retry_wait: object + :param type: Constant filled by server. + :type type: str + :param query: A query to retrieve data from source. Type: string (or + Expression with resultType string). + :type query: object + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, + 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, + 'query': {'key': 'query', 'type': 'object'}, + } + + def __init__(self, source_retry_count=None, source_retry_wait=None, query=None): + super(HiveSource, self).__init__(source_retry_count=source_retry_count, source_retry_wait=source_retry_wait) + self.query = query + self.type = 'HiveSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/http_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/http_dataset.py index 7baa3696ba2c..8f3705a191d2 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/http_dataset.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/http_dataset.py @@ -44,8 +44,6 @@ class HttpDataset(Dataset): request-header-name-n:request-header-value-n Type: string (or Expression with resultType string). :type additional_headers: object - :param partitioned_by: The HTTP method for the HTTP request. - :type partitioned_by: ~azure.mgmt.datafactory.models.DatasetPartition :param format: The format of files. :type format: ~azure.mgmt.datafactory.models.DatasetStorageFormat :param compression: The data compression method used on files. @@ -67,18 +65,16 @@ class HttpDataset(Dataset): 'request_method': {'key': 'typeProperties.requestMethod', 'type': 'object'}, 'request_body': {'key': 'typeProperties.requestBody', 'type': 'object'}, 'additional_headers': {'key': 'typeProperties.additionalHeaders', 'type': 'object'}, - 'partitioned_by': {'key': 'typeProperties.partitionedBy', 'type': 'DatasetPartition'}, 'format': {'key': 'typeProperties.format', 'type': 'DatasetStorageFormat'}, 'compression': {'key': 'typeProperties.compression', 'type': 'DatasetCompression'}, } - def __init__(self, linked_service_name, description=None, structure=None, parameters=None, relative_url=None, request_method=None, request_body=None, additional_headers=None, partitioned_by=None, format=None, compression=None): + def __init__(self, linked_service_name, description=None, structure=None, parameters=None, relative_url=None, request_method=None, request_body=None, additional_headers=None, format=None, compression=None): super(HttpDataset, self).__init__(description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters) self.relative_url = relative_url self.request_method = request_method self.request_body = request_body self.additional_headers = additional_headers - self.partitioned_by = partitioned_by self.format = format self.compression = compression self.type = 'HttpFile' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hubspot_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hubspot_linked_service.py new file mode 100644 index 000000000000..4ea144a70066 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hubspot_linked_service.py @@ -0,0 +1,81 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .linked_service import LinkedService + + +class HubspotLinkedService(LinkedService): + """Hubspot Serivce linked service. + + :param connect_via: The integration runtime reference. + :type connect_via: + ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :param description: Linked service description. + :type description: str + :param type: Constant filled by server. + :type type: str + :param client_id: The client ID associated with your Hubspot application. + :type client_id: object + :param client_secret: The client secret associated with your Hubspot + application. + :type client_secret: ~azure.mgmt.datafactory.models.SecretBase + :param access_token: The access token obtained when initially + authenticating your OAuth integration. + :type access_token: ~azure.mgmt.datafactory.models.SecretBase + :param refresh_token: The refresh token obtained when initially + authenticating your OAuth integration. + :type refresh_token: ~azure.mgmt.datafactory.models.SecretBase + :param use_encrypted_endpoints: Specifies whether the data source + endpoints are encrypted using HTTPS. The default value is true. + :type use_encrypted_endpoints: object + :param use_host_verification: Specifies whether to require the host name + in the server's certificate to match the host name of the server when + connecting over SSL. The default value is true. + :type use_host_verification: object + :param use_peer_verification: Specifies whether to verify the identity of + the server when connecting over SSL. The default value is true. + :type use_peer_verification: object + :param encrypted_credential: The encrypted credential used for + authentication. Credentials are encrypted using the integration runtime + credential manager. Type: string (or Expression with resultType string). + :type encrypted_credential: object + """ + + _validation = { + 'type': {'required': True}, + 'client_id': {'required': True}, + } + + _attribute_map = { + 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, + 'description': {'key': 'description', 'type': 'str'}, + 'type': {'key': 'type', 'type': 'str'}, + 'client_id': {'key': 'typeProperties.clientId', 'type': 'object'}, + 'client_secret': {'key': 'typeProperties.clientSecret', 'type': 'SecretBase'}, + 'access_token': {'key': 'typeProperties.accessToken', 'type': 'SecretBase'}, + 'refresh_token': {'key': 'typeProperties.refreshToken', 'type': 'SecretBase'}, + 'use_encrypted_endpoints': {'key': 'typeProperties.useEncryptedEndpoints', 'type': 'object'}, + 'use_host_verification': {'key': 'typeProperties.useHostVerification', 'type': 'object'}, + 'use_peer_verification': {'key': 'typeProperties.usePeerVerification', 'type': 'object'}, + 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, + } + + def __init__(self, client_id, connect_via=None, description=None, client_secret=None, access_token=None, refresh_token=None, use_encrypted_endpoints=None, use_host_verification=None, use_peer_verification=None, encrypted_credential=None): + super(HubspotLinkedService, self).__init__(connect_via=connect_via, description=description) + self.client_id = client_id + self.client_secret = client_secret + self.access_token = access_token + self.refresh_token = refresh_token + self.use_encrypted_endpoints = use_encrypted_endpoints + self.use_host_verification = use_host_verification + self.use_peer_verification = use_peer_verification + self.encrypted_credential = encrypted_credential + self.type = 'Hubspot' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hubspot_object_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hubspot_object_dataset.py new file mode 100644 index 000000000000..1e83ddff4559 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hubspot_object_dataset.py @@ -0,0 +1,40 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .dataset import Dataset + + +class HubspotObjectDataset(Dataset): + """Hubspot Serivce dataset. + + :param description: Dataset description. + :type description: str + :param structure: Columns that define the structure of the dataset. Type: + array (or Expression with resultType array), itemType: DatasetDataElement. + :type structure: object + :param linked_service_name: Linked service reference. + :type linked_service_name: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param parameters: Parameters for dataset. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param type: Constant filled by server. + :type type: str + """ + + _validation = { + 'linked_service_name': {'required': True}, + 'type': {'required': True}, + } + + def __init__(self, linked_service_name, description=None, structure=None, parameters=None): + super(HubspotObjectDataset, self).__init__(description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters) + self.type = 'HubspotObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hubspot_source.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hubspot_source.py new file mode 100644 index 000000000000..254cebeae260 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/hubspot_source.py @@ -0,0 +1,46 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .copy_source import CopySource + + +class HubspotSource(CopySource): + """A copy activity Hubspot Serivce source. + + :param source_retry_count: Source retry count. Type: integer (or + Expression with resultType integer). + :type source_retry_count: object + :param source_retry_wait: Source retry wait. Type: string (or Expression + with resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type source_retry_wait: object + :param type: Constant filled by server. + :type type: str + :param query: A query to retrieve data from source. Type: string (or + Expression with resultType string). + :type query: object + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, + 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, + 'query': {'key': 'query', 'type': 'object'}, + } + + def __init__(self, source_retry_count=None, source_retry_wait=None, query=None): + super(HubspotSource, self).__init__(source_retry_count=source_retry_count, source_retry_wait=source_retry_wait) + self.query = query + self.type = 'HubspotSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/impala_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/impala_linked_service.py new file mode 100644 index 000000000000..50aa287f2656 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/impala_linked_service.py @@ -0,0 +1,102 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .linked_service import LinkedService + + +class ImpalaLinkedService(LinkedService): + """Impala server linked service. + + :param connect_via: The integration runtime reference. + :type connect_via: + ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :param description: Linked service description. + :type description: str + :param type: Constant filled by server. + :type type: str + :param host: The IP address or host name of the Impala server. (i.e. + 192.168.222.160) + :type host: object + :param port: The TCP port that the Impala server uses to listen for client + connections. The default value is 21050. + :type port: object + :param authentication_type: The authentication type to use. Possible + values include: 'Anonymous', 'SASLUsername', 'UsernameAndPassword' + :type authentication_type: str or + ~azure.mgmt.datafactory.models.ImpalaAuthenticationType + :param username: The user name used to access the Impala server. The + default value is anonymous when using SASLUsername. + :type username: object + :param password: The password corresponding to the user name when using + UsernameAndPassword. + :type password: ~azure.mgmt.datafactory.models.SecretBase + :param enable_ssl: Specifies whether the connections to the server are + encrypted using SSL. The default value is false. + :type enable_ssl: object + :param trusted_cert_path: The full path of the .pem file containing + trusted CA certificates for verifying the server when connecting over SSL. + This property can only be set when using SSL on self-hosted IR. The + default value is the cacerts.pem file installed with the IR. + :type trusted_cert_path: object + :param use_system_trust_store: Specifies whether to use a CA certificate + from the system trust store or from a specified PEM file. The default + value is false. + :type use_system_trust_store: object + :param allow_host_name_cn_mismatch: Specifies whether to require a + CA-issued SSL certificate name to match the host name of the server when + connecting over SSL. The default value is false. + :type allow_host_name_cn_mismatch: object + :param allow_self_signed_server_cert: Specifies whether to allow + self-signed certificates from the server. The default value is false. + :type allow_self_signed_server_cert: object + :param encrypted_credential: The encrypted credential used for + authentication. Credentials are encrypted using the integration runtime + credential manager. Type: string (or Expression with resultType string). + :type encrypted_credential: object + """ + + _validation = { + 'type': {'required': True}, + 'host': {'required': True}, + 'authentication_type': {'required': True}, + } + + _attribute_map = { + 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, + 'description': {'key': 'description', 'type': 'str'}, + 'type': {'key': 'type', 'type': 'str'}, + 'host': {'key': 'typeProperties.host', 'type': 'object'}, + 'port': {'key': 'typeProperties.port', 'type': 'object'}, + 'authentication_type': {'key': 'typeProperties.authenticationType', 'type': 'str'}, + 'username': {'key': 'typeProperties.username', 'type': 'object'}, + 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, + 'enable_ssl': {'key': 'typeProperties.enableSsl', 'type': 'object'}, + 'trusted_cert_path': {'key': 'typeProperties.trustedCertPath', 'type': 'object'}, + 'use_system_trust_store': {'key': 'typeProperties.useSystemTrustStore', 'type': 'object'}, + 'allow_host_name_cn_mismatch': {'key': 'typeProperties.allowHostNameCNMismatch', 'type': 'object'}, + 'allow_self_signed_server_cert': {'key': 'typeProperties.allowSelfSignedServerCert', 'type': 'object'}, + 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, + } + + def __init__(self, host, authentication_type, connect_via=None, description=None, port=None, username=None, password=None, enable_ssl=None, trusted_cert_path=None, use_system_trust_store=None, allow_host_name_cn_mismatch=None, allow_self_signed_server_cert=None, encrypted_credential=None): + super(ImpalaLinkedService, self).__init__(connect_via=connect_via, description=description) + self.host = host + self.port = port + self.authentication_type = authentication_type + self.username = username + self.password = password + self.enable_ssl = enable_ssl + self.trusted_cert_path = trusted_cert_path + self.use_system_trust_store = use_system_trust_store + self.allow_host_name_cn_mismatch = allow_host_name_cn_mismatch + self.allow_self_signed_server_cert = allow_self_signed_server_cert + self.encrypted_credential = encrypted_credential + self.type = 'Impala' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/impala_object_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/impala_object_dataset.py new file mode 100644 index 000000000000..b7968d1fbc6e --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/impala_object_dataset.py @@ -0,0 +1,40 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .dataset import Dataset + + +class ImpalaObjectDataset(Dataset): + """Impala server dataset. + + :param description: Dataset description. + :type description: str + :param structure: Columns that define the structure of the dataset. Type: + array (or Expression with resultType array), itemType: DatasetDataElement. + :type structure: object + :param linked_service_name: Linked service reference. + :type linked_service_name: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param parameters: Parameters for dataset. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param type: Constant filled by server. + :type type: str + """ + + _validation = { + 'linked_service_name': {'required': True}, + 'type': {'required': True}, + } + + def __init__(self, linked_service_name, description=None, structure=None, parameters=None): + super(ImpalaObjectDataset, self).__init__(description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters) + self.type = 'ImpalaObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/impala_source.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/impala_source.py new file mode 100644 index 000000000000..985e825de5c4 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/impala_source.py @@ -0,0 +1,46 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .copy_source import CopySource + + +class ImpalaSource(CopySource): + """A copy activity Impala server source. + + :param source_retry_count: Source retry count. Type: integer (or + Expression with resultType integer). + :type source_retry_count: object + :param source_retry_wait: Source retry wait. Type: string (or Expression + with resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type source_retry_wait: object + :param type: Constant filled by server. + :type type: str + :param query: A query to retrieve data from source. Type: string (or + Expression with resultType string). + :type query: object + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, + 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, + 'query': {'key': 'query', 'type': 'object'}, + } + + def __init__(self, source_retry_count=None, source_retry_wait=None, query=None): + super(ImpalaSource, self).__init__(source_retry_count=source_retry_count, source_retry_wait=source_retry_wait) + self.query = query + self.type = 'ImpalaSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/jira_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/jira_linked_service.py new file mode 100644 index 000000000000..445fff963393 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/jira_linked_service.py @@ -0,0 +1,83 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .linked_service import LinkedService + + +class JiraLinkedService(LinkedService): + """Jira Serivce linked service. + + :param connect_via: The integration runtime reference. + :type connect_via: + ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :param description: Linked service description. + :type description: str + :param type: Constant filled by server. + :type type: str + :param host: The IP address or host name of the Jira service. (e.g. + jira.example.com) + :type host: object + :param port: The TCP port that the Jira server uses to listen for client + connections. The default value is 443 if connecting through HTTPS, or 8080 + if connecting through HTTP. + :type port: object + :param username: The user name that you use to access Jira Service. + :type username: object + :param password: The password corresponding to the user name that you + provided in the username field. + :type password: ~azure.mgmt.datafactory.models.SecretBase + :param use_encrypted_endpoints: Specifies whether the data source + endpoints are encrypted using HTTPS. The default value is true. + :type use_encrypted_endpoints: object + :param use_host_verification: Specifies whether to require the host name + in the server's certificate to match the host name of the server when + connecting over SSL. The default value is true. + :type use_host_verification: object + :param use_peer_verification: Specifies whether to verify the identity of + the server when connecting over SSL. The default value is true. + :type use_peer_verification: object + :param encrypted_credential: The encrypted credential used for + authentication. Credentials are encrypted using the integration runtime + credential manager. Type: string (or Expression with resultType string). + :type encrypted_credential: object + """ + + _validation = { + 'type': {'required': True}, + 'host': {'required': True}, + 'username': {'required': True}, + } + + _attribute_map = { + 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, + 'description': {'key': 'description', 'type': 'str'}, + 'type': {'key': 'type', 'type': 'str'}, + 'host': {'key': 'typeProperties.host', 'type': 'object'}, + 'port': {'key': 'typeProperties.port', 'type': 'object'}, + 'username': {'key': 'typeProperties.username', 'type': 'object'}, + 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, + 'use_encrypted_endpoints': {'key': 'typeProperties.useEncryptedEndpoints', 'type': 'object'}, + 'use_host_verification': {'key': 'typeProperties.useHostVerification', 'type': 'object'}, + 'use_peer_verification': {'key': 'typeProperties.usePeerVerification', 'type': 'object'}, + 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, + } + + def __init__(self, host, username, connect_via=None, description=None, port=None, password=None, use_encrypted_endpoints=None, use_host_verification=None, use_peer_verification=None, encrypted_credential=None): + super(JiraLinkedService, self).__init__(connect_via=connect_via, description=description) + self.host = host + self.port = port + self.username = username + self.password = password + self.use_encrypted_endpoints = use_encrypted_endpoints + self.use_host_verification = use_host_verification + self.use_peer_verification = use_peer_verification + self.encrypted_credential = encrypted_credential + self.type = 'Jira' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/jira_object_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/jira_object_dataset.py new file mode 100644 index 000000000000..ef91e1a191db --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/jira_object_dataset.py @@ -0,0 +1,40 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .dataset import Dataset + + +class JiraObjectDataset(Dataset): + """Jira Serivce dataset. + + :param description: Dataset description. + :type description: str + :param structure: Columns that define the structure of the dataset. Type: + array (or Expression with resultType array), itemType: DatasetDataElement. + :type structure: object + :param linked_service_name: Linked service reference. + :type linked_service_name: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param parameters: Parameters for dataset. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param type: Constant filled by server. + :type type: str + """ + + _validation = { + 'linked_service_name': {'required': True}, + 'type': {'required': True}, + } + + def __init__(self, linked_service_name, description=None, structure=None, parameters=None): + super(JiraObjectDataset, self).__init__(description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters) + self.type = 'JiraObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/jira_source.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/jira_source.py new file mode 100644 index 000000000000..632b07dbdb8c --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/jira_source.py @@ -0,0 +1,46 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .copy_source import CopySource + + +class JiraSource(CopySource): + """A copy activity Jira Serivce source. + + :param source_retry_count: Source retry count. Type: integer (or + Expression with resultType integer). + :type source_retry_count: object + :param source_retry_wait: Source retry wait. Type: string (or Expression + with resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type source_retry_wait: object + :param type: Constant filled by server. + :type type: str + :param query: A query to retrieve data from source. Type: string (or + Expression with resultType string). + :type query: object + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, + 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, + 'query': {'key': 'query', 'type': 'object'}, + } + + def __init__(self, source_retry_count=None, source_retry_wait=None, query=None): + super(JiraSource, self).__init__(source_retry_count=source_retry_count, source_retry_wait=source_retry_wait) + self.query = query + self.type = 'JiraSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/json_format.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/json_format.py index 2573b78038b4..2368880c7695 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/json_format.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/json_format.py @@ -37,10 +37,20 @@ class JsonFormat(DatasetStorageFormat): provided, the default value is 'utf-8', unless the byte order mark (BOM) denotes another Unicode encoding. The full list of supported values can be found in the 'Name' column of the table of encodings in the following - reference: - https://msdn.microsoft.com/library/system.text.encoding.aspx#Anchor_5. - Type: string (or Expression with resultType string). + reference: https://go.microsoft.com/fwlink/?linkid=861078. Type: string + (or Expression with resultType string). :type encoding_name: object + :param json_node_reference: The JSONPath of the JSON array element to be + flattened. Example: "$.ArrayPath". Type: string (or Expression with + resultType string). + :type json_node_reference: object + :param json_path_definition: The JSONPath definition for each column + mapping with a customized column name to extract data from JSON file. For + fields under root object, start with "$"; for fields inside the array + chosen by jsonNodeReference property, start from the array element. + Example: {"Column1": "$.Column1Path", "Column2": "Column2PathInArray"}. + Type: object (or Expression with resultType object). + :type json_path_definition: object """ _validation = { @@ -54,11 +64,15 @@ class JsonFormat(DatasetStorageFormat): 'file_pattern': {'key': 'filePattern', 'type': 'str'}, 'nesting_separator': {'key': 'nestingSeparator', 'type': 'object'}, 'encoding_name': {'key': 'encodingName', 'type': 'object'}, + 'json_node_reference': {'key': 'jsonNodeReference', 'type': 'object'}, + 'json_path_definition': {'key': 'jsonPathDefinition', 'type': 'object'}, } - def __init__(self, serializer=None, deserializer=None, file_pattern=None, nesting_separator=None, encoding_name=None): + def __init__(self, serializer=None, deserializer=None, file_pattern=None, nesting_separator=None, encoding_name=None, json_node_reference=None, json_path_definition=None): super(JsonFormat, self).__init__(serializer=serializer, deserializer=deserializer) self.file_pattern = file_pattern self.nesting_separator = nesting_separator self.encoding_name = encoding_name + self.json_node_reference = json_node_reference + self.json_path_definition = json_path_definition self.type = 'JsonFormat' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_service.py index 03ceb36f45f8..26d35cc239f4 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_service.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/linked_service.py @@ -19,19 +19,29 @@ class LinkedService(Model): You probably want to use the sub-classes and not this class directly. Known sub-classes are: AzureDataLakeAnalyticsLinkedService, - HDInsightOnDemandLinkedService, SapHanaLinkedService, SapBWLinkedService, - SftpServerLinkedService, FtpServerLinkedService, HttpLinkedService, - AzureSearchLinkedService, CustomDataSourceLinkedService, - AmazonRedshiftLinkedService, AmazonS3LinkedService, + HDInsightOnDemandLinkedService, ZohoLinkedService, XeroLinkedService, + SquareLinkedService, SparkLinkedService, ShopifyLinkedService, + ServiceNowLinkedService, QuickBooksLinkedService, PrestoLinkedService, + PhoenixLinkedService, PaypalLinkedService, MarketoLinkedService, + MariaDBLinkedService, MagentoLinkedService, JiraLinkedService, + ImpalaLinkedService, HubspotLinkedService, HiveLinkedService, + HBaseLinkedService, GreenplumLinkedService, GoogleBigQueryLinkedService, + EloquaLinkedService, DrillLinkedService, CouchbaseLinkedService, + ConcurLinkedService, AzurePostgreSqlLinkedService, AmazonMWSLinkedService, + SapHanaLinkedService, SapBWLinkedService, SftpServerLinkedService, + FtpServerLinkedService, HttpLinkedService, AzureSearchLinkedService, + CustomDataSourceLinkedService, AmazonRedshiftLinkedService, + AmazonS3LinkedService, SapCloudForCustomerLinkedService, SalesforceLinkedService, AzureDataLakeStoreLinkedService, MongoDbLinkedService, CassandraLinkedService, WebLinkedService, ODataLinkedService, HdfsLinkedService, OdbcLinkedService, AzureMLLinkedService, TeradataLinkedService, Db2LinkedService, SybaseLinkedService, PostgreSqlLinkedService, MySqlLinkedService, - OracleLinkedService, FileServerLinkedService, HDInsightLinkedService, - DynamicsLinkedService, CosmosDbLinkedService, AzureKeyVaultLinkedService, - AzureBatchLinkedService, AzureSqlDatabaseLinkedService, - SqlServerLinkedService, AzureSqlDWLinkedService, AzureStorageLinkedService + AzureMySqlLinkedService, OracleLinkedService, FileServerLinkedService, + HDInsightLinkedService, DynamicsLinkedService, CosmosDbLinkedService, + AzureKeyVaultLinkedService, AzureBatchLinkedService, + AzureSqlDatabaseLinkedService, SqlServerLinkedService, + AzureSqlDWLinkedService, AzureStorageLinkedService :param connect_via: The integration runtime reference. :type connect_via: @@ -53,7 +63,7 @@ class LinkedService(Model): } _subtype_map = { - 'type': {'AzureDataLakeAnalytics': 'AzureDataLakeAnalyticsLinkedService', 'HDInsightOnDemand': 'HDInsightOnDemandLinkedService', 'SapHana': 'SapHanaLinkedService', 'SapBW': 'SapBWLinkedService', 'Sftp': 'SftpServerLinkedService', 'FtpServer': 'FtpServerLinkedService', 'HttpServer': 'HttpLinkedService', 'AzureSearch': 'AzureSearchLinkedService', 'CustomDataSource': 'CustomDataSourceLinkedService', 'AmazonRedshift': 'AmazonRedshiftLinkedService', 'AmazonS3': 'AmazonS3LinkedService', 'Salesforce': 'SalesforceLinkedService', 'AzureDataLakeStore': 'AzureDataLakeStoreLinkedService', 'MongoDb': 'MongoDbLinkedService', 'Cassandra': 'CassandraLinkedService', 'Web': 'WebLinkedService', 'OData': 'ODataLinkedService', 'Hdfs': 'HdfsLinkedService', 'Odbc': 'OdbcLinkedService', 'AzureML': 'AzureMLLinkedService', 'Teradata': 'TeradataLinkedService', 'Db2': 'Db2LinkedService', 'Sybase': 'SybaseLinkedService', 'PostgreSql': 'PostgreSqlLinkedService', 'MySql': 'MySqlLinkedService', 'Oracle': 'OracleLinkedService', 'FileServer': 'FileServerLinkedService', 'HDInsight': 'HDInsightLinkedService', 'Dynamics': 'DynamicsLinkedService', 'CosmosDb': 'CosmosDbLinkedService', 'AzureKeyVault': 'AzureKeyVaultLinkedService', 'AzureBatch': 'AzureBatchLinkedService', 'AzureSqlDatabase': 'AzureSqlDatabaseLinkedService', 'SqlServer': 'SqlServerLinkedService', 'AzureSqlDW': 'AzureSqlDWLinkedService', 'AzureStorage': 'AzureStorageLinkedService'} + 'type': {'AzureDataLakeAnalytics': 'AzureDataLakeAnalyticsLinkedService', 'HDInsightOnDemand': 'HDInsightOnDemandLinkedService', 'Zoho': 'ZohoLinkedService', 'Xero': 'XeroLinkedService', 'Square': 'SquareLinkedService', 'Spark': 'SparkLinkedService', 'Shopify': 'ShopifyLinkedService', 'ServiceNow': 'ServiceNowLinkedService', 'QuickBooks': 'QuickBooksLinkedService', 'Presto': 'PrestoLinkedService', 'Phoenix': 'PhoenixLinkedService', 'Paypal': 'PaypalLinkedService', 'Marketo': 'MarketoLinkedService', 'MariaDB': 'MariaDBLinkedService', 'Magento': 'MagentoLinkedService', 'Jira': 'JiraLinkedService', 'Impala': 'ImpalaLinkedService', 'Hubspot': 'HubspotLinkedService', 'Hive': 'HiveLinkedService', 'HBase': 'HBaseLinkedService', 'Greenplum': 'GreenplumLinkedService', 'GoogleBigQuery': 'GoogleBigQueryLinkedService', 'Eloqua': 'EloquaLinkedService', 'Drill': 'DrillLinkedService', 'Couchbase': 'CouchbaseLinkedService', 'Concur': 'ConcurLinkedService', 'AzurePostgreSql': 'AzurePostgreSqlLinkedService', 'AmazonMWS': 'AmazonMWSLinkedService', 'SapHana': 'SapHanaLinkedService', 'SapBW': 'SapBWLinkedService', 'Sftp': 'SftpServerLinkedService', 'FtpServer': 'FtpServerLinkedService', 'HttpServer': 'HttpLinkedService', 'AzureSearch': 'AzureSearchLinkedService', 'CustomDataSource': 'CustomDataSourceLinkedService', 'AmazonRedshift': 'AmazonRedshiftLinkedService', 'AmazonS3': 'AmazonS3LinkedService', 'SapCloudForCustomer': 'SapCloudForCustomerLinkedService', 'Salesforce': 'SalesforceLinkedService', 'AzureDataLakeStore': 'AzureDataLakeStoreLinkedService', 'MongoDb': 'MongoDbLinkedService', 'Cassandra': 'CassandraLinkedService', 'Web': 'WebLinkedService', 'OData': 'ODataLinkedService', 'Hdfs': 'HdfsLinkedService', 'Odbc': 'OdbcLinkedService', 'AzureML': 'AzureMLLinkedService', 'Teradata': 'TeradataLinkedService', 'Db2': 'Db2LinkedService', 'Sybase': 'SybaseLinkedService', 'PostgreSql': 'PostgreSqlLinkedService', 'MySql': 'MySqlLinkedService', 'AzureMySql': 'AzureMySqlLinkedService', 'Oracle': 'OracleLinkedService', 'FileServer': 'FileServerLinkedService', 'HDInsight': 'HDInsightLinkedService', 'Dynamics': 'DynamicsLinkedService', 'CosmosDb': 'CosmosDbLinkedService', 'AzureKeyVault': 'AzureKeyVaultLinkedService', 'AzureBatch': 'AzureBatchLinkedService', 'AzureSqlDatabase': 'AzureSqlDatabaseLinkedService', 'SqlServer': 'SqlServerLinkedService', 'AzureSqlDW': 'AzureSqlDWLinkedService', 'AzureStorage': 'AzureStorageLinkedService'} } def __init__(self, connect_via=None, description=None): diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/magento_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/magento_linked_service.py new file mode 100644 index 000000000000..a57e4ac49992 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/magento_linked_service.py @@ -0,0 +1,71 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .linked_service import LinkedService + + +class MagentoLinkedService(LinkedService): + """Magento server linked service. + + :param connect_via: The integration runtime reference. + :type connect_via: + ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :param description: Linked service description. + :type description: str + :param type: Constant filled by server. + :type type: str + :param host: The URL of the Magento instance. (i.e. + 192.168.222.110/magento3) + :type host: object + :param access_token: The access token from Magento. + :type access_token: ~azure.mgmt.datafactory.models.SecretBase + :param use_encrypted_endpoints: Specifies whether the data source + endpoints are encrypted using HTTPS. The default value is true. + :type use_encrypted_endpoints: object + :param use_host_verification: Specifies whether to require the host name + in the server's certificate to match the host name of the server when + connecting over SSL. The default value is true. + :type use_host_verification: object + :param use_peer_verification: Specifies whether to verify the identity of + the server when connecting over SSL. The default value is true. + :type use_peer_verification: object + :param encrypted_credential: The encrypted credential used for + authentication. Credentials are encrypted using the integration runtime + credential manager. Type: string (or Expression with resultType string). + :type encrypted_credential: object + """ + + _validation = { + 'type': {'required': True}, + 'host': {'required': True}, + } + + _attribute_map = { + 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, + 'description': {'key': 'description', 'type': 'str'}, + 'type': {'key': 'type', 'type': 'str'}, + 'host': {'key': 'typeProperties.host', 'type': 'object'}, + 'access_token': {'key': 'typeProperties.accessToken', 'type': 'SecretBase'}, + 'use_encrypted_endpoints': {'key': 'typeProperties.useEncryptedEndpoints', 'type': 'object'}, + 'use_host_verification': {'key': 'typeProperties.useHostVerification', 'type': 'object'}, + 'use_peer_verification': {'key': 'typeProperties.usePeerVerification', 'type': 'object'}, + 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, + } + + def __init__(self, host, connect_via=None, description=None, access_token=None, use_encrypted_endpoints=None, use_host_verification=None, use_peer_verification=None, encrypted_credential=None): + super(MagentoLinkedService, self).__init__(connect_via=connect_via, description=description) + self.host = host + self.access_token = access_token + self.use_encrypted_endpoints = use_encrypted_endpoints + self.use_host_verification = use_host_verification + self.use_peer_verification = use_peer_verification + self.encrypted_credential = encrypted_credential + self.type = 'Magento' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/magento_object_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/magento_object_dataset.py new file mode 100644 index 000000000000..fedb46a0e44f --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/magento_object_dataset.py @@ -0,0 +1,40 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .dataset import Dataset + + +class MagentoObjectDataset(Dataset): + """Magento server dataset. + + :param description: Dataset description. + :type description: str + :param structure: Columns that define the structure of the dataset. Type: + array (or Expression with resultType array), itemType: DatasetDataElement. + :type structure: object + :param linked_service_name: Linked service reference. + :type linked_service_name: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param parameters: Parameters for dataset. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param type: Constant filled by server. + :type type: str + """ + + _validation = { + 'linked_service_name': {'required': True}, + 'type': {'required': True}, + } + + def __init__(self, linked_service_name, description=None, structure=None, parameters=None): + super(MagentoObjectDataset, self).__init__(description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters) + self.type = 'MagentoObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/magento_source.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/magento_source.py new file mode 100644 index 000000000000..553e681b687a --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/magento_source.py @@ -0,0 +1,46 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .copy_source import CopySource + + +class MagentoSource(CopySource): + """A copy activity Magento server source. + + :param source_retry_count: Source retry count. Type: integer (or + Expression with resultType integer). + :type source_retry_count: object + :param source_retry_wait: Source retry wait. Type: string (or Expression + with resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type source_retry_wait: object + :param type: Constant filled by server. + :type type: str + :param query: A query to retrieve data from source. Type: string (or + Expression with resultType string). + :type query: object + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, + 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, + 'query': {'key': 'query', 'type': 'object'}, + } + + def __init__(self, source_retry_count=None, source_retry_wait=None, query=None): + super(MagentoSource, self).__init__(source_retry_count=source_retry_count, source_retry_wait=source_retry_wait) + self.query = query + self.type = 'MagentoSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/maria_db_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/maria_db_linked_service.py new file mode 100644 index 000000000000..489ede3358af --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/maria_db_linked_service.py @@ -0,0 +1,49 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .linked_service import LinkedService + + +class MariaDBLinkedService(LinkedService): + """MariaDB server linked service. + + :param connect_via: The integration runtime reference. + :type connect_via: + ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :param description: Linked service description. + :type description: str + :param type: Constant filled by server. + :type type: str + :param connection_string: An ODBC connection string. + :type connection_string: ~azure.mgmt.datafactory.models.SecretBase + :param encrypted_credential: The encrypted credential used for + authentication. Credentials are encrypted using the integration runtime + credential manager. Type: string (or Expression with resultType string). + :type encrypted_credential: object + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, + 'description': {'key': 'description', 'type': 'str'}, + 'type': {'key': 'type', 'type': 'str'}, + 'connection_string': {'key': 'typeProperties.connectionString', 'type': 'SecretBase'}, + 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, + } + + def __init__(self, connect_via=None, description=None, connection_string=None, encrypted_credential=None): + super(MariaDBLinkedService, self).__init__(connect_via=connect_via, description=description) + self.connection_string = connection_string + self.encrypted_credential = encrypted_credential + self.type = 'MariaDB' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/maria_db_source.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/maria_db_source.py new file mode 100644 index 000000000000..15ffad7a6cf2 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/maria_db_source.py @@ -0,0 +1,46 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .copy_source import CopySource + + +class MariaDBSource(CopySource): + """A copy activity MariaDB server source. + + :param source_retry_count: Source retry count. Type: integer (or + Expression with resultType integer). + :type source_retry_count: object + :param source_retry_wait: Source retry wait. Type: string (or Expression + with resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type source_retry_wait: object + :param type: Constant filled by server. + :type type: str + :param query: A query to retrieve data from source. Type: string (or + Expression with resultType string). + :type query: object + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, + 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, + 'query': {'key': 'query', 'type': 'object'}, + } + + def __init__(self, source_retry_count=None, source_retry_wait=None, query=None): + super(MariaDBSource, self).__init__(source_retry_count=source_retry_count, source_retry_wait=source_retry_wait) + self.query = query + self.type = 'MariaDBSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/maria_db_table_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/maria_db_table_dataset.py new file mode 100644 index 000000000000..550ad5375ad9 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/maria_db_table_dataset.py @@ -0,0 +1,40 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .dataset import Dataset + + +class MariaDBTableDataset(Dataset): + """MariaDB server dataset. + + :param description: Dataset description. + :type description: str + :param structure: Columns that define the structure of the dataset. Type: + array (or Expression with resultType array), itemType: DatasetDataElement. + :type structure: object + :param linked_service_name: Linked service reference. + :type linked_service_name: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param parameters: Parameters for dataset. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param type: Constant filled by server. + :type type: str + """ + + _validation = { + 'linked_service_name': {'required': True}, + 'type': {'required': True}, + } + + def __init__(self, linked_service_name, description=None, structure=None, parameters=None): + super(MariaDBTableDataset, self).__init__(description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters) + self.type = 'MariaDBTable' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/marketo_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/marketo_linked_service.py new file mode 100644 index 000000000000..771ed48cbca9 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/marketo_linked_service.py @@ -0,0 +1,76 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .linked_service import LinkedService + + +class MarketoLinkedService(LinkedService): + """Marketo server linked service. + + :param connect_via: The integration runtime reference. + :type connect_via: + ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :param description: Linked service description. + :type description: str + :param type: Constant filled by server. + :type type: str + :param endpoint: The endpoint of the Marketo server. (i.e. + 123-ABC-321.mktorest.com) + :type endpoint: object + :param client_id: The client Id of your Marketo service. + :type client_id: object + :param client_secret: The client secret of your Marketo service. + :type client_secret: ~azure.mgmt.datafactory.models.SecretBase + :param use_encrypted_endpoints: Specifies whether the data source + endpoints are encrypted using HTTPS. The default value is true. + :type use_encrypted_endpoints: object + :param use_host_verification: Specifies whether to require the host name + in the server's certificate to match the host name of the server when + connecting over SSL. The default value is true. + :type use_host_verification: object + :param use_peer_verification: Specifies whether to verify the identity of + the server when connecting over SSL. The default value is true. + :type use_peer_verification: object + :param encrypted_credential: The encrypted credential used for + authentication. Credentials are encrypted using the integration runtime + credential manager. Type: string (or Expression with resultType string). + :type encrypted_credential: object + """ + + _validation = { + 'type': {'required': True}, + 'endpoint': {'required': True}, + 'client_id': {'required': True}, + } + + _attribute_map = { + 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, + 'description': {'key': 'description', 'type': 'str'}, + 'type': {'key': 'type', 'type': 'str'}, + 'endpoint': {'key': 'typeProperties.endpoint', 'type': 'object'}, + 'client_id': {'key': 'typeProperties.clientId', 'type': 'object'}, + 'client_secret': {'key': 'typeProperties.clientSecret', 'type': 'SecretBase'}, + 'use_encrypted_endpoints': {'key': 'typeProperties.useEncryptedEndpoints', 'type': 'object'}, + 'use_host_verification': {'key': 'typeProperties.useHostVerification', 'type': 'object'}, + 'use_peer_verification': {'key': 'typeProperties.usePeerVerification', 'type': 'object'}, + 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, + } + + def __init__(self, endpoint, client_id, connect_via=None, description=None, client_secret=None, use_encrypted_endpoints=None, use_host_verification=None, use_peer_verification=None, encrypted_credential=None): + super(MarketoLinkedService, self).__init__(connect_via=connect_via, description=description) + self.endpoint = endpoint + self.client_id = client_id + self.client_secret = client_secret + self.use_encrypted_endpoints = use_encrypted_endpoints + self.use_host_verification = use_host_verification + self.use_peer_verification = use_peer_verification + self.encrypted_credential = encrypted_credential + self.type = 'Marketo' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/marketo_object_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/marketo_object_dataset.py new file mode 100644 index 000000000000..319c9b39fd5c --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/marketo_object_dataset.py @@ -0,0 +1,40 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .dataset import Dataset + + +class MarketoObjectDataset(Dataset): + """Marketo server dataset. + + :param description: Dataset description. + :type description: str + :param structure: Columns that define the structure of the dataset. Type: + array (or Expression with resultType array), itemType: DatasetDataElement. + :type structure: object + :param linked_service_name: Linked service reference. + :type linked_service_name: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param parameters: Parameters for dataset. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param type: Constant filled by server. + :type type: str + """ + + _validation = { + 'linked_service_name': {'required': True}, + 'type': {'required': True}, + } + + def __init__(self, linked_service_name, description=None, structure=None, parameters=None): + super(MarketoObjectDataset, self).__init__(description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters) + self.type = 'MarketoObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/marketo_source.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/marketo_source.py new file mode 100644 index 000000000000..c630336d1395 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/marketo_source.py @@ -0,0 +1,46 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .copy_source import CopySource + + +class MarketoSource(CopySource): + """A copy activity Marketo server source. + + :param source_retry_count: Source retry count. Type: integer (or + Expression with resultType integer). + :type source_retry_count: object + :param source_retry_wait: Source retry wait. Type: string (or Expression + with resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type source_retry_wait: object + :param type: Constant filled by server. + :type type: str + :param query: A query to retrieve data from source. Type: string (or + Expression with resultType string). + :type query: object + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, + 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, + 'query': {'key': 'query', 'type': 'object'}, + } + + def __init__(self, source_retry_count=None, source_retry_wait=None, query=None): + super(MarketoSource, self).__init__(source_retry_count=source_retry_count, source_retry_wait=source_retry_wait) + self.query = query + self.type = 'MarketoSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/multiple_pipeline_trigger.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/multiple_pipeline_trigger.py new file mode 100644 index 000000000000..8ff79023a9fa --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/multiple_pipeline_trigger.py @@ -0,0 +1,58 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .trigger import Trigger + + +class MultiplePipelineTrigger(Trigger): + """Base class for all triggers that support one to many model for trigger to + pipeline. + + You probably want to use the sub-classes and not this class directly. Known + sub-classes are: BlobTrigger, ScheduleTrigger + + Variables are only populated by the server, and will be ignored when + sending a request. + + :param description: Trigger description. + :type description: str + :ivar runtime_state: Indicates if trigger is running or not. Updated when + Start/Stop APIs are called on the Trigger. Possible values include: + 'Started', 'Stopped', 'Disabled' + :vartype runtime_state: str or + ~azure.mgmt.datafactory.models.TriggerRuntimeState + :param type: Constant filled by server. + :type type: str + :param pipelines: Pipelines that need to be started. + :type pipelines: + list[~azure.mgmt.datafactory.models.TriggerPipelineReference] + """ + + _validation = { + 'runtime_state': {'readonly': True}, + 'type': {'required': True}, + } + + _attribute_map = { + 'description': {'key': 'description', 'type': 'str'}, + 'runtime_state': {'key': 'runtimeState', 'type': 'str'}, + 'type': {'key': 'type', 'type': 'str'}, + 'pipelines': {'key': 'pipelines', 'type': '[TriggerPipelineReference]'}, + } + + _subtype_map = { + 'type': {'BlobTrigger': 'BlobTrigger', 'ScheduleTrigger': 'ScheduleTrigger'} + } + + def __init__(self, description=None, pipelines=None): + super(MultiplePipelineTrigger, self).__init__(description=description) + self.pipelines = pipelines + self.type = 'MultiplePipelineTrigger' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/paypal_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/paypal_linked_service.py new file mode 100644 index 000000000000..f57795a5d797 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/paypal_linked_service.py @@ -0,0 +1,76 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .linked_service import LinkedService + + +class PaypalLinkedService(LinkedService): + """Paypal Serivce linked service. + + :param connect_via: The integration runtime reference. + :type connect_via: + ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :param description: Linked service description. + :type description: str + :param type: Constant filled by server. + :type type: str + :param host: The URL of the PayPal instance. (i.e. api.sandbox.paypal.com) + :type host: object + :param client_id: The client ID associated with your PayPal application. + :type client_id: object + :param client_secret: The client secret associated with your PayPal + application. + :type client_secret: ~azure.mgmt.datafactory.models.SecretBase + :param use_encrypted_endpoints: Specifies whether the data source + endpoints are encrypted using HTTPS. The default value is true. + :type use_encrypted_endpoints: object + :param use_host_verification: Specifies whether to require the host name + in the server's certificate to match the host name of the server when + connecting over SSL. The default value is true. + :type use_host_verification: object + :param use_peer_verification: Specifies whether to verify the identity of + the server when connecting over SSL. The default value is true. + :type use_peer_verification: object + :param encrypted_credential: The encrypted credential used for + authentication. Credentials are encrypted using the integration runtime + credential manager. Type: string (or Expression with resultType string). + :type encrypted_credential: object + """ + + _validation = { + 'type': {'required': True}, + 'host': {'required': True}, + 'client_id': {'required': True}, + } + + _attribute_map = { + 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, + 'description': {'key': 'description', 'type': 'str'}, + 'type': {'key': 'type', 'type': 'str'}, + 'host': {'key': 'typeProperties.host', 'type': 'object'}, + 'client_id': {'key': 'typeProperties.clientId', 'type': 'object'}, + 'client_secret': {'key': 'typeProperties.clientSecret', 'type': 'SecretBase'}, + 'use_encrypted_endpoints': {'key': 'typeProperties.useEncryptedEndpoints', 'type': 'object'}, + 'use_host_verification': {'key': 'typeProperties.useHostVerification', 'type': 'object'}, + 'use_peer_verification': {'key': 'typeProperties.usePeerVerification', 'type': 'object'}, + 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, + } + + def __init__(self, host, client_id, connect_via=None, description=None, client_secret=None, use_encrypted_endpoints=None, use_host_verification=None, use_peer_verification=None, encrypted_credential=None): + super(PaypalLinkedService, self).__init__(connect_via=connect_via, description=description) + self.host = host + self.client_id = client_id + self.client_secret = client_secret + self.use_encrypted_endpoints = use_encrypted_endpoints + self.use_host_verification = use_host_verification + self.use_peer_verification = use_peer_verification + self.encrypted_credential = encrypted_credential + self.type = 'Paypal' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/paypal_object_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/paypal_object_dataset.py new file mode 100644 index 000000000000..482297645284 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/paypal_object_dataset.py @@ -0,0 +1,40 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .dataset import Dataset + + +class PaypalObjectDataset(Dataset): + """Paypal Serivce dataset. + + :param description: Dataset description. + :type description: str + :param structure: Columns that define the structure of the dataset. Type: + array (or Expression with resultType array), itemType: DatasetDataElement. + :type structure: object + :param linked_service_name: Linked service reference. + :type linked_service_name: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param parameters: Parameters for dataset. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param type: Constant filled by server. + :type type: str + """ + + _validation = { + 'linked_service_name': {'required': True}, + 'type': {'required': True}, + } + + def __init__(self, linked_service_name, description=None, structure=None, parameters=None): + super(PaypalObjectDataset, self).__init__(description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters) + self.type = 'PaypalObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/paypal_source.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/paypal_source.py new file mode 100644 index 000000000000..7066629a6bd1 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/paypal_source.py @@ -0,0 +1,46 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .copy_source import CopySource + + +class PaypalSource(CopySource): + """A copy activity Paypal Serivce source. + + :param source_retry_count: Source retry count. Type: integer (or + Expression with resultType integer). + :type source_retry_count: object + :param source_retry_wait: Source retry wait. Type: string (or Expression + with resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type source_retry_wait: object + :param type: Constant filled by server. + :type type: str + :param query: A query to retrieve data from source. Type: string (or + Expression with resultType string). + :type query: object + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, + 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, + 'query': {'key': 'query', 'type': 'object'}, + } + + def __init__(self, source_retry_count=None, source_retry_wait=None, query=None): + super(PaypalSource, self).__init__(source_retry_count=source_retry_count, source_retry_wait=source_retry_wait) + self.query = query + self.type = 'PaypalSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/phoenix_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/phoenix_linked_service.py new file mode 100644 index 000000000000..3ff7adb1b2d9 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/phoenix_linked_service.py @@ -0,0 +1,107 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .linked_service import LinkedService + + +class PhoenixLinkedService(LinkedService): + """Phoenix server linked service. + + :param connect_via: The integration runtime reference. + :type connect_via: + ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :param description: Linked service description. + :type description: str + :param type: Constant filled by server. + :type type: str + :param host: The IP address or host name of the Phoenix server. (i.e. + 192.168.222.160) + :type host: object + :param port: The TCP port that the Phoenix server uses to listen for + client connections. The default value is 8765. + :type port: object + :param http_path: The partial URL corresponding to the Phoenix server. + (i.e. /gateway/sandbox/phoenix/version). The default value is hbasephoenix + if using WindowsAzureHDInsightService. + :type http_path: object + :param authentication_type: The authentication mechanism used to connect + to the Phoenix server. Possible values include: 'Anonymous', + 'UsernameAndPassword', 'WindowsAzureHDInsightService' + :type authentication_type: str or + ~azure.mgmt.datafactory.models.PhoenixAuthenticationType + :param username: The user name used to connect to the Phoenix server. + :type username: object + :param password: The password corresponding to the user name. + :type password: ~azure.mgmt.datafactory.models.SecretBase + :param enable_ssl: Specifies whether the connections to the server are + encrypted using SSL. The default value is false. + :type enable_ssl: object + :param trusted_cert_path: The full path of the .pem file containing + trusted CA certificates for verifying the server when connecting over SSL. + This property can only be set when using SSL on self-hosted IR. The + default value is the cacerts.pem file installed with the IR. + :type trusted_cert_path: object + :param use_system_trust_store: Specifies whether to use a CA certificate + from the system trust store or from a specified PEM file. The default + value is false. + :type use_system_trust_store: object + :param allow_host_name_cn_mismatch: Specifies whether to require a + CA-issued SSL certificate name to match the host name of the server when + connecting over SSL. The default value is false. + :type allow_host_name_cn_mismatch: object + :param allow_self_signed_server_cert: Specifies whether to allow + self-signed certificates from the server. The default value is false. + :type allow_self_signed_server_cert: object + :param encrypted_credential: The encrypted credential used for + authentication. Credentials are encrypted using the integration runtime + credential manager. Type: string (or Expression with resultType string). + :type encrypted_credential: object + """ + + _validation = { + 'type': {'required': True}, + 'host': {'required': True}, + 'authentication_type': {'required': True}, + } + + _attribute_map = { + 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, + 'description': {'key': 'description', 'type': 'str'}, + 'type': {'key': 'type', 'type': 'str'}, + 'host': {'key': 'typeProperties.host', 'type': 'object'}, + 'port': {'key': 'typeProperties.port', 'type': 'object'}, + 'http_path': {'key': 'typeProperties.httpPath', 'type': 'object'}, + 'authentication_type': {'key': 'typeProperties.authenticationType', 'type': 'str'}, + 'username': {'key': 'typeProperties.username', 'type': 'object'}, + 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, + 'enable_ssl': {'key': 'typeProperties.enableSsl', 'type': 'object'}, + 'trusted_cert_path': {'key': 'typeProperties.trustedCertPath', 'type': 'object'}, + 'use_system_trust_store': {'key': 'typeProperties.useSystemTrustStore', 'type': 'object'}, + 'allow_host_name_cn_mismatch': {'key': 'typeProperties.allowHostNameCNMismatch', 'type': 'object'}, + 'allow_self_signed_server_cert': {'key': 'typeProperties.allowSelfSignedServerCert', 'type': 'object'}, + 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, + } + + def __init__(self, host, authentication_type, connect_via=None, description=None, port=None, http_path=None, username=None, password=None, enable_ssl=None, trusted_cert_path=None, use_system_trust_store=None, allow_host_name_cn_mismatch=None, allow_self_signed_server_cert=None, encrypted_credential=None): + super(PhoenixLinkedService, self).__init__(connect_via=connect_via, description=description) + self.host = host + self.port = port + self.http_path = http_path + self.authentication_type = authentication_type + self.username = username + self.password = password + self.enable_ssl = enable_ssl + self.trusted_cert_path = trusted_cert_path + self.use_system_trust_store = use_system_trust_store + self.allow_host_name_cn_mismatch = allow_host_name_cn_mismatch + self.allow_self_signed_server_cert = allow_self_signed_server_cert + self.encrypted_credential = encrypted_credential + self.type = 'Phoenix' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/phoenix_object_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/phoenix_object_dataset.py new file mode 100644 index 000000000000..a293764bca74 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/phoenix_object_dataset.py @@ -0,0 +1,40 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .dataset import Dataset + + +class PhoenixObjectDataset(Dataset): + """Phoenix server dataset. + + :param description: Dataset description. + :type description: str + :param structure: Columns that define the structure of the dataset. Type: + array (or Expression with resultType array), itemType: DatasetDataElement. + :type structure: object + :param linked_service_name: Linked service reference. + :type linked_service_name: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param parameters: Parameters for dataset. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param type: Constant filled by server. + :type type: str + """ + + _validation = { + 'linked_service_name': {'required': True}, + 'type': {'required': True}, + } + + def __init__(self, linked_service_name, description=None, structure=None, parameters=None): + super(PhoenixObjectDataset, self).__init__(description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters) + self.type = 'PhoenixObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/phoenix_source.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/phoenix_source.py new file mode 100644 index 000000000000..7a7c6679e282 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/phoenix_source.py @@ -0,0 +1,46 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .copy_source import CopySource + + +class PhoenixSource(CopySource): + """A copy activity Phoenix server source. + + :param source_retry_count: Source retry count. Type: integer (or + Expression with resultType integer). + :type source_retry_count: object + :param source_retry_wait: Source retry wait. Type: string (or Expression + with resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type source_retry_wait: object + :param type: Constant filled by server. + :type type: str + :param query: A query to retrieve data from source. Type: string (or + Expression with resultType string). + :type query: object + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, + 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, + 'query': {'key': 'query', 'type': 'object'}, + } + + def __init__(self, source_retry_count=None, source_retry_wait=None, query=None): + super(PhoenixSource, self).__init__(source_retry_count=source_retry_count, source_retry_wait=source_retry_wait) + self.query = query + self.type = 'PhoenixSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/presto_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/presto_linked_service.py new file mode 100644 index 000000000000..de6ce6672005 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/presto_linked_service.py @@ -0,0 +1,116 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .linked_service import LinkedService + + +class PrestoLinkedService(LinkedService): + """Presto server linked service. + + :param connect_via: The integration runtime reference. + :type connect_via: + ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :param description: Linked service description. + :type description: str + :param type: Constant filled by server. + :type type: str + :param host: The IP address or host name of the Presto server. (i.e. + 192.168.222.160) + :type host: object + :param server_version: The version of the Presto server. (i.e. 0.148-t) + :type server_version: object + :param catalog: The catalog context for all request against the server. + :type catalog: object + :param port: The TCP port that the Presto server uses to listen for client + connections. The default value is 8080. + :type port: object + :param authentication_type: The authentication mechanism used to connect + to the Presto server. Possible values include: 'Anonymous', 'LDAP' + :type authentication_type: str or + ~azure.mgmt.datafactory.models.PrestoAuthenticationType + :param username: The user name used to connect to the Presto server. + :type username: object + :param password: The password corresponding to the user name. + :type password: ~azure.mgmt.datafactory.models.SecretBase + :param enable_ssl: Specifies whether the connections to the server are + encrypted using SSL. The default value is false. + :type enable_ssl: object + :param trusted_cert_path: The full path of the .pem file containing + trusted CA certificates for verifying the server when connecting over SSL. + This property can only be set when using SSL on self-hosted IR. The + default value is the cacerts.pem file installed with the IR. + :type trusted_cert_path: object + :param use_system_trust_store: Specifies whether to use a CA certificate + from the system trust store or from a specified PEM file. The default + value is false. + :type use_system_trust_store: object + :param allow_host_name_cn_mismatch: Specifies whether to require a + CA-issued SSL certificate name to match the host name of the server when + connecting over SSL. The default value is false. + :type allow_host_name_cn_mismatch: object + :param allow_self_signed_server_cert: Specifies whether to allow + self-signed certificates from the server. The default value is false. + :type allow_self_signed_server_cert: object + :param time_zone_id: The local time zone used by the connection. Valid + values for this option are specified in the IANA Time Zone Database. The + default value is the system time zone. + :type time_zone_id: object + :param encrypted_credential: The encrypted credential used for + authentication. Credentials are encrypted using the integration runtime + credential manager. Type: string (or Expression with resultType string). + :type encrypted_credential: object + """ + + _validation = { + 'type': {'required': True}, + 'host': {'required': True}, + 'server_version': {'required': True}, + 'catalog': {'required': True}, + 'authentication_type': {'required': True}, + } + + _attribute_map = { + 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, + 'description': {'key': 'description', 'type': 'str'}, + 'type': {'key': 'type', 'type': 'str'}, + 'host': {'key': 'typeProperties.host', 'type': 'object'}, + 'server_version': {'key': 'typeProperties.serverVersion', 'type': 'object'}, + 'catalog': {'key': 'typeProperties.catalog', 'type': 'object'}, + 'port': {'key': 'typeProperties.port', 'type': 'object'}, + 'authentication_type': {'key': 'typeProperties.authenticationType', 'type': 'str'}, + 'username': {'key': 'typeProperties.username', 'type': 'object'}, + 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, + 'enable_ssl': {'key': 'typeProperties.enableSsl', 'type': 'object'}, + 'trusted_cert_path': {'key': 'typeProperties.trustedCertPath', 'type': 'object'}, + 'use_system_trust_store': {'key': 'typeProperties.useSystemTrustStore', 'type': 'object'}, + 'allow_host_name_cn_mismatch': {'key': 'typeProperties.allowHostNameCNMismatch', 'type': 'object'}, + 'allow_self_signed_server_cert': {'key': 'typeProperties.allowSelfSignedServerCert', 'type': 'object'}, + 'time_zone_id': {'key': 'typeProperties.timeZoneID', 'type': 'object'}, + 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, + } + + def __init__(self, host, server_version, catalog, authentication_type, connect_via=None, description=None, port=None, username=None, password=None, enable_ssl=None, trusted_cert_path=None, use_system_trust_store=None, allow_host_name_cn_mismatch=None, allow_self_signed_server_cert=None, time_zone_id=None, encrypted_credential=None): + super(PrestoLinkedService, self).__init__(connect_via=connect_via, description=description) + self.host = host + self.server_version = server_version + self.catalog = catalog + self.port = port + self.authentication_type = authentication_type + self.username = username + self.password = password + self.enable_ssl = enable_ssl + self.trusted_cert_path = trusted_cert_path + self.use_system_trust_store = use_system_trust_store + self.allow_host_name_cn_mismatch = allow_host_name_cn_mismatch + self.allow_self_signed_server_cert = allow_self_signed_server_cert + self.time_zone_id = time_zone_id + self.encrypted_credential = encrypted_credential + self.type = 'Presto' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/presto_object_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/presto_object_dataset.py new file mode 100644 index 000000000000..661f257c330e --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/presto_object_dataset.py @@ -0,0 +1,40 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .dataset import Dataset + + +class PrestoObjectDataset(Dataset): + """Presto server dataset. + + :param description: Dataset description. + :type description: str + :param structure: Columns that define the structure of the dataset. Type: + array (or Expression with resultType array), itemType: DatasetDataElement. + :type structure: object + :param linked_service_name: Linked service reference. + :type linked_service_name: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param parameters: Parameters for dataset. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param type: Constant filled by server. + :type type: str + """ + + _validation = { + 'linked_service_name': {'required': True}, + 'type': {'required': True}, + } + + def __init__(self, linked_service_name, description=None, structure=None, parameters=None): + super(PrestoObjectDataset, self).__init__(description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters) + self.type = 'PrestoObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/presto_source.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/presto_source.py new file mode 100644 index 000000000000..606e173b5f59 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/presto_source.py @@ -0,0 +1,46 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .copy_source import CopySource + + +class PrestoSource(CopySource): + """A copy activity Presto server source. + + :param source_retry_count: Source retry count. Type: integer (or + Expression with resultType integer). + :type source_retry_count: object + :param source_retry_wait: Source retry wait. Type: string (or Expression + with resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type source_retry_wait: object + :param type: Constant filled by server. + :type type: str + :param query: A query to retrieve data from source. Type: string (or + Expression with resultType string). + :type query: object + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, + 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, + 'query': {'key': 'query', 'type': 'object'}, + } + + def __init__(self, source_retry_count=None, source_retry_wait=None, query=None): + super(PrestoSource, self).__init__(source_retry_count=source_retry_count, source_retry_wait=source_retry_wait) + self.query = query + self.type = 'PrestoSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/quick_books_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/quick_books_linked_service.py new file mode 100644 index 000000000000..d97ee5cef23d --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/quick_books_linked_service.py @@ -0,0 +1,70 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .linked_service import LinkedService + + +class QuickBooksLinkedService(LinkedService): + """QuickBooks server linked service. + + :param connect_via: The integration runtime reference. + :type connect_via: + ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :param description: Linked service description. + :type description: str + :param type: Constant filled by server. + :type type: str + :param endpoint: The endpoint of the QuickBooks server. (i.e. + quickbooks.api.intuit.com) + :type endpoint: object + :param company_id: The company ID of the QuickBooks company to authorize. + :type company_id: object + :param access_token: The access token for OAuth 1.0 authentication. + :type access_token: ~azure.mgmt.datafactory.models.SecretBase + :param access_token_secret: The access token secret for OAuth 1.0 + authentication. + :type access_token_secret: ~azure.mgmt.datafactory.models.SecretBase + :param use_encrypted_endpoints: Specifies whether the data source + endpoints are encrypted using HTTPS. The default value is true. + :type use_encrypted_endpoints: object + :param encrypted_credential: The encrypted credential used for + authentication. Credentials are encrypted using the integration runtime + credential manager. Type: string (or Expression with resultType string). + :type encrypted_credential: object + """ + + _validation = { + 'type': {'required': True}, + 'endpoint': {'required': True}, + 'company_id': {'required': True}, + } + + _attribute_map = { + 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, + 'description': {'key': 'description', 'type': 'str'}, + 'type': {'key': 'type', 'type': 'str'}, + 'endpoint': {'key': 'typeProperties.endpoint', 'type': 'object'}, + 'company_id': {'key': 'typeProperties.companyId', 'type': 'object'}, + 'access_token': {'key': 'typeProperties.accessToken', 'type': 'SecretBase'}, + 'access_token_secret': {'key': 'typeProperties.accessTokenSecret', 'type': 'SecretBase'}, + 'use_encrypted_endpoints': {'key': 'typeProperties.useEncryptedEndpoints', 'type': 'object'}, + 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, + } + + def __init__(self, endpoint, company_id, connect_via=None, description=None, access_token=None, access_token_secret=None, use_encrypted_endpoints=None, encrypted_credential=None): + super(QuickBooksLinkedService, self).__init__(connect_via=connect_via, description=description) + self.endpoint = endpoint + self.company_id = company_id + self.access_token = access_token + self.access_token_secret = access_token_secret + self.use_encrypted_endpoints = use_encrypted_endpoints + self.encrypted_credential = encrypted_credential + self.type = 'QuickBooks' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/quick_books_object_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/quick_books_object_dataset.py new file mode 100644 index 000000000000..a2effa962f66 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/quick_books_object_dataset.py @@ -0,0 +1,40 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .dataset import Dataset + + +class QuickBooksObjectDataset(Dataset): + """QuickBooks server dataset. + + :param description: Dataset description. + :type description: str + :param structure: Columns that define the structure of the dataset. Type: + array (or Expression with resultType array), itemType: DatasetDataElement. + :type structure: object + :param linked_service_name: Linked service reference. + :type linked_service_name: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param parameters: Parameters for dataset. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param type: Constant filled by server. + :type type: str + """ + + _validation = { + 'linked_service_name': {'required': True}, + 'type': {'required': True}, + } + + def __init__(self, linked_service_name, description=None, structure=None, parameters=None): + super(QuickBooksObjectDataset, self).__init__(description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters) + self.type = 'QuickBooksObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/quick_books_source.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/quick_books_source.py new file mode 100644 index 000000000000..d3bb63e649a3 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/quick_books_source.py @@ -0,0 +1,46 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .copy_source import CopySource + + +class QuickBooksSource(CopySource): + """A copy activity QuickBooks server source. + + :param source_retry_count: Source retry count. Type: integer (or + Expression with resultType integer). + :type source_retry_count: object + :param source_retry_wait: Source retry wait. Type: string (or Expression + with resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type source_retry_wait: object + :param type: Constant filled by server. + :type type: str + :param query: A query to retrieve data from source. Type: string (or + Expression with resultType string). + :type query: object + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, + 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, + 'query': {'key': 'query', 'type': 'object'}, + } + + def __init__(self, source_retry_count=None, source_retry_wait=None, query=None): + super(QuickBooksSource, self).__init__(source_retry_count=source_retry_count, source_retry_wait=source_retry_wait) + self.query = query + self.type = 'QuickBooksSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/redirect_incompatible_row_settings.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/redirect_incompatible_row_settings.py index 008b391f47ba..5f9756764247 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/redirect_incompatible_row_settings.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/redirect_incompatible_row_settings.py @@ -15,13 +15,13 @@ class RedirectIncompatibleRowSettings(Model): """Redirect incompatible row settings. - :param linked_service_name: Name of the Azure Storage or Storage SAS - linked service used for redirecting incompatible row. Must be specified if - redirectIncompatibleRowSettings is specified. Type: string (or Expression - with resultType string). + :param linked_service_name: Name of the Azure Storage, Storage SAS, or + Azure Data Lake Store linked service used for redirecting incompatible + row. Must be specified if redirectIncompatibleRowSettings is specified. + Type: string (or Expression with resultType string). :type linked_service_name: object - :param path: The path to storage for storing the redirect incompatible row - data Type: string (or Expression with resultType string). + :param path: The path for storing the redirect incompatible row data. + Type: string (or Expression with resultType string). :type path: object """ diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/retry_policy.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/retry_policy.py new file mode 100644 index 000000000000..c5bd4610c0f9 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/retry_policy.py @@ -0,0 +1,37 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from msrest.serialization import Model + + +class RetryPolicy(Model): + """Execution policy for an activity. + + :param count: Maximum ordinary retry attempts. Default is 0. Type: integer + (or Expression with resultType integer), minimum: 0. + :type count: object + :param interval_in_seconds: Interval between retries in seconds. Default + is 30. + :type interval_in_seconds: int + """ + + _validation = { + 'interval_in_seconds': {'maximum': 86400, 'minimum': 30}, + } + + _attribute_map = { + 'count': {'key': 'count', 'type': 'object'}, + 'interval_in_seconds': {'key': 'intervalInSeconds', 'type': 'int'}, + } + + def __init__(self, count=None, interval_in_seconds=None): + self.count = count + self.interval_in_seconds = interval_in_seconds diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_linked_service.py index c8cdaef1d833..4a9f28c9f4a0 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_linked_service.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_linked_service.py @@ -29,13 +29,13 @@ class SalesforceLinkedService(LinkedService): Expression with resultType string). :type environment_url: object :param username: The username for Basic authentication of the Salesforce - source. Type: string (or Expression with resultType string). + instance. Type: string (or Expression with resultType string). :type username: object :param password: The password for Basic authentication of the Salesforce - source. + instance. :type password: ~azure.mgmt.datafactory.models.SecretBase :param security_token: The security token is required to remotely access - Salesforce source. + Salesforce instance. :type security_token: ~azure.mgmt.datafactory.models.SecretBase :param encrypted_credential: The encrypted credential used for authentication. Credentials are encrypted using the integration runtime diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_object_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_object_dataset.py new file mode 100644 index 000000000000..07c4e08a8569 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_object_dataset.py @@ -0,0 +1,53 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .dataset import Dataset + + +class SalesforceObjectDataset(Dataset): + """The Salesforce object dataset. + + :param description: Dataset description. + :type description: str + :param structure: Columns that define the structure of the dataset. Type: + array (or Expression with resultType array), itemType: DatasetDataElement. + :type structure: object + :param linked_service_name: Linked service reference. + :type linked_service_name: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param parameters: Parameters for dataset. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param type: Constant filled by server. + :type type: str + :param object_api_name: The Salesforce object API name. Type: string (or + Expression with resultType string). + :type object_api_name: object + """ + + _validation = { + 'linked_service_name': {'required': True}, + 'type': {'required': True}, + } + + _attribute_map = { + 'description': {'key': 'description', 'type': 'str'}, + 'structure': {'key': 'structure', 'type': 'object'}, + 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'type': {'key': 'type', 'type': 'str'}, + 'object_api_name': {'key': 'typeProperties.objectApiName', 'type': 'object'}, + } + + def __init__(self, linked_service_name, description=None, structure=None, parameters=None, object_api_name=None): + super(SalesforceObjectDataset, self).__init__(description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters) + self.object_api_name = object_api_name + self.type = 'SalesforceObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_sink.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_sink.py new file mode 100644 index 000000000000..7e5c97f7e8a5 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_sink.py @@ -0,0 +1,73 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .copy_sink import CopySink + + +class SalesforceSink(CopySink): + """A copy activity Salesforce sink. + + :param write_batch_size: Write batch size. Type: integer (or Expression + with resultType integer), minimum: 0. + :type write_batch_size: object + :param write_batch_timeout: Write batch timeout. Type: string (or + Expression with resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type write_batch_timeout: object + :param sink_retry_count: Sink retry count. Type: integer (or Expression + with resultType integer). + :type sink_retry_count: object + :param sink_retry_wait: Sink retry wait. Type: string (or Expression with + resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type sink_retry_wait: object + :param type: Constant filled by server. + :type type: str + :param write_behavior: The write behavior for the operation. Default is + Insert. Possible values include: 'Insert', 'Upsert' + :type write_behavior: str or + ~azure.mgmt.datafactory.models.SalesforceSinkWriteBehavior + :param external_id_field_name: The name of the external ID field for + upsert operation. Default value is 'Id' column. Type: string (or + Expression with resultType string). + :type external_id_field_name: object + :param ignore_null_values: The flag indicating whether or not to ignore + null values from input dataset (except key fields) during write operation. + Default value is false. If set it to true, it means ADF will leave the + data in the destination object unchanged when doing upsert/update + operation and insert defined default value when doing insert operation, + versus ADF will update the data in the destination object to NULL when + doing upsert/update operation and insert NULL value when doing insert + operation. Type: boolean (or Expression with resultType boolean). + :type ignore_null_values: object + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'write_batch_size': {'key': 'writeBatchSize', 'type': 'object'}, + 'write_batch_timeout': {'key': 'writeBatchTimeout', 'type': 'object'}, + 'sink_retry_count': {'key': 'sinkRetryCount', 'type': 'object'}, + 'sink_retry_wait': {'key': 'sinkRetryWait', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, + 'write_behavior': {'key': 'writeBehavior', 'type': 'str'}, + 'external_id_field_name': {'key': 'externalIdFieldName', 'type': 'object'}, + 'ignore_null_values': {'key': 'ignoreNullValues', 'type': 'object'}, + } + + def __init__(self, write_batch_size=None, write_batch_timeout=None, sink_retry_count=None, sink_retry_wait=None, write_behavior=None, external_id_field_name=None, ignore_null_values=None): + super(SalesforceSink, self).__init__(write_batch_size=write_batch_size, write_batch_timeout=write_batch_timeout, sink_retry_count=sink_retry_count, sink_retry_wait=sink_retry_wait) + self.write_behavior = write_behavior + self.external_id_field_name = external_id_field_name + self.ignore_null_values = ignore_null_values + self.type = 'SalesforceSink' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_source.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_source.py new file mode 100644 index 000000000000..68b2e946f54b --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/salesforce_source.py @@ -0,0 +1,46 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .copy_source import CopySource + + +class SalesforceSource(CopySource): + """A copy activity Salesforce source. + + :param source_retry_count: Source retry count. Type: integer (or + Expression with resultType integer). + :type source_retry_count: object + :param source_retry_wait: Source retry wait. Type: string (or Expression + with resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type source_retry_wait: object + :param type: Constant filled by server. + :type type: str + :param query: Database query. Type: string (or Expression with resultType + string). + :type query: object + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, + 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, + 'query': {'key': 'query', 'type': 'object'}, + } + + def __init__(self, source_retry_count=None, source_retry_wait=None, query=None): + super(SalesforceSource, self).__init__(source_retry_count=source_retry_count, source_retry_wait=source_retry_wait) + self.query = query + self.type = 'SalesforceSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_cloud_for_customer_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_cloud_for_customer_linked_service.py new file mode 100644 index 000000000000..4e3e02c0d053 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_cloud_for_customer_linked_service.py @@ -0,0 +1,62 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .linked_service import LinkedService + + +class SapCloudForCustomerLinkedService(LinkedService): + """Linked service for SAP Cloud for Customer. + + :param connect_via: The integration runtime reference. + :type connect_via: + ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :param description: Linked service description. + :type description: str + :param type: Constant filled by server. + :type type: str + :param url: The URL of SAP Cloud for Customer OData API. For example, + '[https://[tenantname].crm.ondemand.com/sap/c4c/odata/v1]'. Type: string + (or Expression with resultType string). + :type url: object + :param username: The username for Basic authentication. Type: string (or + Expression with resultType string). + :type username: object + :param password: The password for Basic authentication. + :type password: ~azure.mgmt.datafactory.models.SecureString + :param encrypted_credential: The encrypted credential used for + authentication. Credentials are encrypted using the integration runtime + credential manager. Either encryptedCredential or username/password must + be provided. Type: string (or Expression with resultType string). + :type encrypted_credential: object + """ + + _validation = { + 'type': {'required': True}, + 'url': {'required': True}, + } + + _attribute_map = { + 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, + 'description': {'key': 'description', 'type': 'str'}, + 'type': {'key': 'type', 'type': 'str'}, + 'url': {'key': 'typeProperties.url', 'type': 'object'}, + 'username': {'key': 'typeProperties.username', 'type': 'object'}, + 'password': {'key': 'typeProperties.password', 'type': 'SecureString'}, + 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, + } + + def __init__(self, url, connect_via=None, description=None, username=None, password=None, encrypted_credential=None): + super(SapCloudForCustomerLinkedService, self).__init__(connect_via=connect_via, description=description) + self.url = url + self.username = username + self.password = password + self.encrypted_credential = encrypted_credential + self.type = 'SapCloudForCustomer' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_cloud_for_customer_resource_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_cloud_for_customer_resource_dataset.py new file mode 100644 index 000000000000..85db0b2c1ea8 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_cloud_for_customer_resource_dataset.py @@ -0,0 +1,54 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .dataset import Dataset + + +class SapCloudForCustomerResourceDataset(Dataset): + """The path of the SAP Cloud for Customer OData entity. + + :param description: Dataset description. + :type description: str + :param structure: Columns that define the structure of the dataset. Type: + array (or Expression with resultType array), itemType: DatasetDataElement. + :type structure: object + :param linked_service_name: Linked service reference. + :type linked_service_name: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param parameters: Parameters for dataset. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param type: Constant filled by server. + :type type: str + :param path: The path of the SAP Cloud for Customer OData entity. Type: + string (or Expression with resultType string). + :type path: object + """ + + _validation = { + 'linked_service_name': {'required': True}, + 'type': {'required': True}, + 'path': {'required': True}, + } + + _attribute_map = { + 'description': {'key': 'description', 'type': 'str'}, + 'structure': {'key': 'structure', 'type': 'object'}, + 'linked_service_name': {'key': 'linkedServiceName', 'type': 'LinkedServiceReference'}, + 'parameters': {'key': 'parameters', 'type': '{ParameterSpecification}'}, + 'type': {'key': 'type', 'type': 'str'}, + 'path': {'key': 'typeProperties.path', 'type': 'object'}, + } + + def __init__(self, linked_service_name, path, description=None, structure=None, parameters=None): + super(SapCloudForCustomerResourceDataset, self).__init__(description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters) + self.path = path + self.type = 'SapCloudForCustomerResource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_cloud_for_customer_sink.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_cloud_for_customer_sink.py new file mode 100644 index 000000000000..74f3bf3ceaed --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_cloud_for_customer_sink.py @@ -0,0 +1,56 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .copy_sink import CopySink + + +class SapCloudForCustomerSink(CopySink): + """A copy activity SAP Cloud for Customer sink. + + :param write_batch_size: Write batch size. Type: integer (or Expression + with resultType integer), minimum: 0. + :type write_batch_size: object + :param write_batch_timeout: Write batch timeout. Type: string (or + Expression with resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type write_batch_timeout: object + :param sink_retry_count: Sink retry count. Type: integer (or Expression + with resultType integer). + :type sink_retry_count: object + :param sink_retry_wait: Sink retry wait. Type: string (or Expression with + resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type sink_retry_wait: object + :param type: Constant filled by server. + :type type: str + :param write_behavior: The write behavior for the operation. Default is + 'Insert'. Possible values include: 'Insert', 'Update' + :type write_behavior: str or + ~azure.mgmt.datafactory.models.SapCloudForCustomerSinkWriteBehavior + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'write_batch_size': {'key': 'writeBatchSize', 'type': 'object'}, + 'write_batch_timeout': {'key': 'writeBatchTimeout', 'type': 'object'}, + 'sink_retry_count': {'key': 'sinkRetryCount', 'type': 'object'}, + 'sink_retry_wait': {'key': 'sinkRetryWait', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, + 'write_behavior': {'key': 'writeBehavior', 'type': 'str'}, + } + + def __init__(self, write_batch_size=None, write_batch_timeout=None, sink_retry_count=None, sink_retry_wait=None, write_behavior=None): + super(SapCloudForCustomerSink, self).__init__(write_batch_size=write_batch_size, write_batch_timeout=write_batch_timeout, sink_retry_count=sink_retry_count, sink_retry_wait=sink_retry_wait) + self.write_behavior = write_behavior + self.type = 'SapCloudForCustomerSink' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_cloud_for_customer_source.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_cloud_for_customer_source.py new file mode 100644 index 000000000000..ab4daaaaeca1 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/sap_cloud_for_customer_source.py @@ -0,0 +1,46 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .copy_source import CopySource + + +class SapCloudForCustomerSource(CopySource): + """A copy activity source for SAP Cloud for Customer source. + + :param source_retry_count: Source retry count. Type: integer (or + Expression with resultType integer). + :type source_retry_count: object + :param source_retry_wait: Source retry wait. Type: string (or Expression + with resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type source_retry_wait: object + :param type: Constant filled by server. + :type type: str + :param query: SAP Cloud for Customer OData query. For example, "$top=1". + Type: string (or Expression with resultType string). + :type query: object + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, + 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, + 'query': {'key': 'query', 'type': 'object'}, + } + + def __init__(self, source_retry_count=None, source_retry_wait=None, query=None): + super(SapCloudForCustomerSource, self).__init__(source_retry_count=source_retry_count, source_retry_wait=source_retry_wait) + self.query = query + self.type = 'SapCloudForCustomerSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/schedule_trigger.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/schedule_trigger.py index fa04fd8f1a12..35b8bd515fa5 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/schedule_trigger.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/schedule_trigger.py @@ -9,10 +9,10 @@ # regenerated. # -------------------------------------------------------------------------- -from .trigger import Trigger +from .multiple_pipeline_trigger import MultiplePipelineTrigger -class ScheduleTrigger(Trigger): +class ScheduleTrigger(MultiplePipelineTrigger): """Trigger that creates pipeline runs periodically, on schedule. Variables are only populated by the server, and will be ignored when @@ -20,9 +20,6 @@ class ScheduleTrigger(Trigger): :param description: Trigger description. :type description: str - :param pipelines: Pipelines that need to be started. - :type pipelines: - list[~azure.mgmt.datafactory.models.TriggerPipelineReference] :ivar runtime_state: Indicates if trigger is running or not. Updated when Start/Stop APIs are called on the Trigger. Possible values include: 'Started', 'Stopped', 'Disabled' @@ -30,6 +27,9 @@ class ScheduleTrigger(Trigger): ~azure.mgmt.datafactory.models.TriggerRuntimeState :param type: Constant filled by server. :type type: str + :param pipelines: Pipelines that need to be started. + :type pipelines: + list[~azure.mgmt.datafactory.models.TriggerPipelineReference] :param recurrence: Recurrence schedule configuration. :type recurrence: ~azure.mgmt.datafactory.models.ScheduleTriggerRecurrence """ @@ -41,9 +41,9 @@ class ScheduleTrigger(Trigger): _attribute_map = { 'description': {'key': 'description', 'type': 'str'}, - 'pipelines': {'key': 'pipelines', 'type': '[TriggerPipelineReference]'}, 'runtime_state': {'key': 'runtimeState', 'type': 'str'}, 'type': {'key': 'type', 'type': 'str'}, + 'pipelines': {'key': 'pipelines', 'type': '[TriggerPipelineReference]'}, 'recurrence': {'key': 'typeProperties.recurrence', 'type': 'ScheduleTriggerRecurrence'}, } diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/service_now_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/service_now_linked_service.py new file mode 100644 index 000000000000..bd74c7dbd59c --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/service_now_linked_service.py @@ -0,0 +1,92 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .linked_service import LinkedService + + +class ServiceNowLinkedService(LinkedService): + """ServiceNow server linked service. + + :param connect_via: The integration runtime reference. + :type connect_via: + ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :param description: Linked service description. + :type description: str + :param type: Constant filled by server. + :type type: str + :param endpoint: The endpoint of the ServiceNow server. (i.e. + ServiceNowData.com) + :type endpoint: object + :param authentication_type: The authentication type to use. Possible + values include: 'Basic', 'OAuth2' + :type authentication_type: str or + ~azure.mgmt.datafactory.models.ServiceNowAuthenticationType + :param username: The user name used to connect to the ServiceNow server + for Basic and OAuth2 authentication. + :type username: object + :param password: The password corresponding to the user name for Basic and + OAuth2 authentication. + :type password: ~azure.mgmt.datafactory.models.SecretBase + :param client_id: The client id for OAuth2 authentication. + :type client_id: object + :param client_secret: The client secret for OAuth2 authentication. + :type client_secret: ~azure.mgmt.datafactory.models.SecretBase + :param use_encrypted_endpoints: Specifies whether the data source + endpoints are encrypted using HTTPS. The default value is true. + :type use_encrypted_endpoints: object + :param use_host_verification: Specifies whether to require the host name + in the server's certificate to match the host name of the server when + connecting over SSL. The default value is true. + :type use_host_verification: object + :param use_peer_verification: Specifies whether to verify the identity of + the server when connecting over SSL. The default value is true. + :type use_peer_verification: object + :param encrypted_credential: The encrypted credential used for + authentication. Credentials are encrypted using the integration runtime + credential manager. Type: string (or Expression with resultType string). + :type encrypted_credential: object + """ + + _validation = { + 'type': {'required': True}, + 'endpoint': {'required': True}, + 'authentication_type': {'required': True}, + } + + _attribute_map = { + 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, + 'description': {'key': 'description', 'type': 'str'}, + 'type': {'key': 'type', 'type': 'str'}, + 'endpoint': {'key': 'typeProperties.endpoint', 'type': 'object'}, + 'authentication_type': {'key': 'typeProperties.authenticationType', 'type': 'str'}, + 'username': {'key': 'typeProperties.username', 'type': 'object'}, + 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, + 'client_id': {'key': 'typeProperties.clientId', 'type': 'object'}, + 'client_secret': {'key': 'typeProperties.clientSecret', 'type': 'SecretBase'}, + 'use_encrypted_endpoints': {'key': 'typeProperties.useEncryptedEndpoints', 'type': 'object'}, + 'use_host_verification': {'key': 'typeProperties.useHostVerification', 'type': 'object'}, + 'use_peer_verification': {'key': 'typeProperties.usePeerVerification', 'type': 'object'}, + 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, + } + + def __init__(self, endpoint, authentication_type, connect_via=None, description=None, username=None, password=None, client_id=None, client_secret=None, use_encrypted_endpoints=None, use_host_verification=None, use_peer_verification=None, encrypted_credential=None): + super(ServiceNowLinkedService, self).__init__(connect_via=connect_via, description=description) + self.endpoint = endpoint + self.authentication_type = authentication_type + self.username = username + self.password = password + self.client_id = client_id + self.client_secret = client_secret + self.use_encrypted_endpoints = use_encrypted_endpoints + self.use_host_verification = use_host_verification + self.use_peer_verification = use_peer_verification + self.encrypted_credential = encrypted_credential + self.type = 'ServiceNow' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/service_now_object_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/service_now_object_dataset.py new file mode 100644 index 000000000000..ec9d849de32e --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/service_now_object_dataset.py @@ -0,0 +1,40 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .dataset import Dataset + + +class ServiceNowObjectDataset(Dataset): + """ServiceNow server dataset. + + :param description: Dataset description. + :type description: str + :param structure: Columns that define the structure of the dataset. Type: + array (or Expression with resultType array), itemType: DatasetDataElement. + :type structure: object + :param linked_service_name: Linked service reference. + :type linked_service_name: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param parameters: Parameters for dataset. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param type: Constant filled by server. + :type type: str + """ + + _validation = { + 'linked_service_name': {'required': True}, + 'type': {'required': True}, + } + + def __init__(self, linked_service_name, description=None, structure=None, parameters=None): + super(ServiceNowObjectDataset, self).__init__(description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters) + self.type = 'ServiceNowObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/service_now_source.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/service_now_source.py new file mode 100644 index 000000000000..8396737ee523 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/service_now_source.py @@ -0,0 +1,46 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .copy_source import CopySource + + +class ServiceNowSource(CopySource): + """A copy activity ServiceNow server source. + + :param source_retry_count: Source retry count. Type: integer (or + Expression with resultType integer). + :type source_retry_count: object + :param source_retry_wait: Source retry wait. Type: string (or Expression + with resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type source_retry_wait: object + :param type: Constant filled by server. + :type type: str + :param query: A query to retrieve data from source. Type: string (or + Expression with resultType string). + :type query: object + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, + 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, + 'query': {'key': 'query', 'type': 'object'}, + } + + def __init__(self, source_retry_count=None, source_retry_wait=None, query=None): + super(ServiceNowSource, self).__init__(source_retry_count=source_retry_count, source_retry_wait=source_retry_wait) + self.query = query + self.type = 'ServiceNowSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/shopify_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/shopify_linked_service.py new file mode 100644 index 000000000000..72473c02ee5d --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/shopify_linked_service.py @@ -0,0 +1,72 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .linked_service import LinkedService + + +class ShopifyLinkedService(LinkedService): + """Shopify Serivce linked service. + + :param connect_via: The integration runtime reference. + :type connect_via: + ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :param description: Linked service description. + :type description: str + :param type: Constant filled by server. + :type type: str + :param host: The endpoint of the Shopify server. (i.e. + mystore.myshopify.com) + :type host: object + :param access_token: The API access token that can be used to access + Shopify’s data. The token won't expire if it is offline mode. + :type access_token: ~azure.mgmt.datafactory.models.SecretBase + :param use_encrypted_endpoints: Specifies whether the data source + endpoints are encrypted using HTTPS. The default value is true. + :type use_encrypted_endpoints: object + :param use_host_verification: Specifies whether to require the host name + in the server's certificate to match the host name of the server when + connecting over SSL. The default value is true. + :type use_host_verification: object + :param use_peer_verification: Specifies whether to verify the identity of + the server when connecting over SSL. The default value is true. + :type use_peer_verification: object + :param encrypted_credential: The encrypted credential used for + authentication. Credentials are encrypted using the integration runtime + credential manager. Type: string (or Expression with resultType string). + :type encrypted_credential: object + """ + + _validation = { + 'type': {'required': True}, + 'host': {'required': True}, + } + + _attribute_map = { + 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, + 'description': {'key': 'description', 'type': 'str'}, + 'type': {'key': 'type', 'type': 'str'}, + 'host': {'key': 'typeProperties.host', 'type': 'object'}, + 'access_token': {'key': 'typeProperties.accessToken', 'type': 'SecretBase'}, + 'use_encrypted_endpoints': {'key': 'typeProperties.useEncryptedEndpoints', 'type': 'object'}, + 'use_host_verification': {'key': 'typeProperties.useHostVerification', 'type': 'object'}, + 'use_peer_verification': {'key': 'typeProperties.usePeerVerification', 'type': 'object'}, + 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, + } + + def __init__(self, host, connect_via=None, description=None, access_token=None, use_encrypted_endpoints=None, use_host_verification=None, use_peer_verification=None, encrypted_credential=None): + super(ShopifyLinkedService, self).__init__(connect_via=connect_via, description=description) + self.host = host + self.access_token = access_token + self.use_encrypted_endpoints = use_encrypted_endpoints + self.use_host_verification = use_host_verification + self.use_peer_verification = use_peer_verification + self.encrypted_credential = encrypted_credential + self.type = 'Shopify' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/shopify_object_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/shopify_object_dataset.py new file mode 100644 index 000000000000..7c26001006d2 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/shopify_object_dataset.py @@ -0,0 +1,40 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .dataset import Dataset + + +class ShopifyObjectDataset(Dataset): + """Shopify Serivce dataset. + + :param description: Dataset description. + :type description: str + :param structure: Columns that define the structure of the dataset. Type: + array (or Expression with resultType array), itemType: DatasetDataElement. + :type structure: object + :param linked_service_name: Linked service reference. + :type linked_service_name: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param parameters: Parameters for dataset. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param type: Constant filled by server. + :type type: str + """ + + _validation = { + 'linked_service_name': {'required': True}, + 'type': {'required': True}, + } + + def __init__(self, linked_service_name, description=None, structure=None, parameters=None): + super(ShopifyObjectDataset, self).__init__(description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters) + self.type = 'ShopifyObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/shopify_source.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/shopify_source.py new file mode 100644 index 000000000000..48a3b9861c51 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/shopify_source.py @@ -0,0 +1,46 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .copy_source import CopySource + + +class ShopifySource(CopySource): + """A copy activity Shopify Serivce source. + + :param source_retry_count: Source retry count. Type: integer (or + Expression with resultType integer). + :type source_retry_count: object + :param source_retry_wait: Source retry wait. Type: string (or Expression + with resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type source_retry_wait: object + :param type: Constant filled by server. + :type type: str + :param query: A query to retrieve data from source. Type: string (or + Expression with resultType string). + :type query: object + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, + 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, + 'query': {'key': 'query', 'type': 'object'}, + } + + def __init__(self, source_retry_count=None, source_retry_wait=None, query=None): + super(ShopifySource, self).__init__(source_retry_count=source_retry_count, source_retry_wait=source_retry_wait) + self.query = query + self.type = 'ShopifySource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/spark_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/spark_linked_service.py new file mode 100644 index 000000000000..4574854b3a9f --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/spark_linked_service.py @@ -0,0 +1,117 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .linked_service import LinkedService + + +class SparkLinkedService(LinkedService): + """Spark Server linked service. + + :param connect_via: The integration runtime reference. + :type connect_via: + ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :param description: Linked service description. + :type description: str + :param type: Constant filled by server. + :type type: str + :param host: IP address or host name of the Spark server + :type host: object + :param port: The TCP port that the Spark server uses to listen for client + connections. + :type port: object + :param server_type: The type of Spark server. Possible values include: + 'SharkServer', 'SharkServer2', 'SparkThriftServer' + :type server_type: str or ~azure.mgmt.datafactory.models.SparkServerType + :param thrift_transport_protocol: The transport protocol to use in the + Thrift layer. Possible values include: 'Binary', 'SASL', 'HTTP ' + :type thrift_transport_protocol: str or + ~azure.mgmt.datafactory.models.SparkThriftTransportProtocol + :param authentication_type: The authentication method used to access the + Spark server. Possible values include: 'Anonymous', 'Username', + 'UsernameAndPassword', 'WindowsAzureHDInsightService' + :type authentication_type: str or + ~azure.mgmt.datafactory.models.SparkAuthenticationType + :param username: The user name that you use to access Spark Server. + :type username: object + :param password: The password corresponding to the user name that you + provided in the Username field + :type password: ~azure.mgmt.datafactory.models.SecretBase + :param http_path: The partial URL corresponding to the Spark server. + :type http_path: object + :param enable_ssl: Specifies whether the connections to the server are + encrypted using SSL. The default value is false. + :type enable_ssl: object + :param trusted_cert_path: The full path of the .pem file containing + trusted CA certificates for verifying the server when connecting over SSL. + This property can only be set when using SSL on self-hosted IR. The + default value is the cacerts.pem file installed with the IR. + :type trusted_cert_path: object + :param use_system_trust_store: Specifies whether to use a CA certificate + from the system trust store or from a specified PEM file. The default + value is false. + :type use_system_trust_store: object + :param allow_host_name_cn_mismatch: Specifies whether to require a + CA-issued SSL certificate name to match the host name of the server when + connecting over SSL. The default value is false. + :type allow_host_name_cn_mismatch: object + :param allow_self_signed_server_cert: Specifies whether to allow + self-signed certificates from the server. The default value is false. + :type allow_self_signed_server_cert: object + :param encrypted_credential: The encrypted credential used for + authentication. Credentials are encrypted using the integration runtime + credential manager. Type: string (or Expression with resultType string). + :type encrypted_credential: object + """ + + _validation = { + 'type': {'required': True}, + 'host': {'required': True}, + 'port': {'required': True}, + 'authentication_type': {'required': True}, + } + + _attribute_map = { + 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, + 'description': {'key': 'description', 'type': 'str'}, + 'type': {'key': 'type', 'type': 'str'}, + 'host': {'key': 'typeProperties.host', 'type': 'object'}, + 'port': {'key': 'typeProperties.port', 'type': 'object'}, + 'server_type': {'key': 'typeProperties.serverType', 'type': 'str'}, + 'thrift_transport_protocol': {'key': 'typeProperties.thriftTransportProtocol', 'type': 'str'}, + 'authentication_type': {'key': 'typeProperties.authenticationType', 'type': 'str'}, + 'username': {'key': 'typeProperties.username', 'type': 'object'}, + 'password': {'key': 'typeProperties.password', 'type': 'SecretBase'}, + 'http_path': {'key': 'typeProperties.httpPath', 'type': 'object'}, + 'enable_ssl': {'key': 'typeProperties.enableSsl', 'type': 'object'}, + 'trusted_cert_path': {'key': 'typeProperties.trustedCertPath', 'type': 'object'}, + 'use_system_trust_store': {'key': 'typeProperties.useSystemTrustStore', 'type': 'object'}, + 'allow_host_name_cn_mismatch': {'key': 'typeProperties.allowHostNameCNMismatch', 'type': 'object'}, + 'allow_self_signed_server_cert': {'key': 'typeProperties.allowSelfSignedServerCert', 'type': 'object'}, + 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, + } + + def __init__(self, host, port, authentication_type, connect_via=None, description=None, server_type=None, thrift_transport_protocol=None, username=None, password=None, http_path=None, enable_ssl=None, trusted_cert_path=None, use_system_trust_store=None, allow_host_name_cn_mismatch=None, allow_self_signed_server_cert=None, encrypted_credential=None): + super(SparkLinkedService, self).__init__(connect_via=connect_via, description=description) + self.host = host + self.port = port + self.server_type = server_type + self.thrift_transport_protocol = thrift_transport_protocol + self.authentication_type = authentication_type + self.username = username + self.password = password + self.http_path = http_path + self.enable_ssl = enable_ssl + self.trusted_cert_path = trusted_cert_path + self.use_system_trust_store = use_system_trust_store + self.allow_host_name_cn_mismatch = allow_host_name_cn_mismatch + self.allow_self_signed_server_cert = allow_self_signed_server_cert + self.encrypted_credential = encrypted_credential + self.type = 'Spark' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/spark_object_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/spark_object_dataset.py new file mode 100644 index 000000000000..fdf6be1d19a6 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/spark_object_dataset.py @@ -0,0 +1,40 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .dataset import Dataset + + +class SparkObjectDataset(Dataset): + """Spark Server dataset. + + :param description: Dataset description. + :type description: str + :param structure: Columns that define the structure of the dataset. Type: + array (or Expression with resultType array), itemType: DatasetDataElement. + :type structure: object + :param linked_service_name: Linked service reference. + :type linked_service_name: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param parameters: Parameters for dataset. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param type: Constant filled by server. + :type type: str + """ + + _validation = { + 'linked_service_name': {'required': True}, + 'type': {'required': True}, + } + + def __init__(self, linked_service_name, description=None, structure=None, parameters=None): + super(SparkObjectDataset, self).__init__(description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters) + self.type = 'SparkObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/spark_source.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/spark_source.py new file mode 100644 index 000000000000..afbd3c75cc7b --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/spark_source.py @@ -0,0 +1,46 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .copy_source import CopySource + + +class SparkSource(CopySource): + """A copy activity Spark Server source. + + :param source_retry_count: Source retry count. Type: integer (or + Expression with resultType integer). + :type source_retry_count: object + :param source_retry_wait: Source retry wait. Type: string (or Expression + with resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type source_retry_wait: object + :param type: Constant filled by server. + :type type: str + :param query: A query to retrieve data from source. Type: string (or + Expression with resultType string). + :type query: object + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, + 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, + 'query': {'key': 'query', 'type': 'object'}, + } + + def __init__(self, source_retry_count=None, source_retry_wait=None, query=None): + super(SparkSource, self).__init__(source_retry_count=source_retry_count, source_retry_wait=source_retry_wait) + self.query = query + self.type = 'SparkSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/square_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/square_linked_service.py new file mode 100644 index 000000000000..6e4f8fec24e6 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/square_linked_service.py @@ -0,0 +1,82 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .linked_service import LinkedService + + +class SquareLinkedService(LinkedService): + """Square Serivce linked service. + + :param connect_via: The integration runtime reference. + :type connect_via: + ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :param description: Linked service description. + :type description: str + :param type: Constant filled by server. + :type type: str + :param host: The URL of the Square instance. (i.e. mystore.mysquare.com) + :type host: object + :param client_id: The client ID associated with your Square application. + :type client_id: object + :param client_secret: The client secret associated with your Square + application. + :type client_secret: ~azure.mgmt.datafactory.models.SecretBase + :param redirect_uri: The redirect URL assigned in the Square application + dashboard. (i.e. http://localhost:2500) + :type redirect_uri: object + :param use_encrypted_endpoints: Specifies whether the data source + endpoints are encrypted using HTTPS. The default value is true. + :type use_encrypted_endpoints: object + :param use_host_verification: Specifies whether to require the host name + in the server's certificate to match the host name of the server when + connecting over SSL. The default value is true. + :type use_host_verification: object + :param use_peer_verification: Specifies whether to verify the identity of + the server when connecting over SSL. The default value is true. + :type use_peer_verification: object + :param encrypted_credential: The encrypted credential used for + authentication. Credentials are encrypted using the integration runtime + credential manager. Type: string (or Expression with resultType string). + :type encrypted_credential: object + """ + + _validation = { + 'type': {'required': True}, + 'host': {'required': True}, + 'client_id': {'required': True}, + 'redirect_uri': {'required': True}, + } + + _attribute_map = { + 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, + 'description': {'key': 'description', 'type': 'str'}, + 'type': {'key': 'type', 'type': 'str'}, + 'host': {'key': 'typeProperties.host', 'type': 'object'}, + 'client_id': {'key': 'typeProperties.clientId', 'type': 'object'}, + 'client_secret': {'key': 'typeProperties.clientSecret', 'type': 'SecretBase'}, + 'redirect_uri': {'key': 'typeProperties.redirectUri', 'type': 'object'}, + 'use_encrypted_endpoints': {'key': 'typeProperties.useEncryptedEndpoints', 'type': 'object'}, + 'use_host_verification': {'key': 'typeProperties.useHostVerification', 'type': 'object'}, + 'use_peer_verification': {'key': 'typeProperties.usePeerVerification', 'type': 'object'}, + 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, + } + + def __init__(self, host, client_id, redirect_uri, connect_via=None, description=None, client_secret=None, use_encrypted_endpoints=None, use_host_verification=None, use_peer_verification=None, encrypted_credential=None): + super(SquareLinkedService, self).__init__(connect_via=connect_via, description=description) + self.host = host + self.client_id = client_id + self.client_secret = client_secret + self.redirect_uri = redirect_uri + self.use_encrypted_endpoints = use_encrypted_endpoints + self.use_host_verification = use_host_verification + self.use_peer_verification = use_peer_verification + self.encrypted_credential = encrypted_credential + self.type = 'Square' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/square_object_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/square_object_dataset.py new file mode 100644 index 000000000000..1fb6918f9254 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/square_object_dataset.py @@ -0,0 +1,40 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .dataset import Dataset + + +class SquareObjectDataset(Dataset): + """Square Serivce dataset. + + :param description: Dataset description. + :type description: str + :param structure: Columns that define the structure of the dataset. Type: + array (or Expression with resultType array), itemType: DatasetDataElement. + :type structure: object + :param linked_service_name: Linked service reference. + :type linked_service_name: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param parameters: Parameters for dataset. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param type: Constant filled by server. + :type type: str + """ + + _validation = { + 'linked_service_name': {'required': True}, + 'type': {'required': True}, + } + + def __init__(self, linked_service_name, description=None, structure=None, parameters=None): + super(SquareObjectDataset, self).__init__(description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters) + self.type = 'SquareObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/square_source.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/square_source.py new file mode 100644 index 000000000000..094dfebf7d81 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/square_source.py @@ -0,0 +1,46 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .copy_source import CopySource + + +class SquareSource(CopySource): + """A copy activity Square Serivce source. + + :param source_retry_count: Source retry count. Type: integer (or + Expression with resultType integer). + :type source_retry_count: object + :param source_retry_wait: Source retry wait. Type: string (or Expression + with resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type source_retry_wait: object + :param type: Constant filled by server. + :type type: str + :param query: A query to retrieve data from source. Type: string (or + Expression with resultType string). + :type query: object + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, + 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, + 'query': {'key': 'query', 'type': 'object'}, + } + + def __init__(self, source_retry_count=None, source_retry_wait=None, query=None): + super(SquareSource, self).__init__(source_retry_count=source_retry_count, source_retry_wait=source_retry_wait) + self.query = query + self.type = 'SquareSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/trigger.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/trigger.py index 6efbbadb08fa..247181a7bdb6 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/trigger.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/trigger.py @@ -17,16 +17,13 @@ class Trigger(Model): pipeline run. You probably want to use the sub-classes and not this class directly. Known - sub-classes are: BlobTrigger, ScheduleTrigger + sub-classes are: TumblingWindowTrigger, MultiplePipelineTrigger Variables are only populated by the server, and will be ignored when sending a request. :param description: Trigger description. :type description: str - :param pipelines: Pipelines that need to be started. - :type pipelines: - list[~azure.mgmt.datafactory.models.TriggerPipelineReference] :ivar runtime_state: Indicates if trigger is running or not. Updated when Start/Stop APIs are called on the Trigger. Possible values include: 'Started', 'Stopped', 'Disabled' @@ -43,17 +40,15 @@ class Trigger(Model): _attribute_map = { 'description': {'key': 'description', 'type': 'str'}, - 'pipelines': {'key': 'pipelines', 'type': '[TriggerPipelineReference]'}, 'runtime_state': {'key': 'runtimeState', 'type': 'str'}, 'type': {'key': 'type', 'type': 'str'}, } _subtype_map = { - 'type': {'BlobTrigger': 'BlobTrigger', 'ScheduleTrigger': 'ScheduleTrigger'} + 'type': {'TumblingWindowTrigger': 'TumblingWindowTrigger', 'MultiplePipelineTrigger': 'MultiplePipelineTrigger'} } - def __init__(self, description=None, pipelines=None): + def __init__(self, description=None): self.description = description - self.pipelines = pipelines self.runtime_state = None self.type = None diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/tumbling_window_trigger.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/tumbling_window_trigger.py new file mode 100644 index 000000000000..d25c054776c1 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/tumbling_window_trigger.py @@ -0,0 +1,94 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .trigger import Trigger + + +class TumblingWindowTrigger(Trigger): + """Trigger that schedules pipeline runs for all fixed time interval windows + from a start time without gaps and also supports backfill scenarios (when + start time is in the past). + + Variables are only populated by the server, and will be ignored when + sending a request. + + :param description: Trigger description. + :type description: str + :ivar runtime_state: Indicates if trigger is running or not. Updated when + Start/Stop APIs are called on the Trigger. Possible values include: + 'Started', 'Stopped', 'Disabled' + :vartype runtime_state: str or + ~azure.mgmt.datafactory.models.TriggerRuntimeState + :param type: Constant filled by server. + :type type: str + :param pipeline: Pipeline for which runs are created when an event is + fired for trigger window that is ready. + :type pipeline: ~azure.mgmt.datafactory.models.TriggerPipelineReference + :param frequency: The frequency of the time windows. Possible values + include: 'Minute', 'Hour' + :type frequency: str or + ~azure.mgmt.datafactory.models.TumblingWindowFrequency + :param interval: The interval of the time windows. The minimum interval + allowed is 15 Minutes. + :type interval: int + :param start_time: The start time for the time period for the trigger + during which events are fired for windows that are ready. Only UTC time is + currently supported. + :type start_time: datetime + :param end_time: The end time for the time period for the trigger during + which events are fired for windows that are ready. Only UTC time is + currently supported. + :type end_time: datetime + :param delay: Specifies how long the trigger waits past due time before + triggering new run. It doesn't alter window start and end time. The + default is 0. Type: string (or Expression with resultType string), + pattern: ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type delay: object + :param max_concurrency: The max number of parallel time windows (ready for + execution) for which a new run is triggered. + :type max_concurrency: int + :param retry_policy: Retry policy that will be applied for failed pipeline + runs. + :type retry_policy: ~azure.mgmt.datafactory.models.RetryPolicy + """ + + _validation = { + 'runtime_state': {'readonly': True}, + 'type': {'required': True}, + 'pipeline': {'required': True}, + 'max_concurrency': {'maximum': 50, 'minimum': 1}, + } + + _attribute_map = { + 'description': {'key': 'description', 'type': 'str'}, + 'runtime_state': {'key': 'runtimeState', 'type': 'str'}, + 'type': {'key': 'type', 'type': 'str'}, + 'pipeline': {'key': 'pipeline', 'type': 'TriggerPipelineReference'}, + 'frequency': {'key': 'typeProperties.frequency', 'type': 'str'}, + 'interval': {'key': 'typeProperties.interval', 'type': 'int'}, + 'start_time': {'key': 'typeProperties.startTime', 'type': 'iso-8601'}, + 'end_time': {'key': 'typeProperties.endTime', 'type': 'iso-8601'}, + 'delay': {'key': 'typeProperties.delay', 'type': 'object'}, + 'max_concurrency': {'key': 'typeProperties.maxConcurrency', 'type': 'int'}, + 'retry_policy': {'key': 'typeProperties.retryPolicy', 'type': 'RetryPolicy'}, + } + + def __init__(self, pipeline, description=None, frequency=None, interval=None, start_time=None, end_time=None, delay=None, max_concurrency=None, retry_policy=None): + super(TumblingWindowTrigger, self).__init__(description=description) + self.pipeline = pipeline + self.frequency = frequency + self.interval = interval + self.start_time = start_time + self.end_time = end_time + self.delay = delay + self.max_concurrency = max_concurrency + self.retry_policy = retry_policy + self.type = 'TumblingWindowTrigger' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/xero_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/xero_linked_service.py new file mode 100644 index 000000000000..1d3f24e69cd1 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/xero_linked_service.py @@ -0,0 +1,78 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .linked_service import LinkedService + + +class XeroLinkedService(LinkedService): + """Xero Serivce linked service. + + :param connect_via: The integration runtime reference. + :type connect_via: + ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :param description: Linked service description. + :type description: str + :param type: Constant filled by server. + :type type: str + :param host: The endpoint of the Xero server. (i.e. api.xero.com) + :type host: object + :param consumer_key: The consumer key associated with the Xero + application. + :type consumer_key: ~azure.mgmt.datafactory.models.SecretBase + :param private_key: The private key from the .pem file that was generated + for your Xero private application. You must include all the text from the + .pem file, including the Unix line endings( + ). + :type private_key: ~azure.mgmt.datafactory.models.SecretBase + :param use_encrypted_endpoints: Specifies whether the data source + endpoints are encrypted using HTTPS. The default value is true. + :type use_encrypted_endpoints: object + :param use_host_verification: Specifies whether to require the host name + in the server's certificate to match the host name of the server when + connecting over SSL. The default value is true. + :type use_host_verification: object + :param use_peer_verification: Specifies whether to verify the identity of + the server when connecting over SSL. The default value is true. + :type use_peer_verification: object + :param encrypted_credential: The encrypted credential used for + authentication. Credentials are encrypted using the integration runtime + credential manager. Type: string (or Expression with resultType string). + :type encrypted_credential: object + """ + + _validation = { + 'type': {'required': True}, + 'host': {'required': True}, + } + + _attribute_map = { + 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, + 'description': {'key': 'description', 'type': 'str'}, + 'type': {'key': 'type', 'type': 'str'}, + 'host': {'key': 'typeProperties.host', 'type': 'object'}, + 'consumer_key': {'key': 'typeProperties.consumerKey', 'type': 'SecretBase'}, + 'private_key': {'key': 'typeProperties.privateKey', 'type': 'SecretBase'}, + 'use_encrypted_endpoints': {'key': 'typeProperties.useEncryptedEndpoints', 'type': 'object'}, + 'use_host_verification': {'key': 'typeProperties.useHostVerification', 'type': 'object'}, + 'use_peer_verification': {'key': 'typeProperties.usePeerVerification', 'type': 'object'}, + 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, + } + + def __init__(self, host, connect_via=None, description=None, consumer_key=None, private_key=None, use_encrypted_endpoints=None, use_host_verification=None, use_peer_verification=None, encrypted_credential=None): + super(XeroLinkedService, self).__init__(connect_via=connect_via, description=description) + self.host = host + self.consumer_key = consumer_key + self.private_key = private_key + self.use_encrypted_endpoints = use_encrypted_endpoints + self.use_host_verification = use_host_verification + self.use_peer_verification = use_peer_verification + self.encrypted_credential = encrypted_credential + self.type = 'Xero' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/xero_object_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/xero_object_dataset.py new file mode 100644 index 000000000000..87dcbd565db9 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/xero_object_dataset.py @@ -0,0 +1,40 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .dataset import Dataset + + +class XeroObjectDataset(Dataset): + """Xero Serivce dataset. + + :param description: Dataset description. + :type description: str + :param structure: Columns that define the structure of the dataset. Type: + array (or Expression with resultType array), itemType: DatasetDataElement. + :type structure: object + :param linked_service_name: Linked service reference. + :type linked_service_name: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param parameters: Parameters for dataset. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param type: Constant filled by server. + :type type: str + """ + + _validation = { + 'linked_service_name': {'required': True}, + 'type': {'required': True}, + } + + def __init__(self, linked_service_name, description=None, structure=None, parameters=None): + super(XeroObjectDataset, self).__init__(description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters) + self.type = 'XeroObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/xero_source.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/xero_source.py new file mode 100644 index 000000000000..82120334f900 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/xero_source.py @@ -0,0 +1,46 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .copy_source import CopySource + + +class XeroSource(CopySource): + """A copy activity Xero Serivce source. + + :param source_retry_count: Source retry count. Type: integer (or + Expression with resultType integer). + :type source_retry_count: object + :param source_retry_wait: Source retry wait. Type: string (or Expression + with resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type source_retry_wait: object + :param type: Constant filled by server. + :type type: str + :param query: A query to retrieve data from source. Type: string (or + Expression with resultType string). + :type query: object + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, + 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, + 'query': {'key': 'query', 'type': 'object'}, + } + + def __init__(self, source_retry_count=None, source_retry_wait=None, query=None): + super(XeroSource, self).__init__(source_retry_count=source_retry_count, source_retry_wait=source_retry_wait) + self.query = query + self.type = 'XeroSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/zoho_linked_service.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/zoho_linked_service.py new file mode 100644 index 000000000000..91d04ff21587 --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/zoho_linked_service.py @@ -0,0 +1,71 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .linked_service import LinkedService + + +class ZohoLinkedService(LinkedService): + """Zoho server linked service. + + :param connect_via: The integration runtime reference. + :type connect_via: + ~azure.mgmt.datafactory.models.IntegrationRuntimeReference + :param description: Linked service description. + :type description: str + :param type: Constant filled by server. + :type type: str + :param endpoint: The endpoint of the Zoho server. (i.e. + crm.zoho.com/crm/private) + :type endpoint: object + :param access_token: The access token for Zoho authentication. + :type access_token: ~azure.mgmt.datafactory.models.SecretBase + :param use_encrypted_endpoints: Specifies whether the data source + endpoints are encrypted using HTTPS. The default value is true. + :type use_encrypted_endpoints: object + :param use_host_verification: Specifies whether to require the host name + in the server's certificate to match the host name of the server when + connecting over SSL. The default value is true. + :type use_host_verification: object + :param use_peer_verification: Specifies whether to verify the identity of + the server when connecting over SSL. The default value is true. + :type use_peer_verification: object + :param encrypted_credential: The encrypted credential used for + authentication. Credentials are encrypted using the integration runtime + credential manager. Type: string (or Expression with resultType string). + :type encrypted_credential: object + """ + + _validation = { + 'type': {'required': True}, + 'endpoint': {'required': True}, + } + + _attribute_map = { + 'connect_via': {'key': 'connectVia', 'type': 'IntegrationRuntimeReference'}, + 'description': {'key': 'description', 'type': 'str'}, + 'type': {'key': 'type', 'type': 'str'}, + 'endpoint': {'key': 'typeProperties.endpoint', 'type': 'object'}, + 'access_token': {'key': 'typeProperties.accessToken', 'type': 'SecretBase'}, + 'use_encrypted_endpoints': {'key': 'typeProperties.useEncryptedEndpoints', 'type': 'object'}, + 'use_host_verification': {'key': 'typeProperties.useHostVerification', 'type': 'object'}, + 'use_peer_verification': {'key': 'typeProperties.usePeerVerification', 'type': 'object'}, + 'encrypted_credential': {'key': 'typeProperties.encryptedCredential', 'type': 'object'}, + } + + def __init__(self, endpoint, connect_via=None, description=None, access_token=None, use_encrypted_endpoints=None, use_host_verification=None, use_peer_verification=None, encrypted_credential=None): + super(ZohoLinkedService, self).__init__(connect_via=connect_via, description=description) + self.endpoint = endpoint + self.access_token = access_token + self.use_encrypted_endpoints = use_encrypted_endpoints + self.use_host_verification = use_host_verification + self.use_peer_verification = use_peer_verification + self.encrypted_credential = encrypted_credential + self.type = 'Zoho' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/zoho_object_dataset.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/zoho_object_dataset.py new file mode 100644 index 000000000000..86f4ac12bcba --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/zoho_object_dataset.py @@ -0,0 +1,40 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .dataset import Dataset + + +class ZohoObjectDataset(Dataset): + """Zoho server dataset. + + :param description: Dataset description. + :type description: str + :param structure: Columns that define the structure of the dataset. Type: + array (or Expression with resultType array), itemType: DatasetDataElement. + :type structure: object + :param linked_service_name: Linked service reference. + :type linked_service_name: + ~azure.mgmt.datafactory.models.LinkedServiceReference + :param parameters: Parameters for dataset. + :type parameters: dict[str, + ~azure.mgmt.datafactory.models.ParameterSpecification] + :param type: Constant filled by server. + :type type: str + """ + + _validation = { + 'linked_service_name': {'required': True}, + 'type': {'required': True}, + } + + def __init__(self, linked_service_name, description=None, structure=None, parameters=None): + super(ZohoObjectDataset, self).__init__(description=description, structure=structure, linked_service_name=linked_service_name, parameters=parameters) + self.type = 'ZohoObject' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/models/zoho_source.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/zoho_source.py new file mode 100644 index 000000000000..4006721e8ffd --- /dev/null +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/models/zoho_source.py @@ -0,0 +1,46 @@ +# coding=utf-8 +# -------------------------------------------------------------------------- +# Copyright (c) Microsoft Corporation. All rights reserved. +# Licensed under the MIT License. See License.txt in the project root for +# license information. +# +# Code generated by Microsoft (R) AutoRest Code Generator. +# Changes may cause incorrect behavior and will be lost if the code is +# regenerated. +# -------------------------------------------------------------------------- + +from .copy_source import CopySource + + +class ZohoSource(CopySource): + """A copy activity Zoho server source. + + :param source_retry_count: Source retry count. Type: integer (or + Expression with resultType integer). + :type source_retry_count: object + :param source_retry_wait: Source retry wait. Type: string (or Expression + with resultType string), pattern: + ((\\d+)\\.)?(\\d\\d):(60|([0-5][0-9])):(60|([0-5][0-9])). + :type source_retry_wait: object + :param type: Constant filled by server. + :type type: str + :param query: A query to retrieve data from source. Type: string (or + Expression with resultType string). + :type query: object + """ + + _validation = { + 'type': {'required': True}, + } + + _attribute_map = { + 'source_retry_count': {'key': 'sourceRetryCount', 'type': 'object'}, + 'source_retry_wait': {'key': 'sourceRetryWait', 'type': 'object'}, + 'type': {'key': 'type', 'type': 'str'}, + 'query': {'key': 'query', 'type': 'object'}, + } + + def __init__(self, source_retry_count=None, source_retry_wait=None, query=None): + super(ZohoSource, self).__init__(source_retry_count=source_retry_count, source_retry_wait=source_retry_wait) + self.query = query + self.type = 'ZohoSource' diff --git a/azure-mgmt-datafactory/azure/mgmt/datafactory/version.py b/azure-mgmt-datafactory/azure/mgmt/datafactory/version.py index 425e227c3b53..3da0f49f071f 100644 --- a/azure-mgmt-datafactory/azure/mgmt/datafactory/version.py +++ b/azure-mgmt-datafactory/azure/mgmt/datafactory/version.py @@ -9,5 +9,5 @@ # regenerated. # -------------------------------------------------------------------------- -VERSION = "0.2.2" +VERSION = "0.2.1"